A new law aimed at preventing extremist groups from training artificial intelligence (AI) chatbots to spread terrorism would be introduced under a Labour government, Yvette Cooper will say.
In a major speech on Monday, the shadow home secretary will set out the party’s approach to national security, including action to stop online radicalisation.
Ms Cooper will argue that while encouraging terrorism is a criminal offence under existing legislation, it is harder to establish culpability when AI is being used as a tool for promotion.
She will announce that Labour would close the “loophole” by criminalising the deliberate training of chatbots to radicalise vulnerable people.
The party would work closely with the intelligence community towards preventing malicious actors using the technology for such purposes, Ms Cooper will pledge.
Speaking at a Royal United Services Institute (Rusi) defence and security think tank event, she is expected to say: “Artificial intelligence creates new opportunities for Britain, including for law enforcement, but it also presents significant new threats and risks.
“A series of recent cases have revealed the potential for online chatbots to be used to radicalise people with pro-terror content.
“Our law enforcement and legislation must not be outpaced by terrorists and extremists using new technologies to prey on vulnerable people.
“That’s why Labour will criminalise those who purposely train chatbots to spout terrorist material, with stronger action to monitor and stop radicalising chatbots that are inciting violence or amplifying extremist views.”
With our five missions to build a better Britain, my Labour government will restore hope and deliver a country that works for working people again. pic.twitter.com/5aAVAYKiHk
— Keir Starmer (@Keir_Starmer) July 16, 2023
In her speech, the shadow minister will call on the Government to include action to tackle the deliberate misuse of AI and the rise in online radicalisation in an update on counter-terrorism strategy expected this week.
But a new cross-government strategy for state threats is also needed to run alongside the UK’s plan, she will argue.
Ms Cooper will point to Boris Johnson’s meeting with former KGB officer Alexander Lebedev while he was foreign secretary as an example of behaviour which showed disregard for the importance of national security.
Mr Johnson has said officials were aware in advance that he was attending the house of Evgeny Lebedev and said his contact with the newspaper owner’s father at the event was not “formal” or “pre-arranged.”
Labour’s announcement comes after the sentencing of Jaswant Singh Chail, a former supermarket worker who planned an attack on the Queen at Windsor Castle after being encouraged by his “AI girlfriend”.
The Government’s independent reviewer of terrorism legislation, Jonathan Hall KC, has previously highlighted some of the legal challenges posed by the potential use of AI for radicalisation.
In a report in June, he said: “It is unclear how legal culpability would be established in a scenario where an individual was radicalised (in part) by an AI system.
“Although no case currently exists for revisiting terrorism legislation, the information-gathering capability of AI large language models, and the possibility of truly autonomous target selection, mean that these laws need to be kept under close review.”