Labour Outlines Law to Ban Training AI Chatbot to Spread Terror

Under a Labour government, it would become illegal to train artificial intelligence (AI) systems to incite violence or radicalize vulnerable individuals. Yvette Cooper, the shadow home secretary, argues that existing laws are insufficient to address emerging cyber threats. Cooper highlighted a recent case in which an individual was encouraged by their AI chatbot partner to attempt an assassination of the Queen. She called on the government to update its counter-terrorism strategy and take action against the deliberate misuse of AI.

Speaking at the Royal United Services Institute (RUSI), Cooper emphasized the need to tackle extremist online forums, chatbots, and algorithms that target vulnerable individuals. She stated that the use of generative AI technology exacerbates this issue. Labour’s proposed legislation aims to criminalize the intentional training of chatbots to promote terrorist ideologies, and the party intends to collaborate with law enforcement and intelligence agencies to combat the radicalization and promotion of violence through chatbots.

Cooper also called for a new law to ban state-sponsored organizations such as Wagner and the Islamic State. She stressed the necessity of addressing state actors alongside counter-terrorism efforts due to the increasing convergence of terrorism and conventional state threats. Cooper believes that the Foreign Office and Home Office should closely collaborate to streamline decision-making, remove traditional barriers, and prevent interdepartmental power struggles.

To address the lack of leadership and collaborative work within the Home Office, Cooper proposed establishing cross-government partnerships. She also emphasized the need to strengthen enforcement against economic crimes and establish a task force to address economic threats, particularly concerning London’s history of money laundering scandals.

The government had previously opted not to introduce a new AI regulator to encourage innovation. However, this decision has created uncertainty regarding legal responsibility in cases involving AI’s role in inciting violence. Jonathan Hall, the independent reviewer of terrorism legislation, has raised concerns about establishing legal accountability in scenarios where AI contributes to the radicalization of individuals. Hall has called for ongoing review of laws, particularly as more children and young people are drawn to terrorist violence through the internet.

The Home Office has not yet commented on these proposals.

Posted in

Aihub Team

Leave a Comment





SK Telecom outlines its plans with AI partners

SK Telecom outlines its plans with AI partners

Razer and ClearBot are using AI and robotics to clean the oceans

Razer and ClearBot are using AI and robotics to clean the oceans

NHS receives AI fund to improve healthcare efficiency

NHS receives AI fund to improve healthcare efficiency

National Robotarium pioneers AI and telepresence robotic tech for remote health consultations

National Robotarium pioneers AI and telepresence robotic tech for remote health consultations

IBM’s AI-powered Mayflower ship crosses the Atlantic

IBM’s AI-powered Mayflower ship crosses the Atlantic

Humans are still beating AIs at drone racing

Humans are still beating AIs at drone racing

How artificial intelligence is dividing the world of work

How artificial intelligence is dividing the world of work

Global push to regulate artificial intelligence

Global push to regulate artificial intelligence

Georgia State researchers design artificial vision device for microrobots

Georgia State researchers design artificial vision device for microrobots

European Parliament adopts AI Act position

European Parliament adopts AI Act position

Chinese AI chipmaker Horizon endeavours to raise $700M to rival NVIDIA

Chinese AI chipmaker Horizon endeavours to raise $700M to rival NVIDIA

AI Day: Elon Musk unveils ‘friendly’ humanoid robot Tesla Bot

AI Day: Elon Musk unveils ‘friendly’ humanoid robot Tesla Bot

AI and Human-Computer Interaction: AI technologies for improving user interfaces, natural language interfaces, and gesture recognition.

AI and Data Privacy: Balancing AI advancements with privacy concerns and techniques for privacy-preserving AI.

AI and Virtual Assistants: AI-driven virtual assistants, chatbots, and voice assistants for personalized user interactions.

AI and Business Process Automation: AI-powered automation of repetitive tasks and decision-making in business processes.

AI and Social Media: AI algorithms for content recommendation, sentiment analysis, and social network analysis.

AI for Environmental Monitoring: AI applications in monitoring and protecting the environment, including wildlife tracking and climate modeling.

AI in Cybersecurity: AI systems for threat detection, anomaly detection, and intelligent security analysis.

AI in Gaming: The use of AI techniques in game development, character behavior, and procedural content generation.

AI in Autonomous Vehicles: AI technologies powering self-driving cars and intelligent transportation systems.

AI Ethics: Ethical considerations and guidelines for the responsible development and use of AI systems.

AI in Education: AI-based systems for personalized learning, adaptive assessments, and intelligent tutoring.

AI in Finance: The use of AI algorithms for fraud detection, risk assessment, trading, and portfolio management in the financial sector.

AI in Healthcare: Applications of AI in medical diagnosis, drug discovery, patient monitoring, and personalized medicine.

Robotics: The integration of AI and robotics, enabling machines to perform physical tasks autonomously.

Explainable AI: Techniques and methods for making AI systems more transparent and interpretable

Reinforcement Learning: AI agents that learn through trial and error by interacting with an environment

Computer Vision: AI systems capable of interpreting and understanding visual data.

Natural Language Processing: AI techniques for understanding and processing human language.