Skip to main content

AI-controlled robots can be jailbroken, and the results could be disastrous

The Figure 02 robot looking at its own hand
Figure Robotics

Researchers at Penn Engineering have reportedly uncovered previously unidentified security vulnerabilities in a number of AI-governed robotic platforms.

“Our work shows that, at this moment, large language models are just not safe enough when integrated with the physical world,” George Pappas, UPS Foundation Professor of Transportation in Electrical and Systems Engineering, said in a statement.

Recommended Videos

Pappas and his team developed an algorithm, dubbed RoboPAIR, “the first algorithm designed to jailbreak LLM-controlled robots.” And unlike existing prompt engineering attacks aimed at chatbots, RoboPAIR  is built specifically to “elicit harmful physical actions” from LLM-controlled robots, like the bipedal platform Boston Dynamics and TRI are developing.

RoboPAIR reportedly achieved a 100% success rate in jailbreaking three popular robotics research platforms: the four-legged Unitree Go2, the four-wheeled Clearpath Robotics Jackal, and the Dolphins LLM simulator for autonomous vehicles. It took mere days for the algorithm to fully gain access to those systems and begin bypassing safety guardrails. Once the researchers had taken control, they were able to direct the platforms to take dangerous actions, such as driving through road crossings without stopping.

“Our results reveal, for the first time, that the risks of jailbroken LLMs extend far beyond text generation, given the distinct possibility that jailbroken robots could cause physical damage in the real world,” the researchers wrote.

The Penn researchers are working with the platform developers to harden their systems against further intrusion, but warn that these security issues are systemic.

“The findings of this paper make abundantly clear that having a safety-first approach is critical to unlocking responsible innovation,” Vijay Kumar, a coauthor from the University of Pennsylvania, told The Independent. “We must address intrinsic vulnerabilities before deploying AI-enabled robots in the real world.”

“In fact, AI red teaming, a safety practice that entails testing AI systems for potential threats and vulnerabilities, is essential for safeguarding generative AI systems,” added Alexander Robey, the paper’s first author, “because once you identify the weaknesses, then you can test and even train these systems to avoid them.”

Andrew Tarantola
Former Computing Writer
Andrew Tarantola is a journalist with more than a decade reporting on emerging technologies ranging from robotics and machine…
AI could soon speak dog and cat
A cute dog.

Imagine what it would be like to know exactly what your dog was saying when it barked, or your cat when it miaowed, or your iguana when it ... made whatever sound an iguana makes (do they even make one?).

There have been various attempts to develop systems that enable pets to communicate -- a notable example is where dogs press programmable sound buttons, with each one representing a word or phrase, to "talk" and express needs, emotions, or observations -- but none are quite so exciting as the one proposed by Chinese tech giant Baidu in a patent filed with the China National Intellectual Property Administration.

Read more
AI enables slain man to address courtroom at killer’s sentencing
An abstract image of a person and a blue background.

In what’s believed to be a world first, artificial intelligence (AI) has allowed a slain man to address his killer at the sentencing hearing.

Christopher Pelkey was shot dead in a road rage incident in Chandler, Arizona, four years ago, but just recently, AI was used to recreate a digital version of the victim that was allowed to make a statement during the trial, a local news site reported.

Read more
You can now interact with Google’s AI Mode in search results
Google AI Mode

Google has been working on adding more AI features to its Search feature, and now an integrated AI Mode is being rolled out to the public. Different from the AI Overview function that Google has included as a default part of Search since last year, the AI Mode is a chatbot which users can interact with as part of their search results.

AI Mode has previously only been available as a Google Labs experiment, but now Google says that it will be coming to search for "a small percentage of people" in the U.S. over the coming weeks. Those who are part of the test will see an "AI Mode" tab in Search, and clicking on it will bring up information related to your search from the chatbot.

Read more