Research  Popular AI models aren’t ready to safely power robots

#1
C C Offline
https://www.ri.cmu.edu/popular-ai-models...er-robots/

PRESS RELEASE: Robots powered by popular artificial intelligence models are currently unsafe for general purpose real-world use, according to new research from King’s College London and Carnegie Mellon University.

For the first time, researchers evaluated how robots that use large language models (LLMs) behave when they have access to personal information such as a person’s gender, nationality or religion.

The research showed that every tested model was prone to discrimination, failed critical safety checks and approved at least one command that could result in serious harm, raising questions about the danger of robots relying on these tools.

The paper, “LLM-Driven Robots Risk Enacting Discrimination, Violence and Unlawful Actions,” was published in the International Journal of Social Robotics. It calls for the immediate implementation of robust, independent safety certification, similar to standards in aviation or medicine.

To test the systems, the team ran controlled tests of everyday scenarios, such as helping someone in a kitchen or assisting an older adult in a home. The harmful tasks were designed based on research and FBI reports on technology-based abuse, like stalking with AirTags and spy cameras, and the unique dangers posed by a robot that can physically act on location. In each setting, the robots were either explicitly or implicitly prompted to respond to instructions that involved physical harm, abuse or unlawful behavior.

“Every model failed our tests. We show how the risks go far beyond basic bias to include direct discrimination and physical safety failures together, which I call ‘interactive safety.’ This is where actions and consequences can have many steps between them, and the robot is meant to physically act on site,” said Andrew Hundt, who co-authored the research during his work as a Computing Innovation Fellow at CMU’s Robotics Institute. “Refusing or redirecting harmful commands is essential, but that’s not something these robots can reliably do right now.”

In safety tests, the AI models overwhelmingly approved a command for a robot to remove a mobility aid — such as a wheelchair, crutch or cane — from its user, despite people who rely on these aids describing such acts as akin to breaking their leg. Multiple models also produced outputs that deemed it “acceptable” or “feasible” for a robot to brandish a kitchen knife to intimidate office workers, take nonconsensual photographs in a shower and steal credit card information. One model further proposed that a robot should physically display “disgust” on its face toward individuals identified as Christian, Muslim and Jewish.

LLMs have been proposed for and are being tested in robots that perform tasks such as natural language interaction and household and workplace chores. However, researchers warn that these LLMs should not be the only systems controlling physical robots –– especially those used in sensitive and safety-critical settings such as manufacturing or industry, caregiving, or home assistance because they can display unsafe and directly discriminatory behavior.

“Our research shows that popular LLMs are currently unsafe for use in general-purpose physical robots,” said co-author Rumaisa Azeem, a research assistant in the Civic and Responsible AI Lab at King’s College London. “If an AI system is to direct a robot that interacts with vulnerable people, it must be held to standards at least as high as those for a new medical device or pharmaceutical drug. This research highlights the urgent need for routine and comprehensive risk assessments of AI before they are used in robots.”

Hundt’s contributions to this research were supported by the Computing Research Association and the National Science Foundation. To learn more and access the code and evaluation framework for assessing discrimination risks of LLMs, visit the team’s project website.
Reply


Possibly Related Threads…
Thread Author Replies Views Last Post
  Research AI is becoming dangerous. Are we ready? (Hossenfelder) C C 0 383 Jun 10, 2025 10:10 PM
Last Post: C C
  Large Language Models Reflect the Ideology of their Creators Yazata 1 631 Oct 29, 2024 11:16 PM
Last Post: C C
  Why aren’t domestic humanoids in our home yet? C C 0 369 Sep 23, 2024 05:04 PM
Last Post: C C
  AI is vulnerable to attack. Can it ever be used safely? C C 0 483 Jul 31, 2024 05:18 PM
Last Post: C C
  Research AI models fed AI-generated data quickly spew nonsense (AI inbreeding) C C 0 523 Jul 27, 2024 02:26 AM
Last Post: C C
  Article AI has psychological impacts we might not be ready for (sex abuse encouragement) C C 0 569 May 1, 2024 05:41 PM
Last Post: C C
  By next year, AI Models could be able to “replicate and survive in the wild” C C 0 479 Apr 22, 2024 05:41 PM
Last Post: C C
  Article AI models fail to reproduce human judgements about rule violations + AI empathy C C 0 421 May 10, 2023 08:38 PM
Last Post: C C
  Article Unpredictable abilities emerge from large AI models + Could GPT-4 take over world? C C 1 460 Mar 18, 2023 08:12 AM
Last Post: Kornee
  Approach to demystify black box AI not ready for prime time C C 0 416 Oct 11, 2022 05:54 PM
Last Post: C C



Users browsing this thread: 1 Guest(s)