Skip to content

x-zheng16/Awesome-Embodied-AI-Safety

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

37 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Safety in Embodied AI: A Survey of Risks, Attacks, and Defenses

Paper Website License: CC BY-NC-SA 4.0 Awesome Papers Maintenance GitHub stars GitHub forks

A comprehensive survey and the first unified safety framework for embodied AI, covering 400+ key works across perception, cognition, planning, interaction, and agentic systems.

[Paper] | [Website]

Authors

Xiao Li1,*, Xiang Zheng3,*, Yifeng Gao1, Xinyu Xia4, Yixu Wang1, Xin Wang1, Ye Sun1, Yunhan Zhao1, Ming Wen1,2, Jiayu Li1, Xun Gong4, Yi Liu3, Yige Li5, Yutao Wu6, Cong Wang3, Jun Sun5, Yixin Cao1,2, Zhineng Chen1, Jingjing Chen1, Tao Gui1,2, Qi Zhang1, Zuxuan Wu1,2, Xipeng Qiu1,2, Xuanjing Huang1, Tiehua Zhang7, Zhipeng Wei9, Hanxun Huang10, Sarah Erfani10, James Bailey10, Jianping Wang3, Wei-Ying Ma3,11, Bo Li8, Xingjun Ma1,2,†, Yu-Gang Jiang1,†

1Fudan University, 2Shanghai Innovation Institute, 3City University of Hong Kong, 4Jilin University, 5Singapore Management University, 6Deakin University, 7Tongji University, 8UIUC, 9UC Berkeley, 10The University of Melbourne, 11Tsinghua University

*Equal Contribution, Corresponding Authors

🔥 News

  • [2026/04/01] Beautified paper list with layer icons and visual separators.
  • [2026/03/31] Added llms.txt and SEO meta tags for AI discoverability.
  • [2026/03/28] Added 11 missing safety papers; unified paper counts to 400+.
  • [2026/03/27] Repository and paper released!
  • [2026/03/27] Launched project website with GitHub Pages.
  • [2026/03/27] Added automated paper review GitHub Action for community contributions.
  • [2026/03/26] ISC-Bench paper on arXiv -- 400+ stars in 48 hours!
  • [2026/03/22] ISC-Bench repository released -- Internal Safety Collapse benchmark for frontier LLMs.
  • [2025/09/15] Safety at Scale survey published in Foundations and Trends in Security.
  • [2025/02/02] Safety at Scale survey on arXiv -- large model & agent safety.

Table of Contents

Overview

Embodied AI integrates perception, cognition, planning, and interaction into agents that operate in open-world, safety-critical environments. As these systems gain autonomy and enter domains such as autonomous driving, healthcare, and robotics, ensuring their safety becomes both technically challenging and socially indispensable.

Capability-Risk Duality: each layer of the embodied pipeline represents a capability expansion that introduces corresponding new vulnerabilities.

Capability vs. risk duality in embodied AI systems. As capabilities expand outward from perception to agentic systems, the attack surface grows correspondingly -- vulnerabilities at inner layers cascade to outer layers.

Illustration of safety threats and attack surfaces across capability layers of embodied AI systems.

Overview of representative attack and defense methods across perception, cognition, planning, action & interaction, and agentic system layers. The width of the strips is proportional to the number of reviewed works.

Surveyed Papers

We review 400+ papers across five capability layers of embodied AI.

Layer Subcategories Papers
👁️ Perception Visual, Auditory, Spatial, Motion, Cross-Modal 191
🧠 Cognition Instruction Understanding, World Model, Reasoning 32
🗺️ Planning Task, Trajectory, Multi-Agent 56
🤖 Action and Interaction Robot Control, Human-Agent, Multi-Agent Collaboration 97
Agentic Tool Use, Memory, Self-Evolving, Cascading Risks 76
👁️ Perception (191 papers)
Visual Perception (55)
Auditory Perception (21)
Spatial Perception (59)
Motion Perception (48)
Cross-Modal Perception (8)

🧠 Cognition (32 papers)
Instruction Understanding (12)
World Model (10)
Reasoning (10)

🗺️ Planning (56 papers)
Task Planning (19)
Trajectory Planning (24)
Multi-Agent Planning (13)

🤖 Action and Interaction (97 papers)
Robot Control (82)
Human-Agent Interaction (12)
Multi-Agent Collaboration (3)

Agentic (76 papers)
Tool Use (9)
Memory (15)
Self-Evolving (16)
Cascading Risks (36)

Open Challenges

  • Multimodal Perception Fusion Fragility: Cross-modal attacks exploiting inconsistencies between visual, auditory, and spatial perception remain underexplored.
  • Planning Under Jailbreak: LLM-based planners are vulnerable to instruction manipulation that bypasses safety constraints in physical execution.
  • Human-Agent Interaction Trust: Open-ended scenarios where agents must negotiate trust with humans lack standardized safety evaluation.
  • Agentic Cascading Failures: Self-evolving agents with persistent memory and tool use can propagate inner-layer compromises to system-wide failures.
  • Benchmark Standardization: Lack of unified safety benchmarks across the full embodied AI pipeline hinders reproducible evaluation.

Contributing

Contributions are welcome and encouraged! If you find relevant papers missing from our list or spot any errors:

  • Add a paper: Open a pull request with the paper title, authors, venue, year, and a Google Scholar link.
  • Report an issue: Open an issue describing what needs to be corrected or added.
  • Suggest a category: If a paper does not fit existing subcategories, propose a new one in your PR description.

Please follow the existing format: [Paper Title](Google Scholar link). Authors. *Venue*, Year.

Citation

If you find this survey useful, please cite our paper:

@article{li2026embodied,
  title={Safety in Embodied AI: A Survey of Risks, Attacks, and Defenses},
  author={Li, Xiao and Zheng, Xiang and Gao, Yifeng and Xia, Xinyu and Wang, Yixu and Wang, Xin and Sun, Ye and Zhao, Yunhan and Wen, Ming and Li, Jiayu and Gong, Xun and Liu, Yi and Li, Yige and Wu, Yutao and Wang, Cong and Sun, Jun and Cao, Yixin and Chen, Zhineng and Chen, Jingjing and Gui, Tao and Zhang, Qi and Wu, Zuxuan and Qiu, Xipeng and Huang, Xuanjing and Zhang, Tiehua and Wei, Zhipeng and Huang, Hanxun and Erfani, Sarah and Bailey, James and Ma, Wei-Ying and Wang, Jianping and Li, Bo and Ma, Xingjun and Jiang, Yu-Gang},
  year={2026}
}

Related Projects

From the same team:

  • ISC-Bench -- Internal Safety Collapse in Frontier LLMs (400+ stars)
  • Awesome-Large-Model-Safety -- Safety at Scale: A Comprehensive Survey of Large Model and Agent Safety
  • BackdoorLLM -- A Comprehensive Benchmark for Backdoor Attacks on LLMs (NeurIPS 2025)
  • BackdoorAgent -- Backdoor Attacks on LLM-based Agent Workflows
  • JustAsk -- Curious Code Agents Reveal System Prompts in Frontier LLMs
  • Unlearnable-Examples -- Making Personal Data Unexploitable (ICLR 2021)
  • XTransferBench -- Super Transferable Adversarial Attacks on CLIP (ICML 2025)

Star History

Star History Chart

About

Safety in Embodied AI: A Survey of Risks, Attacks, and Defenses | 400+ Papers | Perception, Cognition, Planning, Interaction, Agentic System

Topics

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Contributors