AI Agents 相关度: 9/10

Agent2Agent Threats in Safety-Critical LLM Assistants: A Human-Centric Taxonomy

Lukas Stappen, Ahmet Erkan Turan, Johann Hagerer, Georg Groh
arXiv: 2602.05877v1 发布: 2026-02-05 更新: 2026-02-05

AI 摘要

提出AgentHeLLM框架,针对LLM智能助手在车辆环境中Agent间通信的安全威胁进行建模和分析。

主要贡献

  • 提出AgentHeLLM威胁建模框架,分离资产识别和攻击路径分析。
  • 构建基于人权视角的资产分类体系。
  • 开发AgentHeLLM Attack Path Generator工具,自动化多阶段威胁发现。

方法论

提出一种基于图的正式模型,区分毒化路径(恶意数据传播)和触发路径(激活动作)。

原文摘要

The integration of Large Language Model (LLM)-based conversational agents into vehicles creates novel security challenges at the intersection of agentic AI, automotive safety, and inter-agent communication. As these intelligent assistants coordinate with external services via protocols such as Google's Agent-to-Agent (A2A), they establish attack surfaces where manipulations can propagate through natural language payloads, potentially causing severe consequences ranging from driver distraction to unauthorized vehicle control. Existing AI security frameworks, while foundational, lack the rigorous "separation of concerns" standard in safety-critical systems engineering by co-mingling the concepts of what is being protected (assets) with how it is attacked (attack paths). This paper addresses this methodological gap by proposing a threat modeling framework called AgentHeLLM (Agent Hazard Exploration for LLM Assistants) that formally separates asset identification from attack path analysis. We introduce a human-centric asset taxonomy derived from harm-oriented "victim modeling" and inspired by the Universal Declaration of Human Rights, and a formal graph-based model that distinguishes poison paths (malicious data propagation) from trigger paths (activation actions). We demonstrate the framework's practical applicability through an open-source attack path suggestion tool AgentHeLLM Attack Path Generator that automates multi-stage threat discovery using a bi-level search strategy.

标签

LLM AI Agent Security Threat Modeling Automotive

arXiv 分类

cs.AI cs.HC