Agentic AI and Multi-Agent Systems are poised to dominate industry and society imminently. Powered by goal-driven autonomy, they represent a powerful form of generative AI, marking a transition from reactive content generation into proactive multitasking capabilities. As an exemplar, we propose an architecture of a multi-agent system for the implementation phase of the software engineering process. We also present a comprehensive threat model for the proposed system. We demonstrate that while such systems can generate code quite accurately, they are vulnerable to attacks, including code injection. Due to their autonomous design and lack of humans in the loop, these systems cannot identify and respond to attacks by themselves. This paper analyzes the vulnerability of multi-agent systems and concludes that the coder-reviewer-tester architecture is more resilient than both the coder and coder-tester architectures, but is less efficient at writing code. We find that by adding a security analysis agent, we mitigate the loss in efficiency while achieving even better resiliency. We conclude by demonstrating that the security analysis agent is vulnerable to advanced code injection attacks, showing that embedding poisonous few-shot examples in the injected code can increase the attack success rate from 0% to 71.95%.
翻译:智能体AI与多智能体系统即将在工业和社会领域占据主导地位。凭借目标驱动的自主性,它们代表了生成式AI的一种强大形式,标志着从被动内容生成向主动多任务处理能力的转变。作为范例,我们提出了一种用于软件工程流程实施阶段的多智能体系统架构。同时,我们为该系统构建了全面的威胁模型。研究表明,虽然此类系统能够相当准确地生成代码,但它们容易受到包括代码注入在内的多种攻击。由于其自主设计特性及缺乏人工介入环节,这些系统无法自主识别并应对攻击。本文分析了多智能体系统的脆弱性,得出结论:编码者-评审者-测试者架构相比编码者架构及编码者-测试者架构具有更强的抗攻击能力,但编码效率较低。研究发现,通过引入安全分析智能体,可以在提升系统抗攻击能力的同时有效缓解效率损失。最后我们证明,安全分析智能体本身仍面临高级代码注入攻击的威胁——通过在注入代码中嵌入恶意少样本示例,攻击成功率可从0%提升至71.95%。