← Back to Library

OMNI-LEAK: Orchestrator Multi-Agent Network Induced Data Leakage

Authors: Akshat Naik, Jay Culligan, Yarin Gal, Philip Torr, Rahaf Aljundi, Alasdair Paren, Adel Bibi

Published: 2026-02-13

arXiv ID: 2602.13477v1

Added to Library: 2026-02-17 03:01 UTC

Red Teaming

📄 Abstract

As Large Language Model (LLM) agents become more capable, their coordinated use in the form of multi-agent systems is anticipated to emerge as a practical paradigm. Prior work has examined the safety and misuse risks associated with agents. However, much of this has focused on the single-agent case and/or setups missing basic engineering safeguards such as access control, revealing a scarcity of threat modeling in multi-agent systems. We investigate the security vulnerabilities of a popular multi-agent pattern known as the orchestrator setup, in which a central agent decomposes and delegates tasks to specialized agents. Through red-teaming a concrete setup representative of a likely future use case, we demonstrate a novel attack vector, OMNI-LEAK, that compromises several agents to leak sensitive data through a single indirect prompt injection, even in the \textit{presence of data access control}. We report the susceptibility of frontier models to different categories of attacks, finding that both reasoning and non-reasoning models are vulnerable, even when the attacker lacks insider knowledge of the implementation details. Our work highlights the importance of safety research to generalize from single-agent to multi-agent settings, in order to reduce the serious risks of real-world privacy breaches and financial losses and overall public trust in AI agents.

🔍 Key Points

  • Introduction of the OMNI-LEAK attack vector that compromises multi-agent systems by leaking sensitive data through indirect prompt injections, even with access controls in place.
  • An experimental evaluation of various frontier models reveals that most, except for one (claude-sonnet-4), are susceptible to at least one form of the OMNI-LEAK attack across different setups and database sizes.
  • Development of the first benchmark for assessing the vulnerability of orchestrator multi-agent data leakage, highlighting the serious risks of privacy breaches in practical applications.

💡 Why This Paper Matters

This paper underscores the critical security vulnerabilities inherent in orchestrator multi-agent systems leveraging large language models. By presenting the OMNI-LEAK attack and demonstrating its effectiveness against various models, the authors illustrate the need for comprehensive threat modeling and mitigation strategies as these systems become increasingly integrated into real-world applications.

🎯 Why It's Interesting for AI Security Researchers

The findings of this paper are pivotal for AI security researchers as they highlight a previously underexplored area of risk associated with multi-agent systems. This work not only reveals significant vulnerabilities in popular models but also serves as a foundation for devising enhanced safety measures and threat modeling frameworks for future AI deployments.

📚 Read the Full Paper