Poster
in
Workshop: Multi-Agent Systems in the Era of Foundation Models: Opportunities, Challenges and Futures
TAMAS: A Dataset for Investigating Security Risks in Multi-Agent LLM Systems
Ishan Kavathekar · Hemang Jain · Ameya Rathod · Ponnurangam Kumaraguru · Tanuja Ganu
Abstract:
Large Language Models (LLMs) have demonstrated strong capabilities as autonomous agents through tool use, planning, and decision-making abilities, leading to their widespread adoption across diverse tasks. As task complexity grows, multi-agent LLM systems are increasingly used to collaboratively solve problems. However, safety and security of these multi-agent systems remains largely unexplored. Existing benchmarks and datasets predominantly focus on single-agent settings, failing to capture the unique vulnerabilities of multi-agent dynamics and co-ordination. To address this gap, we introduce $\textbf{T}$hreats and $\textbf{A}$ttacks in $\textbf{M}$ulti-$\textbf{A}$gent $\textbf{S}$ystems ($\textbf{TAMAS}$), a dataset designed to evaluate the robustness and security of multi-agent LLM systems. TAMAS includes five distinct scenarios comprising 250 adversarial instances across five attack types and 163 different normal and attack tools, along with 100 harmless tasks. We assess system performance across 5 backbone LLMs and 3 agent interaction configurations from Autogen framework, highlighting critical challenges and failure modes in current multi-agent deployments. Our findings show that multi-agent systems are highly vulnerable to adversarial attacks, with Impersonation reaching a 73\% success rate and other attacks ranging from 27\% to 67\%, underscoring the need for stronger defenses. Code and data is available.
Chat is not available.