← Back to Library

Prompt Injection Attacks on Agentic Coding Assistants: A Systematic Analysis of Vulnerabilities in Skills, Tools, and Protocol Ecosystems

Authors: Narek Maloyan, Dmitry Namiot

Published: 2026-01-24

arXiv ID: 2601.17548v1

Added to Library: 2026-01-27 03:01 UTC

Red Teaming

πŸ“„ Abstract

The proliferation of agentic AI coding assistants, including Claude Code, GitHub Copilot, Cursor, and emerging skill-based architectures, has fundamentally transformed software development workflows. These systems leverage Large Language Models (LLMs) integrated with external tools, file systems, and shell access through protocols like the Model Context Protocol (MCP). However, this expanded capability surface introduces critical security vulnerabilities. In this \textbf{Systematization of Knowledge (SoK)} paper, we present a comprehensive analysis of prompt injection attacks targeting agentic coding assistants. We propose a novel three-dimensional taxonomy categorizing attacks across \textit{delivery vectors}, \textit{attack modalities}, and \textit{propagation behaviors}. Our meta-analysis synthesizes findings from 78 recent studies (2021--2026), consolidating evidence that attack success rates against state-of-the-art defenses exceed 85\% when adaptive attack strategies are employed. We systematically catalog 42 distinct attack techniques spanning input manipulation, tool poisoning, protocol exploitation, multimodal injection, and cross-origin context poisoning. Through critical analysis of 18 defense mechanisms reported in prior work, we identify that most achieve less than 50\% mitigation against sophisticated adaptive attacks. We contribute: (1) a unified taxonomy bridging disparate attack classifications, (2) the first systematic analysis of skill-based architecture vulnerabilities with concrete exploit chains, and (3) a defense-in-depth framework grounded in the limitations we identify. Our findings indicate that the security community must treat prompt injection as a first-class vulnerability class requiring architectural-level mitigations rather than ad-hoc filtering approaches.

πŸ” Key Points

  • Proposes a novel three-dimensional taxonomy for categorizing prompt injection attacks targeting coding assistants, enhancing the understanding and classification of these vulnerabilities.
  • Systematically catalogs 42 distinct attack techniques with detailed examples that demonstrate the practical risks of prompt injections in AI coding assistants.
  • Conducts a meta-analysis of 78 studies, revealing that adaptive attack strategies against existing defenses yield success rates exceeding 85%, signaling critical effectiveness gaps in current security measures.
  • Critically assesses 18 defense mechanisms, indicating that most provide less than 50% mitigation against sophisticated attacks and highlights the need for more robust defense frameworks.
  • Introduces a defense-in-depth framework that promotes architectural-level mitigations rather than relying solely on ad-hoc filtering solutions.

πŸ’‘ Why This Paper Matters

This paper is highly relevant as it addresses the growing security risks posed by prompt injection attacks on AI coding assistantsβ€”a significant concern as these tools gain autonomy in software development. By providing comprehensive insights, a novel taxonomy, and an analysis of vulnerabilities and defenses, it establishes a foundational resource for understanding and combating these threats, making substantial contributions to the field of AI security.

🎯 Why It's Interesting for AI Security Researchers

AI security researchers would find this paper invaluable as it not only identifies critical vulnerabilities within widely-used AI coding assistants but also offers a systematic classification of attacks. The empirical evidence presented regarding attack success rates against defenses serves as a call to action for the community to innovate and implement architectural solutions. Additionally, the discussion of novel defense frameworks underscores the urgency for more effective countermeasures, ensuring the safety and integrity of software development processes utilizing these advanced tools.

πŸ“š Read the Full Paper