← Back to Library

AJAR: Adaptive Jailbreak Architecture for Red-teaming

Authors: Yipu Dou, Wang Yang

Published: 2026-01-16

arXiv ID: 2601.10971v1

Added to Library: 2026-01-19 03:01 UTC

Red Teaming

📄 Abstract

As Large Language Models (LLMs) evolve from static chatbots into autonomous agents capable of tool execution, the landscape of AI safety is shifting from content moderation to action security. However, existing red-teaming frameworks remain bifurcated: they either focus on rigid, script-based text attacks or lack the architectural modularity to simulate complex, multi-turn agentic exploitations. In this paper, we introduce AJAR (Adaptive Jailbreak Architecture for Red-teaming), a proof-of-concept framework designed to bridge this gap through Protocol-driven Cognitive Orchestration. Built upon the robust runtime of Petri, AJAR leverages the Model Context Protocol (MCP) to decouple adversarial logic from the execution loop, encapsulating state-of-the-art algorithms like X-Teaming as standardized, plug-and-play services. We validate the architectural feasibility of AJAR through a controlled qualitative case study, demonstrating its ability to perform stateful backtracking within a tool-use environment. Furthermore, our preliminary exploration of the "Agentic Gap" reveals a complex safety dynamic: while tool usage introduces new injection vectors via code execution, the cognitive load of parameter formatting can inadvertently disrupt persona-based attacks. AJAR is open-sourced to facilitate the standardized, environment-aware evaluation of this emerging attack surface. The code and data are available at https://github.com/douyipu/ajar.

🔍 Key Points

  • Introduction of AJAR, a novel framework that separates adversarial logic from execution, allowing for more dynamic and adaptable red-teaming strategies.
  • Utilization of the Model Context Protocol (MCP) to modularize adversarial strategies, facilitating easier integration and flexibility in attack execution.
  • Demonstration of the 'Agentic Gap,' highlighting nuanced safety dynamics when using tools: while tools can be distractions, they also create new vectors for indirect attacks.
  • Validation of AJAR's architecture through controlled qualitative case studies, showing its ability to handle complex, multi-turn jailbreak scenarios effectively.
  • Open-source release of AJAR to encourage further research and development of adaptive red-teaming methodologies.

💡 Why This Paper Matters

This paper presents a crucial advancement in the field of AI safety, providing a structured framework that adapts red-teaming practices to the evolving landscape of autonomous agents. By addressing the limitations of existing frameworks and exploring the interplay between tool usage and safety dynamics, AJAR stands to significantly enhance the robustness of security assessments for large language model-based agents.

🎯 Why It's Interesting for AI Security Researchers

The research presented in this paper is vital for AI security researchers as it offers a new methodology for systematically evaluating the safety of AI agents that are increasingly capable of action execution. The findings regarding the 'Agentic Gap' and the implications of tool usage provide important insights into how adversarial contexts may evolve, making it essential for those focused on the security and ethical deployment of AI technologies.

📚 Read the Full Paper