← Back to Library

Understanding Adversarial Transfer: Why Representation-Space Attacks Fail Where Data-Space Attacks Succeed

Authors: Isha Gupta, Rylan Schaeffer, Joshua Kazdan, Ken Liu, Sanmi Koyejo

Published: 2025-10-01

arXiv ID: 2510.01494v1

Added to Library: 2025-10-03 04:02 UTC

Red Teaming

📄 Abstract

The field of adversarial robustness has long established that adversarial examples can successfully transfer between image classifiers and that text jailbreaks can successfully transfer between language models (LMs). However, a pair of recent studies reported being unable to successfully transfer image jailbreaks between vision-language models (VLMs). To explain this striking difference, we propose a fundamental distinction regarding the transferability of attacks against machine learning models: attacks in the input data-space can transfer, whereas attacks in model representation space do not, at least not without geometric alignment of representations. We then provide theoretical and empirical evidence of this hypothesis in four different settings. First, we mathematically prove this distinction in a simple setting where two networks compute the same input-output map but via different representations. Second, we construct representation-space attacks against image classifiers that are as successful as well-known data-space attacks, but fail to transfer. Third, we construct representation-space attacks against LMs that successfully jailbreak the attacked models but again fail to transfer. Fourth, we construct data-space attacks against VLMs that successfully transfer to new VLMs, and we show that representation space attacks \emph{can} transfer when VLMs' latent geometries are sufficiently aligned in post-projector space. Our work reveals that adversarial transfer is not an inherent property of all attacks but contingent on their operational domain - the shared data-space versus models' unique representation spaces - a critical insight for building more robust models.

🔍 Key Points

  • The paper establishes a critical distinction between data-space and representation-space attacks in the context of adversarial robustness, elucidating why the former are more likely to transfer successfully than the latter.
  • Theoretical and empirical evidence is presented in four scenarios, demonstrating that while data-space attacks can transfer across models, representation-space attacks often require geometric alignment of model representations to succeed in transferring.
  • The authors propose and empirically validate the construction of representation-space attacks that are effective against specific models but fail to transfer when applied to models with differing latent geometries.
  • Data-space attacks are shown to successfully transfer among Vision-Language Models (VLMs), while representation-space attacks do not, reinforcing the significance of understanding model structure in designing adversarial attacks.
  • The findings suggest practical implications for enhancing the robustness of machine learning models by taking into account the nature of their internal representations during attack design.

💡 Why This Paper Matters

This paper is significant for its insightful analysis of adversarial attack transferability, highlighting essential differences between data-space and representation-space attacks. By providing both theoretical groundwork and extensive empirical validation, it sets the stage for future research aimed at improving the robustness of machine learning models against a range of adversarial attacks.

🎯 Why It's Interesting for AI Security Researchers

AI security researchers would find this paper particularly relevant due to its advanced exploration of adversarial transfer dynamics, which is critical for devising effective defense strategies against malicious attacks. The clear delineation between different attack spaces offers a foundational understanding for minimizing vulnerabilities in AI systems, especially in safety-critical applications involving multimodal interactions.

📚 Read the Full Paper