← Back to Library

Understanding Adversarial Transfer: Why Representation-Space Attacks Fail Where Data-Space Attacks Succeed

Authors: Isha Gupta, Rylan Schaeffer, Joshua Kazdan, Ken Ziyu Liu, Sanmi Koyejo

Published: 2025-10-01

arXiv ID: 2510.01494v2

Added to Library: 2025-10-06 01:01 UTC

Red Teaming

📄 Abstract

The field of adversarial robustness has long established that adversarial examples can successfully transfer between image classifiers and that text jailbreaks can successfully transfer between language models (LMs). However, a pair of recent studies reported being unable to successfully transfer image jailbreaks between vision-language models (VLMs). To explain this striking difference, we propose a fundamental distinction regarding the transferability of attacks against machine learning models: attacks in the input data-space can transfer, whereas attacks in model representation space do not, at least not without geometric alignment of representations. We then provide theoretical and empirical evidence of this hypothesis in four different settings. First, we mathematically prove this distinction in a simple setting where two networks compute the same input-output map but via different representations. Second, we construct representation-space attacks against image classifiers that are as successful as well-known data-space attacks, but fail to transfer. Third, we construct representation-space attacks against LMs that successfully jailbreak the attacked models but again fail to transfer. Fourth, we construct data-space attacks against VLMs that successfully transfer to new VLMs, and we show that representation space attacks can transfer when VLMs' latent geometries are sufficiently aligned in post-projector space. Our work reveals that adversarial transfer is not an inherent property of all attacks but contingent on their operational domain - the shared data-space versus models' unique representation spaces - a critical insight for building more robust models.

🔍 Key Points

  • Establishment of a fundamental distinction between data-space and representation-space attacks, where the former transfers effectively between models while the latter does not without additional geometric alignment.
  • Mathematical proof provided showing that data-space attacks always transfer perfectly, while representation-space attacks require strict geometric alignment of representations to transfer successfully.
  • Demonstration of representation-space attacks on image classifiers and language models that were effective but did not transfer, highlighting the nature of model representation sensitivity.
  • Empirical results showing that data-space attacks can successfully transfer among vision-language models (VLMs), contrasting with representation-space attacks which fail to transfer unless models share aligned geometries in latent space.
  • Recognition of geometric alignment as a crucial property for representation-space attack transfer, leading to insights on improving adversarial robustness in multimodal models.

💡 Why This Paper Matters

This paper presents a critical advancement in understanding adversarial transfers in machine learning models, particularly distinguishing between data-space and representation-space attacks. Its implications for model security and robustness against adversarial examples are significant, especially for developers and researchers striving to enhance the safety and reliability of AI systems.

🎯 Why It's Interesting for AI Security Researchers

The findings of this paper are of paramount interest to AI security researchers, specifically those focused on enhancing the robustness of machine learning models. Understanding the nuances of adversarial attack transferability directly relates to the development of more secure AI systems, as it informs strategies for defending against potentially catastrophic exploitations of vulnerabilities in both image and language-based AI models.

📚 Read the Full Paper