← Back to Library

PromptCARE: Prompt Copyright Protection by Watermark Injection and Verification

Authors: Hongwei Yao, Jian Lou, Kui Ren, Zhan Qin

Published: 2023-08-05

arXiv ID: 2308.02816v2

Added to Library: 2025-11-11 14:12 UTC

📄 Abstract

Large language models (LLMs) have witnessed a meteoric rise in popularity among the general public users over the past few months, facilitating diverse downstream tasks with human-level accuracy and proficiency. Prompts play an essential role in this success, which efficiently adapt pre-trained LLMs to task-specific applications by simply prepending a sequence of tokens to the query texts. However, designing and selecting an optimal prompt can be both expensive and demanding, leading to the emergence of Prompt-as-a-Service providers who profit by providing well-designed prompts for authorized use. With the growing popularity of prompts and their indispensable role in LLM-based services, there is an urgent need to protect the copyright of prompts against unauthorized use. In this paper, we propose PromptCARE, the first framework for prompt copyright protection through watermark injection and verification. Prompt watermarking presents unique challenges that render existing watermarking techniques developed for model and dataset copyright verification ineffective. PromptCARE overcomes these hurdles by proposing watermark injection and verification schemes tailor-made for prompts and NLP characteristics. Extensive experiments on six well-known benchmark datasets, using three prevalent pre-trained LLMs (BERT, RoBERTa, and Facebook OPT-1.3b), demonstrate the effectiveness, harmlessness, robustness, and stealthiness of PromptCARE.

🔍 Key Points

  • Demonstrates the vulnerability of ChatGPT and similar LLMs to lightweight prompt injection attacks using a structured framework.
  • Presents three real-world injection methods: direct user prompts, search-based context integration, and system-level instructions in GPT agents.
  • Highlights specific examples where injected prompts bias outputs in high-stakes contexts like product recommendations and academic peer reviews.
  • Calls for prioritization of prompt-level security in LLM design and deployment to mitigate security risks associated with adversarial prompting.

💡 Why This Paper Matters

This paper is relevant and important as it exposes a critical vulnerability within widely-used LLMs like ChatGPT, emphasizing that current security measures are inadequate to prevent sophisticated prompt injection attacks. It raises awareness among developers and researchers about the necessity of improving model robustness against manipulation.

🎯 Why It's Interesting for AI Security Researchers

This paper is of particular interest to AI security researchers as it provides insight into novel attack vectors that can be employed against LLMs. It not only outlines the risks associated with prompt injections in real-world applications but also encourages a dialogue on enhancing security protocols and designing safer AI systems that can withstand such manipulations.

📚 Read the Full Paper