Prompt leaking

The Early Bird Catches the Leak: Unveiling Timing Side Channels in LLM Serving Systems

Authors: Linke Song, Zixuan Pang, Wenhao Wang, Zihao Wang, XiaoFeng Wang, Hongbo Chen, Wei Song, Yier Jin, Dan Meng, Rui Hou | Published: 2024-09-30 | Updated: 2025-08-13
Security Assurance
Prompt leaking
攻撃戦略分析

Confidential Prompting: Privacy-preserving LLM Inference on Cloud

Authors: Caihua Li, In Gim, Lin Zhong | Published: 2024-09-27 | Updated: 2025-08-25
Process Partitioning Method
Prompt leaking
Model Extraction Attack

Evading Toxicity Detection with ASCII-art: A Benchmark of Spatial Attacks on Moderation Systems

Authors: Sergey Berezin, Reza Farahbakhsh, Noel Crespi | Published: 2024-09-27 | Updated: 2025-09-24
Token Compression Framework
Prompt leaking
Natural Language Processing

Why Are My Prompts Leaked? Unraveling Prompt Extraction Threats in Customized Large Language Models

Authors: Zi Liang, Haibo Hu, Qingqing Ye, Yaxin Xiao, Haoyang Li | Published: 2024-08-05 | Updated: 2025-02-12
Prompt Injection
Prompt leaking
Model Evaluation

From Sands to Mansions: Towards Automated Cyberattack Emulation with Classical Planning and Large Language Models

Authors: Lingzhi Wang, Zhenyuan Li, Yi Jiang, Zhengkai Wang, Zonghan Guo, Jiahui Wang, Yangyang Wei, Xiangmin Shen, Wei Ruan, Yan Chen | Published: 2024-07-24 | Updated: 2025-04-17
Prompt leaking
Attack Action Model
Attack Detection Method

ProxyGPT: Enabling User Anonymity in LLM Chatbots via (Un)Trustworthy Volunteer Proxies

Authors: Dzung Pham, Jade Sheffey, Chau Minh Pham, Amir Houmansadr | Published: 2024-07-11 | Updated: 2025-06-11
Privacy Enhancing Technology
Prompt Injection
Prompt leaking

Chain-of-Scrutiny: Detecting Backdoor Attacks for Large Language Models

Authors: Xi Li, Ruofan Mao, Yusen Zhang, Renze Lou, Chen Wu, Jiaqi Wang | Published: 2024-06-10 | Updated: 2025-10-30
Indirect Prompt Injection
Trigger Detection
Prompt leaking

Human-Imperceptible Retrieval Poisoning Attacks in LLM-Powered Applications

Authors: Quan Zhang, Binqi Zeng, Chijin Zhou, Gwihwan Go, Heyuan Shi, Yu Jiang | Published: 2024-04-26
Poisoning attack on RAG
Prompt leaking
Poisoning

Stealing Part of a Production Language Model

Authors: Nicholas Carlini, Daniel Paleka, Krishnamurthy Dj Dvijotham, Thomas Steinke, Jonathan Hayase, A. Feder Cooper, Katherine Lee, Matthew Jagielski, Milad Nasr, Arthur Conmy, Itay Yona, Eric Wallace, David Rolnick, Florian Tramèr | Published: 2024-03-11 | Updated: 2024-07-09
Prompt leaking
Model Robustness
Model Extraction Attack

Secret Collusion among Generative AI Agents: Multi-Agent Deception via Steganography

Authors: Sumeet Ramesh Motwani, Mikhail Baranchuk, Martin Strohmeier, Vijay Bolina, Philip H. S. Torr, Lewis Hammond, Christian Schroeder de Witt | Published: 2024-02-12 | Updated: 2025-04-14
Privacy Enhancing Technology
Prompt leaking
Digital Watermarking for Generative AI