Prompt leaking

The Early Bird Catches the Leak: Unveiling Timing Side Channels in LLM Serving Systems

Authors: Linke Song, Zixuan Pang, Wenhao Wang, Zihao Wang, XiaoFeng Wang, Hongbo Chen, Wei Song, Yier Jin, Dan Meng, Rui Hou | Published: 2024-09-30 | Updated: 2025-08-13
Security Assurance
Prompt leaking
攻撃戦略分析

Confidential Prompting: Privacy-preserving LLM Inference on Cloud

Authors: Caihua Li, In Gim, Lin Zhong | Published: 2024-09-27 | Updated: 2025-08-25
Process Partitioning Method
Prompt leaking
Model Extraction Attack

Why Are My Prompts Leaked? Unraveling Prompt Extraction Threats in Customized Large Language Models

Authors: Zi Liang, Haibo Hu, Qingqing Ye, Yaxin Xiao, Haoyang Li | Published: 2024-08-05 | Updated: 2025-02-12
Prompt Injection
Prompt leaking
Model Evaluation

From Sands to Mansions: Towards Automated Cyberattack Emulation with Classical Planning and Large Language Models

Authors: Lingzhi Wang, Zhenyuan Li, Yi Jiang, Zhengkai Wang, Zonghan Guo, Jiahui Wang, Yangyang Wei, Xiangmin Shen, Wei Ruan, Yan Chen | Published: 2024-07-24 | Updated: 2025-04-17
Prompt leaking
Attack Action Model
Attack Detection Method

ProxyGPT: Enabling User Anonymity in LLM Chatbots via (Un)Trustworthy Volunteer Proxies

Authors: Dzung Pham, Jade Sheffey, Chau Minh Pham, Amir Houmansadr | Published: 2024-07-11 | Updated: 2025-06-11
Privacy Enhancing Technology
Prompt Injection
Prompt leaking

Human-Imperceptible Retrieval Poisoning Attacks in LLM-Powered Applications

Authors: Quan Zhang, Binqi Zeng, Chijin Zhou, Gwihwan Go, Heyuan Shi, Yu Jiang | Published: 2024-04-26
Poisoning attack on RAG
Prompt leaking
Poisoning

Stealing Part of a Production Language Model

Authors: Nicholas Carlini, Daniel Paleka, Krishnamurthy Dj Dvijotham, Thomas Steinke, Jonathan Hayase, A. Feder Cooper, Katherine Lee, Matthew Jagielski, Milad Nasr, Arthur Conmy, Itay Yona, Eric Wallace, David Rolnick, Florian Tramèr | Published: 2024-03-11 | Updated: 2024-07-09
Prompt leaking
Model Robustness
Model Extraction Attack

Secret Collusion among Generative AI Agents: Multi-Agent Deception via Steganography

Authors: Sumeet Ramesh Motwani, Mikhail Baranchuk, Martin Strohmeier, Vijay Bolina, Philip H. S. Torr, Lewis Hammond, Christian Schroeder de Witt | Published: 2024-02-12 | Updated: 2025-04-14
Privacy Enhancing Technology
Prompt leaking
Digital Watermarking for Generative AI

Language Model Inversion

Authors: John X. Morris, Wenting Zhao, Justin T. Chiu, Vitaly Shmatikov, Alexander M. Rush | Published: 2023-11-22
Prompt leaking
Model Inversion
Model Evaluation

Assessing Prompt Injection Risks in 200+ Custom GPTs

Authors: Jiahao Yu, Yuhang Wu, Dong Shu, Mingyu Jin, Sabrina Yang, Xinyu Xing | Published: 2023-11-20 | Updated: 2024-05-25
Prompt Injection
Prompt leaking
Dialogue System