RACONTEUR: A Knowledgeable, Insightful, and Portable LLM-Powered Shell Command Explainer Authors: Jiangyi Deng, Xinfeng Li, Yanjiao Chen, Yijie Bai, Haiqin Weng, Yan Liu, Tao Wei, Wenyuan Xu | Published: 2024-09-03 LLM Performance EvaluationCybersecurityPrompt Injection 2024.09.03 2025.05.27 Literature Database
Membership Inference Attacks Against In-Context Learning Authors: Rui Wen, Zheng Li, Michael Backes, Yang Zhang | Published: 2024-09-02 Prompt InjectionMembership InferenceAttack Method 2024.09.02 2025.05.27 Literature Database
Unveiling the Vulnerability of Private Fine-Tuning in Split-Based Frameworks for Large Language Models: A Bidirectionally Enhanced Attack Authors: Guanzhong Chen, Zhenghan Qin, Mingxin Yang, Yajie Zhou, Tao Fan, Tianyu Du, Zenglin Xu | Published: 2024-09-02 | Updated: 2024-09-04 LLM SecurityPrompt InjectionAttack Method 2024.09.02 2025.05.27 Literature Database
ProphetFuzz: Fully Automated Prediction and Fuzzing of High-Risk Option Combinations with Only Documentation via Large Language Model Authors: Dawei Wang, Geng Zhou, Li Chen, Dan Li, Yukai Miao | Published: 2024-09-02 Option-Based FuzzingPrompt InjectionVulnerability Management 2024.09.02 2025.05.27 Literature Database
The Dark Side of Human Feedback: Poisoning Large Language Models via User Inputs Authors: Bocheng Chen, Hanqing Guo, Guangjing Wang, Yuanda Wang, Qiben Yan | Published: 2024-09-01 LLM Performance EvaluationPrompt InjectionPoisoning 2024.09.01 2025.05.27 Literature Database
Automatic Pseudo-Harmful Prompt Generation for Evaluating False Refusals in Large Language Models Authors: Bang An, Sicheng Zhu, Ruiyi Zhang, Michael-Andrei Panaitescu-Liess, Yuancheng Xu, Furong Huang | Published: 2024-09-01 LLM Performance EvaluationContent ModerationPrompt Injection 2024.09.01 2025.05.27 Literature Database
LLM Defenses Are Not Robust to Multi-Turn Human Jailbreaks Yet Authors: Nathaniel Li, Ziwen Han, Ian Steneker, Willow Primack, Riley Goodside, Hugh Zhang, Zifan Wang, Cristina Menghini, Summer Yue | Published: 2024-08-27 | Updated: 2024-09-04 Prompt InjectionUser EducationAttack Method 2024.08.27 2025.05.27 Literature Database
Is Generative AI the Next Tactical Cyber Weapon For Threat Actors? Unforeseen Implications of AI Generated Cyber Attacks Authors: Yusuf Usman, Aadesh Upadhyay, Prashnna Gyawali, Robin Chataut | Published: 2024-08-23 CybersecurityPrompt InjectionAttack Method 2024.08.23 2025.05.27 Literature Database
LLM-PBE: Assessing Data Privacy in Large Language Models Authors: Qinbin Li, Junyuan Hong, Chulin Xie, Jeffrey Tan, Rachel Xin, Junyi Hou, Xavier Yin, Zhun Wang, Dan Hendrycks, Zhangyang Wang, Bo Li, Bingsheng He, Dawn Song | Published: 2024-08-23 | Updated: 2024-09-06 LLM SecurityPrivacy Protection MethodPrompt Injection 2024.08.23 2025.05.27 Literature Database
Efficient Detection of Toxic Prompts in Large Language Models Authors: Yi Liu, Junzhe Yu, Huijia Sun, Ling Shi, Gelei Deng, Yuqi Chen, Yang Liu | Published: 2024-08-21 | Updated: 2024-09-14 Content ModerationPrompt InjectionModel Performance Evaluation 2024.08.21 2025.05.27 Literature Database