Prompt Injection

DrLLM: Prompt-Enhanced Distributed Denial-of-Service Resistance Method with Large Language Models

Authors: Zhenyu Yin, Shang Liu, Guangyuan Xu | Published: 2024-09-11 | Updated: 2025-01-13
DDoS Attack Detection
LLM Performance Evaluation
Prompt Injection

AdaPPA: Adaptive Position Pre-Fill Jailbreak Attack Approach Targeting LLMs

Authors: Lijia Lv, Weigang Zhang, Xuehai Tang, Jie Wen, Feng Liu, Jizhong Han, Songlin Hu | Published: 2024-09-11
LLM Security
Prompt Injection
Attack Method

Exploring User Privacy Awareness on GitHub: An Empirical Study

Authors: Costanza Alfieri, Juri Di Rocco, Paola Inverardi, Phuong T. Nguyen | Published: 2024-09-06 | Updated: 2024-09-10
Privacy Protection
Prompt Injection
User Activity Analysis

RACONTEUR: A Knowledgeable, Insightful, and Portable LLM-Powered Shell Command Explainer

Authors: Jiangyi Deng, Xinfeng Li, Yanjiao Chen, Yijie Bai, Haiqin Weng, Yan Liu, Tao Wei, Wenyuan Xu | Published: 2024-09-03
LLM Performance Evaluation
Cybersecurity
Prompt Injection

Membership Inference Attacks Against In-Context Learning

Authors: Rui Wen, Zheng Li, Michael Backes, Yang Zhang | Published: 2024-09-02
Prompt Injection
Membership Inference
Attack Method

Unveiling the Vulnerability of Private Fine-Tuning in Split-Based Frameworks for Large Language Models: A Bidirectionally Enhanced Attack

Authors: Guanzhong Chen, Zhenghan Qin, Mingxin Yang, Yajie Zhou, Tao Fan, Tianyu Du, Zenglin Xu | Published: 2024-09-02 | Updated: 2024-09-04
LLM Security
Prompt Injection
Attack Method

ProphetFuzz: Fully Automated Prediction and Fuzzing of High-Risk Option Combinations with Only Documentation via Large Language Model

Authors: Dawei Wang, Geng Zhou, Li Chen, Dan Li, Yukai Miao | Published: 2024-09-02
Option-Based Fuzzing
Prompt Injection
Vulnerability Management

The Dark Side of Human Feedback: Poisoning Large Language Models via User Inputs

Authors: Bocheng Chen, Hanqing Guo, Guangjing Wang, Yuanda Wang, Qiben Yan | Published: 2024-09-01
LLM Performance Evaluation
Prompt Injection
Poisoning

Automatic Pseudo-Harmful Prompt Generation for Evaluating False Refusals in Large Language Models

Authors: Bang An, Sicheng Zhu, Ruiyi Zhang, Michael-Andrei Panaitescu-Liess, Yuancheng Xu, Furong Huang | Published: 2024-09-01
LLM Performance Evaluation
Content Moderation
Prompt Injection

LLM Defenses Are Not Robust to Multi-Turn Human Jailbreaks Yet

Authors: Nathaniel Li, Ziwen Han, Ian Steneker, Willow Primack, Riley Goodside, Hugh Zhang, Zifan Wang, Cristina Menghini, Summer Yue | Published: 2024-08-27 | Updated: 2024-09-04
Prompt Injection
User Education
Attack Method