Backdoor Attack

Watch Out for Your Guidance on Generation! Exploring Conditional Backdoor Attacks against Large Language Models

Authors: Jiaming He, Wenbo Jiang, Guanyu Hou, Wenshu Fan, Rui Zhang, Hongwei Li | Published: 2024-04-23 | Updated: 2025-01-08
LLM Security
Backdoor Attack
Poisoning

Competition Report: Finding Universal Jailbreak Backdoors in Aligned LLMs

Authors: Javier Rando, Francesco Croce, Kryštof Mitka, Stepan Shabalin, Maksym Andriushchenko, Nicolas Flammarion, Florian Tramèr | Published: 2024-04-22 | Updated: 2024-06-06
LLM Security
Backdoor Attack
Prompt Injection

Physical Backdoor Attack can Jeopardize Driving with Vision-Large-Language Models

Authors: Zhenyang Ni, Rui Ye, Yuxi Wei, Zhen Xiang, Yanfeng Wang, Siheng Chen | Published: 2024-04-19 | Updated: 2024-04-22
Backdoor Attack
Vulnerabilities in Autonomous Driving Technology

Exploring Backdoor Vulnerabilities of Chat Models

Authors: Yunzhuo Hao, Wenkai Yang, Yankai Lin | Published: 2024-04-03
Backdoor Attack
Prompt Injection

Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models

Authors: Yuxin Wen, Leo Marchyok, Sanghyun Hong, Jonas Geiping, Tom Goldstein, Nicholas Carlini | Published: 2024-04-01
Backdoor Attack
Poisoning
Membership Inference

A Backdoor Approach with Inverted Labels Using Dirty Label-Flipping Attacks

Authors: Orson Mengara | Published: 2024-03-29 | Updated: 2024-04-07
Dataset Generation
Backdoor Attack
Poisoning

Backdoor Secrets Unveiled: Identifying Backdoor Data with Optimized Scaled Prediction Consistency

Authors: Soumyadeep Pal, Yuguang Yao, Ren Wang, Bingquan Shen, Sijia Liu | Published: 2024-03-15
Watermarking
Backdoor Attack
Poisoning

Interactive Trimming against Evasive Online Data Manipulation Attacks: A Game-Theoretic Approach

Authors: Yue Fu, Qingqing Ye, Rong Du, Haibo Hu | Published: 2024-03-15
Backdoor Attack
Privacy Protection Method
Strategic Game

Machine Unlearning: Taxonomy, Metrics, Applications, Challenges, and Prospects

Authors: Na Li, Chunyi Zhou, Yansong Gao, Hui Chen, Anmin Fu, Zhi Zhang, Yu Shui | Published: 2024-03-13
Backdoor Attack
Membership Inference
Machine Unlearning

Poisoning Programs by Un-Repairing Code: Security Concerns of AI-generated Code

Authors: Cristina Improta | Published: 2024-03-11
Security Analysis
Backdoor Attack
Poisoning