Backdoor Attack

TextGuard: Provable Defense against Backdoor Attacks on Text Classification

Authors: Hengzhi Pei, Jinyuan Jia, Wenbo Guo, Bo Li, Dawn Song | Published: 2023-11-19 | Updated: 2023-11-25
Text Generation Method
Backdoor Attack
Poisoning

Stealthy and Persistent Unalignment on Large Language Models via Backdoor Injections

Authors: Yuanpu Cao, Bochuan Cao, Jinghui Chen | Published: 2023-11-15 | Updated: 2024-06-09
Backdoor Attack
Prompt Injection

Label Poisoning is All You Need

Authors: Rishi D. Jha, Jonathan Hayase, Sewoong Oh | Published: 2023-10-29
Security Analysis
Backdoor Attack
Classification of Malicious Actors

On the Detection of Image-Scaling Attacks in Machine Learning

Authors: Erwin Quiring, Andreas Müller, Konrad Rieck | Published: 2023-10-23
Backdoor Attack
Adversarial Attack Detection
Analysis of Detection Methods

FLEDGE: Ledger-based Federated Learning Resilient to Inference and Backdoor Attacks

Authors: Jorge Castillo, Phillip Rieger, Hossein Fereidooni, Qian Chen, Ahmad Sadeghi | Published: 2023-10-03
Backdoor Attack
Privacy Protection
Poisoning

Protect Federated Learning Against Backdoor Attacks via Data-Free Trigger Generation

Authors: Yanxin Yang, Ming Hu, Yue Cao, Jun Xia, Yihao Huang, Yang Liu, Mingsong Chen | Published: 2023-08-22
Backdoor Attack
Poisoning
Defense Method

SoK: Realistic Adversarial Attacks and Defenses for Intelligent Network Intrusion Detection

Authors: João Vitorino, Isabel Praça, Eva Maia | Published: 2023-08-13
Backdoor Attack
Adversarial Training
Defense Method

Breaking Speaker Recognition with PaddingBack

Authors: Zhe Ye, Diqun Yan, Li Dong, Kailai Shen | Published: 2023-08-08 | Updated: 2024-03-11
Watermarking
Backdoor Attack
Poisoning

Backdoor Federated Learning by Poisoning Backdoor-Critical Layers

Authors: Haomin Zhuang, Mingxian Yu, Hao Wang, Yang Hua, Jian Li, Xu Yuan | Published: 2023-08-08 | Updated: 2024-04-15
Backdoor Attack
Poisoning Attack
Defense Method

Backdoor Attacks for In-Context Learning with Language Models

Authors: Nikhil Kandpal, Matthew Jagielski, Florian Tramèr, Nicholas Carlini | Published: 2023-07-27
LLM Security
Backdoor Attack
Prompt Injection