Backdoor Attack

Injection, Attack and Erasure: Revocable Backdoor Attacks via Machine Unlearning

Authors: Baogang Song, Dongdong Zhao, Jianwen Xiang, Qiben Xu, Zizhuo Yu | Published: 2025-10-15
Backdoor Attack
Model Protection Methods
Defense Mechanism

Cryptographic Backdoor for Neural Networks: Boon and Bane

Authors: Anh Tu Ngo, Anupam Chattopadhyay, Subhamoy Maitra | Published: 2025-09-25
Trigger Detection
Backdoor Attack
Watermark Design

Non-omniscient backdoor injection with a single poison sample: Proving the one-poison hypothesis for linear regression and linear classification

Authors: Thorsten Peinemann, Paula Arnold, Sebastian Berndt, Thomas Eisenbarth, Esfandiar Mohammadi | Published: 2025-08-07
Backdoor Attack
Backdoor Attack Techniques
Poisoning

Evasion Attacks Against Bayesian Predictive Models

Authors: Pablo G. Arce, Roi Naveiro, David Ríos Insua | Published: 2025-06-11
Backdoor Attack
Bayesian Adversarial Learning
Adversarial Perturbation Techniques

Backdoor Cleaning without External Guidance in MLLM Fine-tuning

Authors: Xuankun Rong, Wenke Huang, Jian Liang, Jinhe Bi, Xun Xiao, Yiming Li, Bo Du, Mang Ye | Published: 2025-05-22
LLM Security
Backdoor Attack

Finetuning-Activated Backdoors in LLMs

Authors: Thibaud Gloaguen, Mark Vero, Robin Staab, Martin Vechev | Published: 2025-05-22
LLM Security
Backdoor Attack
Prompt Injection

Analysis of the vulnerability of machine learning regression models to adversarial attacks using data from 5G wireless networks

Authors: Leonid Legashev, Artur Zhigalov, Denis Parfenov | Published: 2025-05-01
Backdoor Attack
Poisoning
Attack Type

How to Backdoor the Knowledge Distillation

Authors: Chen Wu, Qian Ma, Prasenjit Mitra, Sencun Zhu | Published: 2025-04-30
Backdoor Attack
Adversarial Learning
Vulnerabilities of Knowledge Distillation

Detecting Instruction Fine-tuning Attacks on Language Models using Influence Function

Authors: Jiawei Li | Published: 2025-04-12 | Updated: 2025-09-30
Backdoor Attack
Prompt validation
Sentiment Analysis

BadToken: Token-level Backdoor Attacks to Multi-modal Large Language Models

Authors: Zenghui Yuan, Jiawen Shi, Pan Zhou, Neil Zhenqiang Gong, Lichao Sun | Published: 2025-03-20
Backdoor Attack
Prompt Injection
Large Language Model