Poisoning

Adversarially robust generalization theory via Jacobian regularization for deep neural networks

Authors: Dongya Wu, Xin Li | Published: 2024-12-17
Poisoning
Adversarial Example

GLL: A Differentiable Graph Learning Layer for Neural Networks

Authors: Jason Brown, Bohan Chen, Harris Hardiman-Mostow, Jeff Calder, Andrea L. Bertozzi | Published: 2024-12-11
Poisoning
Adversarial Training

Optimal Defenses Against Gradient Reconstruction Attacks

Authors: Yuxiao Chen, Gamze Gürsoy, Qi Lei | Published: 2024-11-06
Poisoning
Defense Method

FEDLAD: Federated Evaluation of Deep Leakage Attacks and Defenses

Authors: Isaac Baglin, Xiatian Zhu, Simon Hadfield | Published: 2024-11-05 | Updated: 2025-01-05
Poisoning
Attack Evaluation
Evaluation Method

Federated Learning in Practice: Reflections and Projections

Authors: Katharine Daly, Hubert Eichner, Peter Kairouz, H. Brendan McMahan, Daniel Ramage, Zheng Xu | Published: 2024-10-11
Privacy Protection
Privacy Protection Method
Poisoning

PoisonBench: Assessing Large Language Model Vulnerability to Data Poisoning

Authors: Tingchen Fu, Mrinank Sharma, Philip Torr, Shay B. Cohen, David Krueger, Fazl Barez | Published: 2024-10-11
LLM Performance Evaluation
Backdoor Attack
Poisoning

Data Taggants: Dataset Ownership Verification via Harmless Targeted Data Poisoning

Authors: Wassim Bouaziz, El-Mahdi El-Mhamdi, Nicolas Usunier | Published: 2024-10-09
Poisoning

CAT: Concept-level backdoor ATtacks for Concept Bottleneck Models

Authors: Songning Lai, Jiayu Yang, Yu Huang, Lijie Hu, Tianlang Xue, Zhangyi Hu, Jiaxu Li, Haicheng Liao, Yutao Yue | Published: 2024-10-07
Backdoor Attack
Poisoning

Federated Learning Nodes Can Reconstruct Peers’ Image Data

Authors: Ethan Wilson, Kai Yue, Chau-Wai Wong, Huaiyu Dai | Published: 2024-10-07
Privacy Protection
Poisoning

Empirical Perturbation Analysis of Linear System Solvers from a Data Poisoning Perspective

Authors: Yixin Liu, Arielle Carr, Lichao Sun | Published: 2024-10-01
Backdoor Attack
Poisoning
Linear Solver