Poisoning

Towards Class-Oriented Poisoning Attacks Against Neural Networks

Authors: Bingyin Zhao, Yingjie Lao | Published: 2020-07-31 | Updated: 2021-10-11
Backdoor Attack
Poisoning
Attack Method

Adversarial Attacks with Multiple Antennas Against Deep Learning-Based Modulation Classifiers

Authors: Brian Kim, Yalin E. Sagduyu, Tugba Erpek, Kemal Davaslioglu, Sennur Ulukus | Published: 2020-07-31
Poisoning
Attack Method
Deep Learning

Practical Detection of Trojan Neural Networks: Data-Limited and Data-Free Cases

Authors: Ren Wang, Gaoyuan Zhang, Sijia Liu, Pin-Yu Chen, Jinjun Xiong, Meng Wang | Published: 2020-07-31
Backdoor Attack
Poisoning
Attack Method

Dynamic Defense Against Byzantine Poisoning Attacks in Federated Learning

Authors: Nuria Rodríguez-Barroso, Eugenio Martínez-Cámara, M. Victoria Luzón, Francisco Herrera | Published: 2020-07-29 | Updated: 2022-02-24
Byzantine Resilience
Poisoning
Defense Mechanism

Backdoor Attacks and Countermeasures on Deep Learning: A Comprehensive Review

Authors: Yansong Gao, Bao Gia Doan, Zhi Zhang, Siqi Ma, Jiliang Zhang, Anmin Fu, Surya Nepal, Hyoungshick Kim | Published: 2020-07-21 | Updated: 2020-08-02
Backdoor Attack
Poisoning
Attack Method

Adversarial Immunization for Certifiable Robustness on Graphs

Authors: Shuchang Tao, Huawei Shen, Qi Cao, Liang Hou, Xueqi Cheng | Published: 2020-07-19 | Updated: 2021-08-25
Graph Transformation
Poisoning
Computational Complexity

Data Poisoning Attacks Against Federated Learning Systems

Authors: Vale Tolpegin, Stacey Truex, Mehmet Emre Gursoy, Ling Liu | Published: 2020-07-16 | Updated: 2020-08-11
Poisoning
Performance Evaluation
Attack Method

A simple defense against adversarial attacks on heatmap explanations

Authors: Laura Rieger, Lars Kai Hansen | Published: 2020-07-13
Poisoning
Attack Method
Defense Mechanism

Mitigating backdoor attacks in LSTM-based Text Classification Systems by Backdoor Keyword Identification

Authors: Chuanshuai Chen, Jiazhu Dai | Published: 2020-07-11 | Updated: 2021-03-15
Text Generation Method
Backdoor Attack
Poisoning

Improving Adversarial Robustness by Enforcing Local and Global Compactness

Authors: Anh Bui, Trung Le, He Zhao, Paul Montague, Olivier deVel, Tamas Abraham, Dinh Phung | Published: 2020-07-10
Poisoning
Performance Evaluation
Deep Learning