Literature Database

Attack of the Tails: Yes, You Really Can Backdoor Federated Learning

Authors: Hongyi Wang, Kartik Sreenivasan, Shashank Rajput, Harit Vishwakarma, Saurabh Agarwal, Jy-yong Sohn, Kangwook Lee, Dimitris Papailiopoulos | Published: 2020-07-09
Poisoning
Model Robustness
Attack Method

Evaluation of Adversarial Training on Different Types of Neural Networks in Deep Learning-based IDSs

Authors: Rana Abou Khamis, Ashraf Matrawy | Published: 2020-07-08
Poisoning
Factors of Performance Degradation
Adversarial Training

On the relationship between class selectivity, dimensionality, and robustness

Authors: Matthew L. Leavitt, Ari S. Morcos | Published: 2020-07-08 | Updated: 2020-10-13
Poisoning
Adversarial Learning
Vulnerability Analysis

How benign is benign overfitting?

Authors: Amartya Sanyal, Puneet K Dokania, Varun Kanade, Philip H. S. Torr | Published: 2020-07-08
Adversarial Example
Adversarial Learning
Overfitting and Memorization

BlockFLow: An Accountable and Privacy-Preserving Solution for Federated Learning

Authors: Vaikkunth Mugunthan, Ravi Rahman, Lalana Kagal | Published: 2020-07-08
Performance Evaluation
Privacy Assessment
Attack Pattern Extraction

Defending against Backdoors in Federated Learning with Robust Learning Rate

Authors: Mustafa Safa Ozdayi, Murat Kantarcioglu, Yulia R. Gel | Published: 2020-07-07 | Updated: 2021-07-29
Backdoor Attack
Adversarial Learning
Defense Mechanism

Backdoor attacks and defenses in feature-partitioned collaborative learning

Authors: Yang Liu, Zhihao Yi, Tianjian Chen | Published: 2020-07-07
Poisoning
Adversarial Learning
Defense Mechanism

Stochastic Linear Bandits Robust to Adversarial Attacks

Authors: Ilija Bogunovic, Arpan Losalka, Andreas Krause, Jonathan Scarlett | Published: 2020-07-07 | Updated: 2020-10-27
Quantification of Uncertainty
Adversarial Learning
Computational Efficiency

Robust Learning with Frequency Domain Regularization

Authors: Weiyu Guo, Yidong Ouyang | Published: 2020-07-07
Adversarial Learning
Fundamentals of Machine Learning
Computational Efficiency

Regional Image Perturbation Reduces $L_p$ Norms of Adversarial Examples While Maintaining Model-to-model Transferability

Authors: Utku Ozbulak, Jonathan Peck, Wesley De Neve, Bart Goossens, Yvan Saeys, Arnout Van Messem | Published: 2020-07-07 | Updated: 2020-07-18
Attack Pattern Extraction
Adversarial Example
Adversarial Learning