Data Poisoning Attacks Against Federated Learning Systems

Authors: Vale Tolpegin, Stacey Truex, Mehmet Emre Gursoy, Ling Liu | Published: 2020-07-16 | Updated: 2020-08-11

On Adversarial Robustness: A Neural Architecture Search perspective

Authors: Chaitanya Devaguptapu, Devansh Agarwal, Gaurav Mittal, Pulkit Gopalani, Vineeth N Balasubramanian | Published: 2020-07-16 | Updated: 2021-08-26

Towards Debiasing Sentence Representations

Authors: Paul Pu Liang, Irene Mengze Li, Emily Zheng, Yao Chong Lim, Ruslan Salakhutdinov, Louis-Philippe Morency | Published: 2020-07-16

AdvFlow: Inconspicuous Black-box Adversarial Attacks using Normalizing Flows

Authors: Hadi M. Dolatabadi, Sarah Erfani, Christopher Leckie | Published: 2020-07-15 | Updated: 2020-10-23

Robustifying Reinforcement Learning Agents via Action Space Adversarial Training

Authors: Kai Liang Tan, Yasaman Esfandiari, Xian Yeow Lee, Aakanksha, Soumik Sarkar | Published: 2020-07-14

Security and Machine Learning in the Real World

Authors: Ivan Evtimov, Weidong Cui, Ece Kamar, Emre Kiciman, Tadayoshi Kohno, Jerry Li | Published: 2020-07-13

A simple defense against adversarial attacks on heatmap explanations

Authors: Laura Rieger, Lars Kai Hansen | Published: 2020-07-13

Simple and Efficient Hard Label Black-box Adversarial Attacks in Low Query Budget Regimes

Authors: Satya Narayan Shukla, Anit Kumar Sahu, Devin Willmott, J. Zico Kolter | Published: 2020-07-13 | Updated: 2021-06-11

ManiGen: A Manifold Aided Black-box Generator of Adversarial Examples

Authors: Guanxiong Liu, Issa Khalil, Abdallah Khreishah, Abdulelah Algosaibi, Adel Aldalbahi, Mohammed Alaneem, Abdulaziz Alhumam, Mohammed Anan | Published: 2020-07-11

Mitigating backdoor attacks in LSTM-based Text Classification Systems by Backdoor Keyword Identification

Authors: Chuanshuai Chen, Jiazhu Dai | Published: 2020-07-11 | Updated: 2021-03-15