Backdoor Attacks to Graph Neural Networks

Authors: Zaixi Zhang, Jinyuan Jia, Binghui Wang, Neil Zhenqiang Gong | Published: 2020-06-19 | Updated: 2021-12-17

Systematic Attack Surface Reduction For Deployed Sentiment Analysis Models

Authors: Josh Kalin, David Noever, Gerry Dozier | Published: 2020-06-19

A general framework for defining and optimizing robustness

Authors: Alessandro Tibo, Manfred Jaeger, Kim G. Larsen | Published: 2020-06-19 | Updated: 2021-05-29

Differentiable Language Model Adversarial Attacks on Categorical Sequence Classifiers

Authors: I. Fursov, A. Zaytsev, N. Kluchnikov, A. Kravchenko, E. Burnaev | Published: 2020-06-19

Towards an Adversarially Robust Normalization Approach

Authors: Muhammad Awais, Fahad Shamshad, Sung-Ho Bae | Published: 2020-06-19

Adversarial Attacks for Multi-view Deep Models

Authors: Xuli Sun, Shiliang Sun | Published: 2020-06-19

Beware the Black-Box: on the Robustness of Recent Defenses to Adversarial Examples

Authors: Kaleel Mahmood, Deniz Gurevin, Marten van Dijk, Phuong Ha Nguyen | Published: 2020-06-18 | Updated: 2021-05-20

REGroup: Rank-aggregating Ensemble of Generative Classifiers for Robust Predictions

Authors: Lokender Tiwari, Anish Madan, Saket Anand, Subhashis Banerjee | Published: 2020-06-18 | Updated: 2021-11-24

Local Competition and Uncertainty for Adversarial Robustness in Deep Learning

Authors: Antonios Alexos, Konstantinos P. Panousis, Sotirios Chatzis | Published: 2020-06-18

OGAN: Disrupting Deepfakes with an Adversarial Attack that Survives Training

Authors: Eran Segalis, Eran Galili | Published: 2020-06-17 | Updated: 2020-11-25