Defensive Approximation: Securing CNNs using Approximate Computing

Authors: Amira Guesmi, Ihsen Alouani, Khaled Khasawneh, Mouna Baklouti, Tarek Frikha, Mohamed Abid, Nael Abu-Ghazaleh | Published: 2020-06-13 | Updated: 2021-07-29

Rethinking Clustering for Robustness

Authors: Motasem Alfarra, Juan C. Pérez, Adel Bibi, Ali Thabet, Pablo Arbeláez, Bernard Ghanem | Published: 2020-06-13 | Updated: 2021-11-19

Adversarial Self-Supervised Contrastive Learning

Authors: Minseon Kim, Jihoon Tack, Sung Ju Hwang | Published: 2020-06-13 | Updated: 2020-10-26

Leakage of Dataset Properties in Multi-Party Machine Learning

Authors: Wanrong Zhang, Shruti Tople, Olga Ohrimenko | Published: 2020-06-12 | Updated: 2021-06-17

Backdoor Attacks on Federated Meta-Learning

Authors: Chien-Lun Chen, Leana Golubchik, Marco Paolieri | Published: 2020-06-12 | Updated: 2020-12-16

Provably Robust Metric Learning

Authors: Lu Wang, Xuanqing Liu, Jinfeng Yi, Yuan Jiang, Cho-Jui Hsieh | Published: 2020-06-12 | Updated: 2020-12-19

Robustness to Adversarial Attacks in Learning-Enabled Controllers

Authors: Zikang Xiong, Joe Eappen, He Zhu, Suresh Jagannathan | Published: 2020-06-11

Backdoors in Neural Models of Source Code

Authors: Goutham Ramakrishnan, Aws Albarghouthi | Published: 2020-06-11

On the Tightness of Semidefinite Relaxations for Certifying Robustness to Adversarial Examples

Authors: Richard Y. Zhang | Published: 2020-06-11 | Updated: 2020-10-26

Backdoor Smoothing: Demystifying Backdoor Attacks on Deep Neural Networks

Authors: Kathrin Grosse, Taesung Lee, Battista Biggio, Youngja Park, Michael Backes, Ian Molloy | Published: 2020-06-11 | Updated: 2021-11-02