Is Spiking Secure? A Comparative Study on the Security Vulnerabilities of Spiking and Deep Neural Networks

Authors: Alberto Marchisio, Giorgio Nanfa, Faiq Khalid, Muhammad Abdullah Hanif, Maurizio Martina, Muhammad Shafique | Published: 2019-02-04 | Updated: 2020-05-18

Robustness of Generalized Learning Vector Quantization Models against Adversarial Attacks

Authors: Sascha Saralajew, Lars Holdijk, Maike Rees, Thomas Villmann | Published: 2019-02-01 | Updated: 2019-03-09

The Efficacy of SHIELD under Different Threat Models

Authors: Cory Cornelius, Nilaksh Das, Shang-Tse Chen, Li Chen, Michael E. Kounavis, Duen Horng Chau | Published: 2019-02-01 | Updated: 2019-08-02

Robustness Certificates Against Adversarial Examples for ReLU Networks

Authors: Sahil Singla, Soheil Feizi | Published: 2019-02-01 | Updated: 2019-02-05

Natural and Adversarial Error Detection using Invariance to Image Transformations

Authors: Yuval Bahat, Michal Irani, Gregory Shakhnarovich | Published: 2019-02-01

Optimal Attack against Autoregressive Models by Manipulating the Environment

Authors: Yiding Chen, Xiaojin Zhu | Published: 2019-02-01 | Updated: 2019-12-01

A New Family of Neural Networks Provably Resistant to Adversarial Attacks

Authors: Rakshit Agrawal, Luca de Alfaro, David Helmbold | Published: 2019-02-01

Training Artificial Neural Networks by Generalized Likelihood Ratio Method: Exploring Brain-like Learning to Improve Robustness

Authors: Li Xiao, Yijie Peng, Jeff Hong, Zewu Ke, Shuhuai Yang | Published: 2019-01-31 | Updated: 2019-07-11

CapsAttacks: Robust and Imperceptible Adversarial Attacks on Capsule Networks

Authors: Alberto Marchisio, Giorgio Nanfa, Faiq Khalid, Muhammad Abdullah Hanif, Maurizio Martina, Muhammad Shafique | Published: 2019-01-28 | Updated: 2019-05-24

Utility Preserving Secure Private Data Release

Authors: Jasjeet Dhaliwal, Geoffrey So, Aleatha Parker-Wood, Melanie Beck | Published: 2019-01-28 | Updated: 2019-03-15