MAZE: Data-Free Model Stealing Attack Using Zeroth-Order Gradient Estimation

Authors: Sanjay Kariyappa, Atul Prakash, Moinuddin Qureshi | Published: 2020-05-06 | Updated: 2022-10-28

Towards Frequency-Based Explanation for Robust CNN

Authors: Zifan Wang, Yilin Yang, Ankit Shrivastava, Varun Rawal, Zihao Ding | Published: 2020-05-06

Insider Threat Detection Based on Stress Recognition Using Keystroke Dynamics

Authors: Azamat Sultanov, Konstantin Kogos | Published: 2020-05-06

Enhancing Intrinsic Adversarial Robustness via Feature Pyramid Decoder

Authors: Guanlin Li, Shuya Ding, Jun Luo, Chang Liu | Published: 2020-05-06

Information-Theoretic Bounds on the Generalization Error and Privacy Leakage in Federated Learning

Authors: Semih Yagli, Alex Dytso, H. Vincent Poor | Published: 2020-05-05

Adversarial Training against Location-Optimized Adversarial Patches

Authors: Sukrut Rao, David Stutz, Bernt Schiele | Published: 2020-05-05 | Updated: 2020-12-14

When Machine Unlearning Jeopardizes Privacy

Authors: Min Chen, Zhikun Zhang, Tianhao Wang, Michael Backes, Mathias Humbert, Yang Zhang | Published: 2020-05-05 | Updated: 2021-09-14

Stealing Links from Graph Neural Networks

Authors: Xinlei He, Jinyuan Jia, Michael Backes, Neil Zhenqiang Gong, Yang Zhang | Published: 2020-05-05 | Updated: 2020-10-05

Mind the Gap: On Bridging the Semantic Gap between Machine Learning and Information Security

Authors: Michael R. Smith, Nicholas T. Johnson, Joe B. Ingram, Armida J. Carbajal, Ramyaa Ramyaa, Evelyn Domschot, Christopher C. Lamb, Stephen J. Verzi, W. Philip Kegelmeyer | Published: 2020-05-04

Depth-2 Neural Networks Under a Data-Poisoning Attack

Authors: Sayar Karmakar, Anirbit Mukherjee, Theodore Papamarkou | Published: 2020-05-04 | Updated: 2022-06-29