Defensive Dropout for Hardening Deep Neural Networks under Adversarial Attacks

Authors: Siyue Wang, Xiao Wang, Pu Zhao, Wujie Wen, David Kaeli, Peter Chin, Xue Lin | Published: 2018-09-13

Query-Efficient Black-Box Attack by Active Learning

Authors: Pengcheng Li, Jinfeng Yi, Lijun Zhang | Published: 2018-09-13

Adversarial Examples: Opportunities and Challenges

Authors: Jiliang Zhang, Chen Li | Published: 2018-09-13 | Updated: 2019-09-23

Deep Learning in Information Security

Authors: Stefan Thaler, Vlado Menkovski, Milan Petkovic | Published: 2018-09-12

Poisoning Attacks to Graph-Based Recommender Systems

Authors: Minghong Fang, Guolei Yang, Neil Zhenqiang Gong, Jia Liu | Published: 2018-09-11

PUF-AES-PUF: a novel PUF architecture against non-invasive attacks

Authors: Weize Yu, Jia Chen | Published: 2018-09-11

Universal Multi-Party Poisoning Attacks

Authors: Saeed Mahloujifar, Mohammad Mahmoody, Ameer Mohammed | Published: 2018-09-10 | Updated: 2021-11-10

Privacy-Preserving Deep Learning via Weight Transmission

Authors: Le Trieu Phong, Tran Thi Phuong | Published: 2018-09-10 | Updated: 2019-02-12

Certified Adversarial Robustness with Additive Noise

Authors: Bai Li, Changyou Chen, Wenlin Wang, Lawrence Carin | Published: 2018-09-10 | Updated: 2019-11-10

The Curse of Concentration in Robust Learning: Evasion and Poisoning Attacks from Concentration of Measure

Authors: Saeed Mahloujifar, Dimitrios I. Diochnos, Mohammad Mahmoody | Published: 2018-09-09 | Updated: 2018-11-06