SAPAG: A Self-Adaptive Privacy Attack From Gradients

Authors: Yijue Wang, Jieren Deng, Dan Guo, Chenghong Wang, Xianrui Meng, Hang Liu, Caiwen Ding, Sanguthevar Rajasekaran | Published: 2020-09-14

Manifold attack

Authors: Khanh-Hung Tran, Fred-Maurice Ngole-Mboula, Jean-Luc Starck | Published: 2020-09-13 | Updated: 2021-03-17

MACE: A Flexible Framework for Membership Privacy Estimation in Generative Models

Authors: Yixi Xu, Sumit Mukherjee, Xiyang Liu, Shruti Tople, Rahul Dodhia, Juan Lavista Ferres | Published: 2020-09-11 | Updated: 2022-10-12

Quantifying Membership Inference Vulnerability via Generalization Gap and Other Model Metrics

Authors: Jason W. Bentley, Daniel Gibney, Gary Hoppenworth, Sumit Kumar Jha | Published: 2020-09-11

Machine Unlearning for Random Forests

Authors: Jonathan Brophy, Daniel Lowd | Published: 2020-09-11 | Updated: 2021-06-11

Federated Model Distillation with Noise-Free Differential Privacy

Authors: Lichao Sun, Lingjuan Lyu | Published: 2020-09-11 | Updated: 2021-05-21

Second Order Optimization for Adversarial Robustness and Interpretability

Authors: Theodoros Tsiligkaridis, Jay Roberts | Published: 2020-09-10

Privacy Analysis of Deep Learning in the Wild: Membership Inference Attacks against Transfer Learning

Authors: Yang Zou, Zhikun Zhang, Michael Backes, Yang Zhang | Published: 2020-09-10

Quantifying the Preferential Direction of the Model Gradient in Adversarial Training With Projected Gradient Descent

Authors: Ricardo Bigolin Lanfredi, Joyce D. Schroeder, Tolga Tasdizen | Published: 2020-09-10 | Updated: 2023-04-20

A black-box adversarial attack for poisoning clustering

Authors: Antonio Emanuele Cinà, Alessandro Torcinovich, Marcello Pelillo | Published: 2020-09-09 | Updated: 2021-11-10