バックドア攻撃

Adaptive Reward-Poisoning Attacks against Reinforcement Learning

Authors: Xuezhou Zhang, Yuzhe Ma, Adish Singla, Xiaojin Zhu | Published: 2020-03-27 | Updated: 2020-06-22
Q-Learningアルゴリズム
バックドア攻撃
強化学習攻撃

RAB: Provable Robustness Against Backdoor Attacks

Authors: Maurice Weber, Xiaojun Xu, Bojan Karlaš, Ce Zhang, Bo Li | Published: 2020-03-19 | Updated: 2023-08-03
バックドア攻撃
ロバスト性
敵対的サンプル

Towards Probabilistic Verification of Machine Unlearning

Authors: David Marco Sommer, Liwei Song, Sameer Wagh, Prateek Mittal | Published: 2020-03-09 | Updated: 2020-12-01
トレーニング手法
バックドア攻撃
バックドア攻撃対策

Towards Backdoor Attacks and Defense in Robust Machine Learning Models

Authors: Ezekiel Soremekun, Sakshi Udeshi, Sudipta Chattopadhyay | Published: 2020-02-25 | Updated: 2023-01-11
バックドア攻撃
ポイズニング
ロバスト性評価

Label-Consistent Backdoor Attacks

Authors: Alexander Turner, Dimitris Tsipras, Aleksander Madry | Published: 2019-12-05 | Updated: 2019-12-06
バックドア攻撃
ポイズニング
敵対的サンプル

Deep Probabilistic Models to Detect Data Poisoning Attacks

Authors: Mahesh Subedar, Nilesh Ahuja, Ranganath Krishnan, Ibrahima J. Ndiour, Omesh Tickoo | Published: 2019-12-03
バックドア攻撃
ポイズニング攻撃
性能評価

Failure Modes in Machine Learning Systems

Authors: Ram Shankar Siva Kumar, David O Brien, Kendra Albert, Salomé Viljöen, Jeffrey Snover | Published: 2019-11-25
バックドア攻撃
プライバシー侵害
敵対的サンプル

Revealing Perceptible Backdoors, without the Training Set, via the Maximum Achievable Misclassification Fraction Statistic

Authors: Zhen Xiang, David J. Miller, Hang Wang, George Kesidis | Published: 2019-11-18 | Updated: 2020-04-06
DDIAの検出と位置特定
バックドア攻撃
ポイズニング

A New Ensemble Adversarial Attack Powered by Long-term Gradient Memories

Authors: Zhaohui Che, Ali Borji, Guangtao Zhai, Suiyi Ling, Jing Li, Patrick Le Callet | Published: 2019-11-18
バックドア攻撃
モデル性能評価
敵対的攻撃手法

Robust Anomaly Detection and Backdoor Attack Detection Via Differential Privacy

Authors: Min Du, Ruoxi Jia, Dawn Song | Published: 2019-11-16
バックドア攻撃
プライバシー保護データマイニング
実験的検証