敵対的サンプル

Survey of Vulnerabilities in Large Language Models Revealed by Adversarial Attacks

Authors: Erfan Shayegani, Md Abdullah Al Mamun, Yu Fu, Pedram Zaree, Yue Dong, Nael Abu-Ghazaleh | Published: 2023-10-16
プロンプトインジェクション
敵対的サンプル
敵対的訓練

BufferSearch: Generating Black-Box Adversarial Texts With Lower Queries

Authors: Wenjie Lv, Zhen Wang, Yitao Zheng, Zhehua Zhong, Qi Xuan, Tianyi Chen | Published: 2023-10-14
攻撃の評価
敵対的サンプル
最適化手法

Misusing Tools in Large Language Models With Visual Adversarial Examples

Authors: Xiaohan Fu, Zihan Wang, Shuheng Li, Rajesh K. Gupta, Niloofar Mireshghallah, Taylor Berg-Kirkpatrick, Earlence Fernandes | Published: 2023-10-04
LLM性能評価
プロンプトインジェクション
敵対的サンプル

DAD++: Improved Data-free Test Time Adversarial Defense

Authors: Gaurav Kumar Nayak, Inder Khatri, Shubham Randive, Ruchit Rawal, Anirban Chakraborty | Published: 2023-09-10
敵対的サンプル
敵対的攻撃
防御手法

Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities

Authors: Maximilian Mozes, Xuanli He, Bennett Kleinberg, Lewis D. Griffin | Published: 2023-08-24
プロンプトインジェクション
悪意のあるコンテンツ生成
敵対的サンプル

Sample Complexity of Robust Learning against Evasion Attacks

Authors: Pascale Gourdeau | Published: 2023-08-23
サンプル複雑性
ポイズニング
敵対的サンプル

A Comparison of Adversarial Learning Techniques for Malware Detection

Authors: Pavla Louthánová, Matouš Kozák, Martin Jureček, Mark Stamp | Published: 2023-08-19
マルウェア検出
敵対的サンプル
敵対的攻撃

Data-Free Model Extraction Attacks in the Context of Object Detection

Authors: Harshit Shah, Aravindhan G, Pavan Kulkarni, Yuvaraj Govidarajulu, Manojkumar Parmar | Published: 2023-08-09
データセット生成
モデル抽出攻撃
敵対的サンプル

A reading survey on adversarial machine learning: Adversarial attacks and their understanding

Authors: Shashank Kotyan | Published: 2023-08-07
敵対的サンプル
敵対的攻撃手法
防御手法

URET: Universal Robustness Evaluation Toolkit (for Evasion)

Authors: Kevin Eykholt, Taesung Lee, Douglas Schales, Jiyong Jang, Ian Molloy, Masha Zorin | Published: 2023-08-03
データ生成
敵対的サンプル
敵対的攻撃