XBreaking: Explainable Artificial Intelligence for Jailbreaking LLMs Authors: Marco Arazzi, Vignesh Kumar Kembu, Antonino Nocera, Vinod P | Published: 2025-04-30 LLMの安全機構の解除プロンプトインジェクション説明手法 2025.04.30 2025.05.12 Literature Database
On the Privacy Risks of Model Explanations Authors: Reza Shokri, Martin Strobel, Yair Zick | Published: 2019-06-29 | Updated: 2021-02-05 メンバーシップ推論敵対的攻撃説明手法 2019.06.29 2025.05.13 Literature Database