XBreaking: Explainable Artificial Intelligence for Jailbreaking LLMs Authors: Marco Arazzi, Vignesh Kumar Kembu, Antonino Nocera, Vinod P | Published: 2025-04-30 Disabling Safety Mechanisms of LLMPrompt InjectionExplanation Method 2025.04.30 2025.05.27 Literature Database
On the Privacy Risks of Model Explanations Authors: Reza Shokri, Martin Strobel, Yair Zick | Published: 2019-06-29 | Updated: 2021-02-05 Membership InferenceAdversarial attackExplanation Method 2019.06.29 2025.05.28 Literature Database