Is Your Prompt Safe? Investigating Prompt Injection Attacks Against Open-Source LLMs Authors: Jiawen Wang, Pritha Gupta, Ivan Habernal, Eyke Hüllermeier | Published: 2025-05-20 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.20 2025.05.28 Literature Database
Exploring Jailbreak Attacks on LLMs through Intent Concealment and Diversion Authors: Tiehan Cui, Yanxu Mao, Peipei Liu, Congying Liu, Datao You | Published: 2025-05-20 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.20 2025.05.28 Literature Database
PandaGuard: Systematic Evaluation of LLM Safety against Jailbreaking Attacks Authors: Guobin Shen, Dongcheng Zhao, Linghao Feng, Xiang He, Jihang Wang, Sicheng Shen, Haibo Tong, Yiting Dong, Jindong Li, Xiang Zheng, Yi Zeng | Published: 2025-05-20 | Updated: 2025-05-22 Disabling Safety Mechanisms of LLMPrompt InjectionEffectiveness Analysis of Defense Methods 2025.05.20 2025.05.28 Literature Database
Evaluating the efficacy of LLM Safety Solutions : The Palit Benchmark Dataset Authors: Sayon Palit, Daniel Woods | Published: 2025-05-19 | Updated: 2025-05-20 LLM SecurityPrompt InjectionAttack Method 2025.05.19 2025.05.28 Literature Database
JULI: Jailbreak Large Language Models by Self-Introspection Authors: Jesson Wang, Zhanhao Hu, David Wagner | Published: 2025-05-17 | Updated: 2025-05-20 API SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.17 2025.05.28 Literature Database
Dark LLMs: The Growing Threat of Unaligned AI Models Authors: Michael Fire, Yitzhak Elbazis, Adi Wasenstein, Lior Rokach | Published: 2025-05-15 Disabling Safety Mechanisms of LLMPrompt InjectionLarge Language Model 2025.05.15 2025.05.28 Literature Database
Analysing Safety Risks in LLMs Fine-Tuned with Pseudo-Malicious Cyber Security Data Authors: Adel ElZemity, Budi Arief, Shujun Li | Published: 2025-05-15 LLM SecurityPrompt InjectionLarge Language Model 2025.05.15 2025.05.28 Literature Database
PIG: Privacy Jailbreak Attack on LLMs via Gradient-based Iterative In-Context Optimization Authors: Yidan Wang, Yanan Cao, Yubing Ren, Fang Fang, Zheng Lin, Binxing Fang | Published: 2025-05-15 Disabling Safety Mechanisms of LLMPrompt InjectionPrivacy Protection in Machine Learning 2025.05.15 2025.05.28 Literature Database
SecReEvalBench: A Multi-turned Security Resilience Evaluation Benchmark for Large Language Models Authors: Huining Cui, Wei Liu | Published: 2025-05-12 LLM SecurityPrompt InjectionPrompt leaking 2025.05.12 2025.05.28 Literature Database
Security through the Eyes of AI: How Visualization is Shaping Malware Detection Authors: Asmitha K. A., Matteo Brosolo, Serena Nicolazzo, Antonino Nocera, Vinod P., Rafidha Rehiman K. A., Muhammed Shafi K. P | Published: 2025-05-12 Prompt InjectionMalware ClassificationAdversarial Example Detection 2025.05.12 2025.05.28 Literature Database