Security through the Eyes of AI: How Visualization is Shaping Malware Detection Authors: Asmitha K. A., Matteo Brosolo, Serena Nicolazzo, Antonino Nocera, Vinod P., Rafidha Rehiman K. A., Muhammed Shafi K. P | Published: 2025-05-12 Prompt InjectionMalware ClassificationAdversarial Example Detection 2025.05.12 2025.05.14 Literature Database
One Trigger Token Is Enough: A Defense Strategy for Balancing Safety and Usability in Large Language Models Authors: Haoran Gu, Handing Wang, Yi Mei, Mengjie Zhang, Yaochu Jin | Published: 2025-05-12 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.12 2025.05.14 Literature Database
Red Teaming the Mind of the Machine: A Systematic Evaluation of Prompt Injection and Jailbreak Vulnerabilities in LLMs Authors: Chetan Pathade | Published: 2025-05-07 | Updated: 2025-05-13 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.07 2025.05.15 Literature Database
LlamaFirewall: An open source guardrail system for building secure AI agents Authors: Sahana Chennabasappa, Cyrus Nikolaidis, Daniel Song, David Molnar, Stephanie Ding, Shengye Wan, Spencer Whitman, Lauren Deason, Nicholas Doucette, Abraham Montilla, Alekhya Gampa, Beto de Paola, Dominik Gabi, James Crnkovich, Jean-Christophe Testud, Kat He, Rashnil Chaturvedi, Wu Zhou, Joshua Saxe | Published: 2025-05-06 LLM SecurityAlignmentPrompt Injection 2025.05.06 2025.05.12 Literature Database
Directed Greybox Fuzzing via Large Language Model Authors: Hanxiang Xu, Yanjie Zhao, Haoyu Wang | Published: 2025-05-06 RAGPrompt InjectionVulnerability Analysis 2025.05.06 2025.05.12 Literature Database
LLM-Based Threat Detection and Prevention Framework for IoT Ecosystems Authors: Yazan Otoum, Arghavan Asad, Amiya Nayak | Published: 2025-05-01 Bias Detection in AI OutputLLM Performance EvaluationPrompt Injection 2025.05.01 2025.05.12 Literature Database
An Empirical Study on the Effectiveness of Large Language Models for Binary Code Understanding Authors: Xiuwei Shang, Zhenkan Fu, Shaoyin Cheng, Guoqiang Chen, Gangyang Li, Li Hu, Weiming Zhang, Nenghai Yu | Published: 2025-04-30 Program AnalysisPrompt InjectionPrompt leaking 2025.04.30 2025.05.12 Literature Database
LASHED: LLMs And Static Hardware Analysis for Early Detection of RTL Bugs Authors: Baleegh Ahmad, Hammond Pearce, Ramesh Karri, Benjamin Tan | Published: 2025-04-30 Program AnalysisPrompt InjectionVulnerability detection 2025.04.30 2025.05.12 Literature Database
XBreaking: Explainable Artificial Intelligence for Jailbreaking LLMs Authors: Marco Arazzi, Vignesh Kumar Kembu, Antonino Nocera, Vinod P | Published: 2025-04-30 Disabling Safety Mechanisms of LLMPrompt InjectionExplanation Method 2025.04.30 2025.05.12 Literature Database
ACE: A Security Architecture for LLM-Integrated App Systems Authors: Evan Li, Tushin Mallick, Evan Rose, William Robertson, Alina Oprea, Cristina Nita-Rotaru | Published: 2025-04-29 | Updated: 2025-05-07 Indirect Prompt InjectionPrompt InjectionInformation Flow Analysis 2025.04.29 2025.05.12 Literature Database