Security through the Eyes of AI: How Visualization is Shaping Malware Detection Authors: Asmitha K. A., Matteo Brosolo, Serena Nicolazzo, Antonino Nocera, Vinod P., Rafidha Rehiman K. A., Muhammed Shafi K. P | Published: 2025-05-12 Prompt InjectionMalware ClassificationAdversarial Example Detection 2025.05.12 2025.05.28 Literature Database
One Trigger Token Is Enough: A Defense Strategy for Balancing Safety and Usability in Large Language Models Authors: Haoran Gu, Handing Wang, Yi Mei, Mengjie Zhang, Yaochu Jin | Published: 2025-05-12 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.12 2025.05.28 Literature Database
Red Teaming the Mind of the Machine: A Systematic Evaluation of Prompt Injection and Jailbreak Vulnerabilities in LLMs Authors: Chetan Pathade | Published: 2025-05-07 | Updated: 2025-05-13 LLM SecurityDisabling Safety Mechanisms of LLMPrompt Injection 2025.05.07 2025.05.28 Literature Database
Safeguard-by-Development: A Privacy-Enhanced Development Paradigm for Multi-Agent Collaboration Systems Authors: Jian Cui, Zichuan Li, Luyi Xing, Xiaojing Liao | Published: 2025-05-07 | Updated: 2025-06-24 Privacy ProtectionPrivacy protection frameworkPrompt Injection 2025.05.07 2025.06.26 Literature Database
LlamaFirewall: An open source guardrail system for building secure AI agents Authors: Sahana Chennabasappa, Cyrus Nikolaidis, Daniel Song, David Molnar, Stephanie Ding, Shengye Wan, Spencer Whitman, Lauren Deason, Nicholas Doucette, Abraham Montilla, Alekhya Gampa, Beto de Paola, Dominik Gabi, James Crnkovich, Jean-Christophe Testud, Kat He, Rashnil Chaturvedi, Wu Zhou, Joshua Saxe | Published: 2025-05-06 LLM SecurityAlignmentPrompt Injection 2025.05.06 2025.05.27 Literature Database
Directed Greybox Fuzzing via Large Language Model Authors: Hanxiang Xu, Yanjie Zhao, Haoyu Wang | Published: 2025-05-06 RAGPrompt InjectionVulnerability Analysis 2025.05.06 2025.05.27 Literature Database
LLM-Based Threat Detection and Prevention Framework for IoT Ecosystems Authors: Yazan Otoum, Arghavan Asad, Amiya Nayak | Published: 2025-05-01 | Updated: 2025-05-13 Bias Detection in AI OutputLLM Performance EvaluationPrompt Injection 2025.05.01 2025.05.27 Literature Database
An Empirical Study on the Effectiveness of Large Language Models for Binary Code Understanding Authors: Xiuwei Shang, Zhenkan Fu, Shaoyin Cheng, Guoqiang Chen, Gangyang Li, Li Hu, Weiming Zhang, Nenghai Yu | Published: 2025-04-30 Program AnalysisPrompt InjectionPrompt leaking 2025.04.30 2025.05.27 Literature Database
LASHED: LLMs And Static Hardware Analysis for Early Detection of RTL Bugs Authors: Baleegh Ahmad, Hammond Pearce, Ramesh Karri, Benjamin Tan | Published: 2025-04-30 Program AnalysisPrompt InjectionVulnerability detection 2025.04.30 2025.05.27 Literature Database
XBreaking: Explainable Artificial Intelligence for Jailbreaking LLMs Authors: Marco Arazzi, Vignesh Kumar Kembu, Antonino Nocera, Vinod P | Published: 2025-04-30 Disabling Safety Mechanisms of LLMPrompt InjectionExplanation Method 2025.04.30 2025.05.27 Literature Database