LLM Security

Weaponizing Language Models for Cybersecurity Offensive Operations: Automating Vulnerability Assessment Report Validation; A Review Paper

Authors: Abdulrahman S Almuhaidib, Azlan Mohd Zain, Zalmiyah Zakaria, Izyan Izzati Kamsani, Abdulaziz S Almuhaidib | Published: 2025-05-07
LLM Security
Vulnerability Analysis

LLMs’ Suitability for Network Security: A Case Study of STRIDE Threat Modeling

Authors: AbdulAziz AbdulGhaffar, Ashraf Matrawy | Published: 2025-05-07
LLM Security
Performance Evaluation
Vulnerability Analysis

LlamaFirewall: An open source guardrail system for building secure AI agents

Authors: Sahana Chennabasappa, Cyrus Nikolaidis, Daniel Song, David Molnar, Stephanie Ding, Shengye Wan, Spencer Whitman, Lauren Deason, Nicholas Doucette, Abraham Montilla, Alekhya Gampa, Beto de Paola, Dominik Gabi, James Crnkovich, Jean-Christophe Testud, Kat He, Rashnil Chaturvedi, Wu Zhou, Joshua Saxe | Published: 2025-05-06
LLM Security
Alignment
Prompt Injection

Output Constraints as Attack Surface: Exploiting Structured Generation to Bypass LLM Safety Mechanisms

Authors: Shuoming Zhang, Jiacheng Zhao, Ruiyuan Xu, Xiaobing Feng, Huimin Cui | Published: 2025-03-31
LLM Security
Disabling Safety Mechanisms of LLM
Prompt Injection

Get the Agents Drunk: Memory Perturbations in Autonomous Agent-based Recommender Systems

Authors: Shiyi Yang, Zhibo Hu, Chen Wang, Tong Yu, Xiwei Xu, Liming Zhu, Lina Yao | Published: 2025-03-31
LLM Security
Indirect Prompt Injection
Model DoS

Intelligent IoT Attack Detection Design via ODLLM with Feature Ranking-based Knowledge Base

Authors: Satvik Verma, Qun Wang, E. Wes Bethel | Published: 2025-03-27
DDoS Attack Detection
LLM Security
Network Traffic Analysis

Beyond Surface-Level Patterns: An Essence-Driven Defense Framework Against Jailbreak Attacks in LLMs

Authors: Shiyu Xiang, Ansen Zhang, Yanfei Cao, Yang Fan, Ronghao Chen | Published: 2025-02-26 | Updated: 2025-05-28
LLM Security
Prompt Injection
Attack Evaluation

DELMAN: Dynamic Defense Against Large Language Model Jailbreaking with Model Editing

Authors: Yi Wang, Fenghua Weng, Sibei Yang, Zhan Qin, Minlie Huang, Wenjie Wang | Published: 2025-02-17 | Updated: 2025-05-29
LLM Security
Prompt Injection
Defense Method

CL-Attack: Textual Backdoor Attacks via Cross-Lingual Triggers

Authors: Jingyi Zheng, Tianyi Hu, Tianshuo Cong, Xinlei He | Published: 2024-12-26 | Updated: 2025-03-31
LLM Security
Backdoor Attack
Vulnerability of Adversarial Examples

Jailbreaking and Mitigation of Vulnerabilities in Large Language Models

Authors: Benji Peng, Keyu Chen, Qian Niu, Ziqian Bi, Ming Liu, Pohsun Feng, Tianyang Wang, Lawrence K. Q. Yan, Yizhu Wen, Yichao Zhang, Caitlyn Heqi Yin | Published: 2024-10-20 | Updated: 2025-05-08
LLM Security
Disabling Safety Mechanisms of LLM
Prompt Injection