Black-Box Guardrail Reverse-engineering Attack

Authors: Hongwei Yao, Yun Xia, Shuo Shao, Haoran Shi, Tong Qiao, Cong Wang | Published: 2025-11-06

Automated and Explainable Denial of Service Analysis for AI-Driven Intrusion Detection Systems

Authors: Paul Badu Yakubu, Lesther Santana, Mohamed Rahouti, Yufeng Xin, Abdellah Chehri, Mohammed Aledhari | Published: 2025-11-06

Specification-Guided Vulnerability Detection with Large Language Models

Authors: Hao Zhu, Jia Li, Cuiyun Gao, Jiaru Qian, Yihong Dong, Huanyu Liu, Lecheng Wang, Ziliang Wang, Xiaolong Hu, Ge Li | Published: 2025-11-06

Hybrid Fuzzing with LLM-Guided Input Mutation and Semantic Feedback

Authors: Shiyin Lin | Published: 2025-11-06

Whisper Leak: a side-channel attack on Large Language Models

Authors: Geoff McDonald, Jonathan Bar Or | Published: 2025-11-05

Watermarking Large Language Models in Europe: Interpreting the AI Act in Light of Technology

Authors: Thomas Souverain | Published: 2025-11-05

Let the Bees Find the Weak Spots: A Path Planning Perspective on Multi-Turn Jailbreak Attacks against LLMs

Authors: Yize Liu, Yunyun Hou, Aina Sui | Published: 2025-11-05

Auditing M-LLMs for Privacy Risks: A Synthetic Benchmark and Evaluation Framework

Authors: Junhao Li, Jiahao Chen, Zhou Feng, Chunyi Zhou | Published: 2025-11-05

Death by a Thousand Prompts: Open Model Vulnerability Analysis

Authors: Amy Chang, Nicholas Conley, Harish Santhanalakshmi Ganesan, Adam Swanda | Published: 2025-11-05

Measuring the Security of Mobile LLM Agents under Adversarial Prompts from Untrusted Third-Party Channels

Authors: Chenghao Du, Quanfeng Huang, Tingxuan Tang, Zihao Wang, Adwait Nadkarni, Yue Xiao | Published: 2025-10-31 | Updated: 2025-11-06