Whisper Leak: a side-channel attack on Large Language Models

Authors: Geoff McDonald, Jonathan Bar Or | Published: 2025-11-05

Watermarking Large Language Models in Europe: Interpreting the AI Act in Light of Technology

Authors: Thomas Souverain | Published: 2025-11-05

Let the Bees Find the Weak Spots: A Path Planning Perspective on Multi-Turn Jailbreak Attacks against LLMs

Authors: Yize Liu, Yunyun Hou, Aina Sui | Published: 2025-11-05

Auditing M-LLMs for Privacy Risks: A Synthetic Benchmark and Evaluation Framework

Authors: Junhao Li, Jiahao Chen, Zhou Feng, Chunyi Zhou | Published: 2025-11-05

Death by a Thousand Prompts: Open Model Vulnerability Analysis

Authors: Amy Chang, Nicholas Conley, Harish Santhanalakshmi Ganesan, Adam Swanda | Published: 2025-11-05

Measuring the Security of Mobile LLM Agents under Adversarial Prompts from Untrusted Third-Party Channels

Authors: Chenghao Du, Quanfeng Huang, Tingxuan Tang, Zihao Wang, Adwait Nadkarni, Yue Xiao | Published: 2025-10-31 | Updated: 2025-11-06

PVMark: Enabling Public Verifiability for LLM Watermarking Schemes

Authors: Haohua Duan, Liyao Xiang, Xin Zhang | Published: 2025-10-30

ALMGuard: Safety Shortcuts and Where to Find Them as Guardrails for Audio-Language Models

Authors: Weifei Jin, Yuxin Cao, Junjie Su, Minhui Xue, Jie Hao, Ke Xu, Jin Song Dong, Derui Wang | Published: 2025-10-30

Model Inversion Attacks Meet Cryptographic Fuzzy Extractors

Authors: Mallika Prabhakar, Louise Xu, Prateek Saxena | Published: 2025-10-29

Spectral Perturbation Bounds for Low-Rank Approximation with Applications to Privacy

Authors: Phuc Tran, Nisheeth K. Vishnoi, Van H. Vu | Published: 2025-10-29