Specification-Guided Vulnerability Detection with Large Language Models Authors: Hao Zhu, Jia Li, Cuiyun Gao, Jiaru Qian, Yihong Dong, Huanyu Liu, Lecheng Wang, Ziliang Wang, Xiaolong Hu, Ge Li | Published: 2025-11-06 Prompt InjectionLarge Language Model脆弱性検出手法 2025.11.06 2025.11.08 Literature Database
Whisper Leak: a side-channel attack on Large Language Models Authors: Geoff McDonald, Jonathan Bar Or | Published: 2025-11-05 Traffic Characteristic AnalysisPrompt leakingLarge Language Model 2025.11.05 2025.11.07 Literature Database
Beyond Text: Multimodal Jailbreaking of Vision-Language and Audio Models through Perceptually Simple Transformations Authors: Divyanshu Kumar, Shreyas Jena, Nitin Aravind Birur, Tanay Baswa, Sahil Agarwal, Prashanth Harshangi | Published: 2025-10-23 Model RobustnessLarge Language Model攻撃手法評価 2025.10.23 2025.10.25 Literature Database
SAID: Empowering Large Language Models with Self-Activating Internal Defense Authors: Yulong Chen, Yadong Liu, Jiawen Zhang, Mu Li, Chao Huang, Jie Wen | Published: 2025-10-23 Prompt InjectionModel RobustnessLarge Language Model 2025.10.23 2025.10.25 Literature Database
HarmNet: A Framework for Adaptive Multi-Turn Jailbreak Attacks on Large Language Models Authors: Sidhant Narula, Javad Rafiei Asl, Mohammad Ghasemigol, Eduardo Blanco, Daniel Takabi | Published: 2025-10-21 Query Generation MethodLarge Language Model脱獄手法 2025.10.21 2025.10.23 Literature Database
VERA-V: Variational Inference Framework for Jailbreaking Vision-Language Models Authors: Qilin Liao, Anamika Lochab, Ruqi Zhang | Published: 2025-10-20 Model DoSLarge Language ModelUntargeted Toxicity Attack 2025.10.20 2025.10.22 Literature Database
CrossGuard: Safeguarding MLLMs against Joint-Modal Implicit Malicious Attacks Authors: Xu Zhang, Hao Li, Zhichao Lu | Published: 2025-10-20 Query Generation MethodPrompt InjectionLarge Language Model 2025.10.20 2025.10.22 Literature Database
SoK: Taxonomy and Evaluation of Prompt Security in Large Language Models Authors: Hanbin Hong, Shuya Feng, Nima Naderloui, Shenao Yan, Jingyu Zhang, Biying Liu, Ali Arastehfard, Heqing Huang, Yuan Hong | Published: 2025-10-17 | Updated: 2025-10-21 LLM Securityシナリオベースの悪用Large Language Model 2025.10.17 2025.10.23 Literature Database
Are My Optimized Prompts Compromised? Exploring Vulnerabilities of LLM-based Optimizers Authors: Andrew Zhao, Reshmi Ghosh, Vitor Carvalho, Emily Lawton, Keegan Hines, Gao Huang, Jack W. Stokes | Published: 2025-10-16 Prompt InjectionPrompt leakingLarge Language Model 2025.10.16 2025.10.18 Literature Database
Terrarium: Revisiting the Blackboard for Multi-Agent Safety, Privacy, and Security Studies Authors: Mason Nakamura, Abhinav Kumar, Saaduddin Mahmud, Sahar Abdelnabi, Shlomo Zilberstein, Eugene Bagdasarian | Published: 2025-10-16 エージェント設計Large Language Model通信プロトコル 2025.10.16 2025.10.18 Literature Database