FlipLLM: Efficient Bit-Flip Attacks on Multimodal LLMs using Reinforcement Learning Authors: Khurram Khalil, Khaza Anuarul Hoque | Published: 2025-12-10 Prompt InjectionLarge Language ModelVulnerability Assessment Method 2025.12.10 2025.12.12 Literature Database
Attention is All You Need to Defend Against Indirect Prompt Injection Attacks in LLMs Authors: Yinan Zhong, Qianhao Miao, Yanjiao Chen, Jiangyi Deng, Yushi Cheng, Wenyuan Xu | Published: 2025-12-09 Indirect Prompt InjectionPrompt validationLarge Language Model 2025.12.09 2025.12.11 Literature Database
SoK: a Comprehensive Causality Analysis Framework for Large Language Model Security Authors: Wei Zhao, Zhe Li, Jun Sun | Published: 2025-12-04 Prompt Injection因果推論Large Language Model 2025.12.04 2025.12.06 Literature Database
Benchmarking and Understanding Safety Risks in AI Character Platforms Authors: Yiluo Wei, Peixian Zhang, Gareth Tyson | Published: 2025-12-01 キャラクターのメタデータ収集Risk AssessmentLarge Language Model 2025.12.01 2025.12.03 Literature Database
Adversarial Attack-Defense Co-Evolution for LLM Safety Alignment via Tree-Group Dual-Aware Search and Optimization Authors: Xurui Li, Kaisong Song, Rui Zhu, Pin-Yu Chen, Haixu Tang | Published: 2025-11-24 Prompt InjectionLarge Language ModelMalicious Prompt 2025.11.24 2025.11.26 Literature Database
Defending Large Language Models Against Jailbreak Exploits with Responsible AI Considerations Authors: Ryan Wong, Hosea David Yu Fei Ng, Dhananjai Sharma, Glenn Jun Jie Ng, Kavishvaran Srinivasan | Published: 2025-11-24 Ethical ConsiderationsLarge Language ModelMalicious Prompt 2025.11.24 2025.11.26 Literature Database
Can MLLMs Detect Phishing? A Comprehensive Security Benchmark Suite Focusing on Dynamic Threats and Multimodal Evaluation in Academic Environments Authors: Jingzhuo Zhou | Published: 2025-11-19 Privacy Risk ManagementPrompt InjectionLarge Language Model 2025.11.19 2025.11.21 Literature Database
Beyond Fixed and Dynamic Prompts: Embedded Jailbreak Templates for Advancing LLM Security Authors: Hajun Kim, Hyunsik Na, Daeseon Choi | Published: 2025-11-18 Prompt EngineeringLarge Language ModelMalicious Prompt 2025.11.18 2025.11.20 Literature Database
ForgeDAN: An Evolutionary Framework for Jailbreaking Aligned Large Language Models Authors: Siyang Cheng, Gaotian Liu, Rui Mei, Yilin Wang, Kejia Zhang, Kaishuo Wei, Yuqi Yu, Weiping Wen, Xiaojie Wu, Junhua Liu | Published: 2025-11-17 Prompt InjectionLarge Language ModelEvolutionary Algorithm 2025.11.17 2025.11.19 Literature Database
Specification-Guided Vulnerability Detection with Large Language Models Authors: Hao Zhu, Jia Li, Cuiyun Gao, Jiaru Qian, Yihong Dong, Huanyu Liu, Lecheng Wang, Ziliang Wang, Xiaolong Hu, Ge Li | Published: 2025-11-06 Prompt InjectionLarge Language Model脆弱性検出手法 2025.11.06 2025.11.08 Literature Database