SpatialJB: How Text Distribution Art Becomes the “Jailbreak Key” for LLM Guardrails Authors: Zhiyi Mou, Jingyuan Yang, Zeheng Qian, Wangze Ni, Tianfang Xiao, Ning Liu, Chen Zhang, Zhan Qin, Kui Ren | Published: 2026-01-14 LLM活用Prompt InjectionLarge Language Model 2026.01.14 2026.01.16 Literature Database
HoneyTrap: Deceiving Large Language Model Attackers to Honeypot Traps with Resilient Multi-Agent Defense Authors: Siyuan Li, Xi Lin, Jun Wu, Zehao Liu, Haoyu Li, Tianjie Ju, Xiang Chen, Jianhua Li | Published: 2026-01-07 Prompt InjectionLarge Language ModelAdversarial Attack Detection 2026.01.07 2026.01.09 Literature Database
Jailbreaking LLMs & VLMs: Mechanisms, Evaluation, and Unified Defense Authors: Zejian Chen, Chaozhuo Li, Chao Li, Xi Zhang, Litian Zhang, Yiming He | Published: 2026-01-07 Prompt InjectionLarge Language ModelAdversarial Attack Detection 2026.01.07 2026.01.09 Literature Database
On the Effectiveness of Instruction-Tuning Local LLMs for Identifying Software Vulnerabilities Authors: Sangryu Park, Gihyuk Ko, Homook Cho | Published: 2025-12-23 Prompt InjectionLarge Language ModelVulnerability Analysis 2025.12.23 2025.12.25 Literature Database
Large Language Models as a (Bad) Security Norm in the Context of Regulation and Compliance Authors: Kaspar Rosager Ludvigsen | Published: 2025-12-18 LLM活用Indirect Prompt InjectionLarge Language Model 2025.12.18 2025.12.20 Literature Database
FlipLLM: Efficient Bit-Flip Attacks on Multimodal LLMs using Reinforcement Learning Authors: Khurram Khalil, Khaza Anuarul Hoque | Published: 2025-12-10 Prompt InjectionLarge Language ModelVulnerability Assessment Method 2025.12.10 2025.12.12 Literature Database
Attention is All You Need to Defend Against Indirect Prompt Injection Attacks in LLMs Authors: Yinan Zhong, Qianhao Miao, Yanjiao Chen, Jiangyi Deng, Yushi Cheng, Wenyuan Xu | Published: 2025-12-09 Indirect Prompt InjectionPrompt validationLarge Language Model 2025.12.09 2025.12.11 Literature Database
SoK: a Comprehensive Causality Analysis Framework for Large Language Model Security Authors: Wei Zhao, Zhe Li, Jun Sun | Published: 2025-12-04 Prompt Injection因果推論Large Language Model 2025.12.04 2025.12.06 Literature Database
Benchmarking and Understanding Safety Risks in AI Character Platforms Authors: Yiluo Wei, Peixian Zhang, Gareth Tyson | Published: 2025-12-01 キャラクターのメタデータ収集Risk AssessmentLarge Language Model 2025.12.01 2025.12.03 Literature Database
Adversarial Attack-Defense Co-Evolution for LLM Safety Alignment via Tree-Group Dual-Aware Search and Optimization Authors: Xurui Li, Kaisong Song, Rui Zhu, Pin-Yu Chen, Haixu Tang | Published: 2025-11-24 Prompt InjectionLarge Language ModelMalicious Prompt 2025.11.24 2025.11.26 Literature Database