Clouding the Mirror: Stealthy Prompt Injection Attacks Targeting LLM-based Phishing Detection Authors: Takashi Koide, Hiroki Nakano, Daiki Chiba | Published: 2026-02-05 Indirect Prompt Injectionフィッシング検出手法Prompt Injection 2026.02.05 2026.02.07 Literature Database
How Few-shot Demonstrations Affect Prompt-based Defenses Against LLM Jailbreak Attacks Authors: Yanshu Wang, Shuaishuai Yang, Jingjing He, Tong Yang | Published: 2026-02-04 LLM Performance EvaluationPrompt InjectionLarge Language Model 2026.02.04 2026.02.06 Literature Database
RedVisor: Reasoning-Aware Prompt Injection Defense via Zero-Copy KV Cache Reuse Authors: Mingrui Liu, Sixiao Zhang, Cheng Long, Kwok-Yan Lam | Published: 2026-02-02 FrameworkPrompt Injection防御手法の統合 2026.02.02 2026.02.04 Literature Database
Eliciting Least-to-Most Reasoning for Phishing URL Detection Authors: Holly Trikilis, Pasindu Marasinghe, Fariza Rashid, Suranga Seneviratne | Published: 2026-01-28 LLM Performance EvaluationPrompt InjectionPrompt leaking 2026.01.28 2026.01.30 Literature Database
$α^3$-SecBench: A Large-Scale Evaluation Suite of Security, Resilience, and Trust for LLM-based UAV Agents over 6G Networks Authors: Mohamed Amine Ferrag, Abderrahmane Lakas, Merouane Debbah | Published: 2026-01-26 LLM活用Security AssurancePrompt Injection 2026.01.26 2026.01.28 Literature Database
NeuroFilter: Privacy Guardrails for Conversational LLM Agents Authors: Saswat Das, Ferdinando Fioretto | Published: 2026-01-21 Privacy ProtectionPrompt Injectionマルチターン攻撃分析 2026.01.21 2026.01.23 Literature Database
Be Your Own Red Teamer: Safety Alignment via Self-Play and Reflective Experience Replay Authors: Hao Wang, Yanting Wang, Hao Li, Rui Li, Lei Sha | Published: 2026-01-15 Prompt InjectionAdversarial Attack AnalysisSelf-Learning Method 2026.01.15 2026.01.17 Literature Database
ReasAlign: Reasoning Enhanced Safety Alignment against Prompt Injection Attack Authors: Hao Li, Yankai Yang, G. Edward Suh, Ning Zhang, Chaowei Xiao | Published: 2026-01-15 LLM活用Indirect Prompt InjectionPrompt Injection 2026.01.15 2026.01.17 Literature Database
The Promptware Kill Chain: How Prompt Injections Gradually Evolved Into a Multi-Step Malware Authors: Ben Nassi, Bruce Schneier, Oleg Brodt | Published: 2026-01-14 LLM活用Indirect Prompt InjectionPrompt Injection 2026.01.14 2026.01.16 Literature Database
SpatialJB: How Text Distribution Art Becomes the “Jailbreak Key” for LLM Guardrails Authors: Zhiyi Mou, Jingyuan Yang, Zeheng Qian, Wangze Ni, Tianfang Xiao, Ning Liu, Chen Zhang, Zhan Qin, Kui Ren | Published: 2026-01-14 LLM活用Prompt InjectionLarge Language Model 2026.01.14 2026.01.16 Literature Database