Prompt Injection

Clouding the Mirror: Stealthy Prompt Injection Attacks Targeting LLM-based Phishing Detection

Authors: Takashi Koide, Hiroki Nakano, Daiki Chiba | Published: 2026-02-05
Indirect Prompt Injection
フィッシング検出手法
Prompt Injection

How Few-shot Demonstrations Affect Prompt-based Defenses Against LLM Jailbreak Attacks

Authors: Yanshu Wang, Shuaishuai Yang, Jingjing He, Tong Yang | Published: 2026-02-04
LLM Performance Evaluation
Prompt Injection
Large Language Model

RedVisor: Reasoning-Aware Prompt Injection Defense via Zero-Copy KV Cache Reuse

Authors: Mingrui Liu, Sixiao Zhang, Cheng Long, Kwok-Yan Lam | Published: 2026-02-02
Framework
Prompt Injection
防御手法の統合

Eliciting Least-to-Most Reasoning for Phishing URL Detection

Authors: Holly Trikilis, Pasindu Marasinghe, Fariza Rashid, Suranga Seneviratne | Published: 2026-01-28
LLM Performance Evaluation
Prompt Injection
Prompt leaking

$α^3$-SecBench: A Large-Scale Evaluation Suite of Security, Resilience, and Trust for LLM-based UAV Agents over 6G Networks

Authors: Mohamed Amine Ferrag, Abderrahmane Lakas, Merouane Debbah | Published: 2026-01-26
LLM活用
Security Assurance
Prompt Injection

NeuroFilter: Privacy Guardrails for Conversational LLM Agents

Authors: Saswat Das, Ferdinando Fioretto | Published: 2026-01-21
Privacy Protection
Prompt Injection
マルチターン攻撃分析

Be Your Own Red Teamer: Safety Alignment via Self-Play and Reflective Experience Replay

Authors: Hao Wang, Yanting Wang, Hao Li, Rui Li, Lei Sha | Published: 2026-01-15
Prompt Injection
Adversarial Attack Analysis
Self-Learning Method

ReasAlign: Reasoning Enhanced Safety Alignment against Prompt Injection Attack

Authors: Hao Li, Yankai Yang, G. Edward Suh, Ning Zhang, Chaowei Xiao | Published: 2026-01-15
LLM活用
Indirect Prompt Injection
Prompt Injection

The Promptware Kill Chain: How Prompt Injections Gradually Evolved Into a Multi-Step Malware

Authors: Ben Nassi, Bruce Schneier, Oleg Brodt | Published: 2026-01-14
LLM活用
Indirect Prompt Injection
Prompt Injection

SpatialJB: How Text Distribution Art Becomes the “Jailbreak Key” for LLM Guardrails

Authors: Zhiyi Mou, Jingyuan Yang, Zeheng Qian, Wangze Ni, Tianfang Xiao, Ning Liu, Chen Zhang, Zhan Qin, Kui Ren | Published: 2026-01-14
LLM活用
Prompt Injection
Large Language Model