Prompt Injection

NeuroFilter: Privacy Guardrails for Conversational LLM Agents

Authors: Saswat Das, Ferdinando Fioretto | Published: 2026-01-21
Privacy Protection
Prompt Injection
マルチターン攻撃分析

Be Your Own Red Teamer: Safety Alignment via Self-Play and Reflective Experience Replay

Authors: Hao Wang, Yanting Wang, Hao Li, Rui Li, Lei Sha | Published: 2026-01-15
Prompt Injection
Adversarial Attack Analysis
Self-Learning Method

ReasAlign: Reasoning Enhanced Safety Alignment against Prompt Injection Attack

Authors: Hao Li, Yankai Yang, G. Edward Suh, Ning Zhang, Chaowei Xiao | Published: 2026-01-15
LLM活用
Indirect Prompt Injection
Prompt Injection

The Promptware Kill Chain: How Prompt Injections Gradually Evolved Into a Multi-Step Malware

Authors: Ben Nassi, Bruce Schneier, Oleg Brodt | Published: 2026-01-14
LLM活用
Indirect Prompt Injection
Prompt Injection

SpatialJB: How Text Distribution Art Becomes the “Jailbreak Key” for LLM Guardrails

Authors: Zhiyi Mou, Jingyuan Yang, Zeheng Qian, Wangze Ni, Tianfang Xiao, Ning Liu, Chen Zhang, Zhan Qin, Kui Ren | Published: 2026-01-14
LLM活用
Prompt Injection
Large Language Model

A Decompilation-Driven Framework for Malware Detection with Large Language Models

Authors: Aniesh Chawla, Udbhav Prasad | Published: 2026-01-14
LLM活用
Prompt Injection
Multi-Domain Data Analysis

Evaluating Implicit Regulatory Compliance in LLM Tool Invocation via Logic-Guided Synthesis

Authors: Da Song, Yuheng Huang, Boqi Chen, Tianshuo Cong, Randy Goebel, Lei Ma, Foutse Khomh | Published: 2026-01-13
LLM活用
Prompt Injection
Performance Evaluation

SecureCAI: Injection-Resilient LLM Assistants for Cybersecurity Operations

Authors: Mohammed Himayath Ali, Mohammed Aqib Abdullah, Mohammed Mudassir Uddin, Shahnawaz Alam | Published: 2026-01-12
Indirect Prompt Injection
Prompt Injection
Adversarial Attack Analysis

When Bots Take the Bait: Exposing and Mitigating the Emerging Social Engineering Attack in Web Automation Agent

Authors: Xinyi Wu, Geng Hong, Yueyue Chen, MingXuan Liu, Feier Jin, Xudong Pan, Jiarun Dai, Baojun Liu | Published: 2026-01-12
Indirect Prompt Injection
Prompt Injection
User Behavior Analysis

Safe-FedLLM: Delving into the Safety of Federated Large Language Models

Authors: Mingxiang Tao, Yu Tian, Wenxuan Tu, Yue Yang, Xue Yang, Xiangyan Tang | Published: 2026-01-12
Security Assurance
Privacy Protection
Prompt Injection