David vs. Goliath: Verifiable Agent-to-Agent Jailbreaking via Reinforcement Learning Authors: Samuel Nellessen, Tal Kachman | Published: 2026-02-02 Indirect Prompt Injectionマルチターン攻撃分析Reinforcement Learning Attack 2026.02.02 2026.02.04 Literature Database
NeuroFilter: Privacy Guardrails for Conversational LLM Agents Authors: Saswat Das, Ferdinando Fioretto | Published: 2026-01-21 Privacy ProtectionPrompt Injectionマルチターン攻撃分析 2026.01.21 2026.01.23 Literature Database
Let the Bees Find the Weak Spots: A Path Planning Perspective on Multi-Turn Jailbreak Attacks against LLMs Authors: Yize Liu, Yunyun Hou, Aina Sui | Published: 2025-11-05 Automation of CybersecurityPrompt Injectionマルチターン攻撃分析 2025.11.05 2025.11.07 Literature Database