マルチターン攻撃分析

David vs. Goliath: Verifiable Agent-to-Agent Jailbreaking via Reinforcement Learning

Authors: Samuel Nellessen, Tal Kachman | Published: 2026-02-02
Indirect Prompt Injection
マルチターン攻撃分析
Reinforcement Learning Attack

NeuroFilter: Privacy Guardrails for Conversational LLM Agents

Authors: Saswat Das, Ferdinando Fioretto | Published: 2026-01-21
Privacy Protection
Prompt Injection
マルチターン攻撃分析

Let the Bees Find the Weak Spots: A Path Planning Perspective on Multi-Turn Jailbreak Attacks against LLMs

Authors: Yize Liu, Yunyun Hou, Aina Sui | Published: 2025-11-05
Automation of Cybersecurity
Prompt Injection
マルチターン攻撃分析