HarmNet: A Framework for Adaptive Multi-Turn Jailbreak Attacks on Large Language Models Authors: Sidhant Narula, Javad Rafiei Asl, Mohammad Ghasemigol, Eduardo Blanco, Daniel Takabi | Published: 2025-10-21 Query Generation MethodLarge Language Model脱獄手法 2025.10.21 2025.10.23 Literature Database
PLAGUE: Plug-and-play framework for Lifelong Adaptive Generation of Multi-turn Exploits Authors: Neeladri Bhuiya, Madhav Aggarwal, Diptanshu Purwar | Published: 2025-10-20 | Updated: 2025-10-22 Poisoning attack on RAG攻撃者ターゲット更新脱獄手法 2025.10.20 2025.10.24 Literature Database
NEXUS: Network Exploration for eXploiting Unsafe Sequences in Multi-Turn LLM Jailbreaks Authors: Javad Rafiei Asl, Sidhant Narula, Mohammad Ghasemigol, Eduardo Blanco, Daniel Takabi | Published: 2025-10-03 | Updated: 2025-10-21 Prompt InjectionLarge Language Model脱獄手法 2025.10.03 2025.10.23 Literature Database
GuidedBench: Measuring and Mitigating the Evaluation Discrepancies of In-the-wild LLM Jailbreak Methods Authors: Ruixuan Huang, Xunguang Wang, Zongjie Li, Daoyuan Wu, Shuai Wang | Published: 2025-02-24 | Updated: 2025-07-09 Prompt Injection脱獄手法Evaluation Method 2025.02.24 2025.07.11 Literature Database
TombRaider: Entering the Vault of History to Jailbreak Large Language Models Authors: Junchen Ding, Jiahao Zhang, Yi Liu, Ziqi Ding, Gelei Deng, Yuekang Li | Published: 2025-01-27 | Updated: 2025-08-25 Prompt InjectionPrompt leaking脱獄手法 2025.01.27 2025.08.27 Literature Database