Immunity memory-based jailbreak detection: multi-agent adaptive guard for large language models Authors: Jun Leng, Litian Zhang, Xi Zhang | Published: 2025-12-03 Prompt Injectionメモリ化メカニズムAttack Detection Method 2025.12.03 2025.12.05 Literature Database
The Landscape of Memorization in LLMs: Mechanisms, Measurement, and Mitigation Authors: Alexander Xiong, Xuandong Zhao, Aneesh Pappu, Dawn Song | Published: 2025-07-08 Prompt leakingメモリ化メカニズムLarge Language Model 2025.07.08 2025.07.10 Literature Database