MetaDefense: Defending Finetuning-based Jailbreak Attack Before and During Generation Authors: Weisen Jiang, Sinno Jialin Pan | Published: 2025-10-09 Prompt InjectionRobustnessDefense Mechanism 2025.10.09 2025.10.11 Literature Database
Proactive defense against LLM Jailbreak Authors: Weiliang Zhao, Jinjun Peng, Daniel Ben-Levi, Zhou Yu, Junfeng Yang | Published: 2025-10-06 Disabling Safety Mechanisms of LLMPrompt Injection防御手法の統合 2025.10.06 2025.10.08 Literature Database
P2P: A Poison-to-Poison Remedy for Reliable Backdoor Defense in LLMs Authors: Shuai Zhao, Xinyi Wu, Shiqian Zhao, Xiaobao Wu, Zhongliang Guo, Yanhao Jia, Anh Tuan Luu | Published: 2025-10-06 Prompt InjectionPrompt validation防御手法の統合 2025.10.06 2025.10.08 Literature Database
NEXUS: Network Exploration for eXploiting Unsafe Sequences in Multi-Turn LLM Jailbreaks Authors: Javad Rafiei Asl, Sidhant Narula, Mohammad Ghasemigol, Eduardo Blanco, Daniel Takabi | Published: 2025-10-03 | Updated: 2025-10-21 Prompt InjectionLarge Language Model脱獄手法 2025.10.03 2025.10.23 Literature Database
FalseCrashReducer: Mitigating False Positive Crashes in OSS-Fuzz-Gen Using Agentic AI Authors: Paschal C. Amusuo, Dongge Liu, Ricardo Andres Calvo Mendez, Jonathan Metzman, Oliver Chang, James C. Davis | Published: 2025-10-02 Program AnalysisPrompt Injection誤検知管理 2025.10.02 2025.10.04 Literature Database
Bypassing Prompt Guards in Production with Controlled-Release Prompting Authors: Jaiden Fairoze, Sanjam Garg, Keewoo Lee, Mingyuan Wang | Published: 2025-10-02 Prompt InjectionLarge Language ModelStructural Attack 2025.10.02 2025.10.04 Literature Database
Fingerprinting LLMs via Prompt Injection Authors: Yuepeng Hu, Zhengyuan Jiang, Mengyuan Li, Osama Ahmed, Zhicong Huang, Cheng Hong, Neil Gong | Published: 2025-09-29 | Updated: 2025-10-01 Indirect Prompt InjectionToken Identification MethodPrompt Injection 2025.09.29 2025.10.03 Literature Database
MaskSQL: Safeguarding Privacy for LLM-Based Text-to-SQL via Abstraction Authors: Sepideh Abedini, Shubhankar Mohapatra, D. B. Emerson, Masoumeh Shafieinejad, Jesse C. Cresswell, Xi He | Published: 2025-09-27 | Updated: 2025-09-30 SQLクエリ生成Prompt InjectionPrompt leaking 2025.09.27 2025.10.02 Literature Database
RLCracker: Exposing the Vulnerability of LLM Watermarks with Adaptive RL Attacks Authors: Hanbo Huang, Yiran Zhang, Hao Zheng, Xuan Gong, Yihan Li, Lin Liu, Shiyu Liang | Published: 2025-09-25 Disabling Safety Mechanisms of LLMPrompt InjectionWatermark Design 2025.09.25 2025.09.27 Literature Database
Can Federated Learning Safeguard Private Data in LLM Training? Vulnerabilities, Attacks, and Defense Evaluation Authors: Wenkai Guo, Xuefeng Liu, Haolin Wang, Jianwei Niu, Shaojie Tang, Jing Yuan | Published: 2025-09-25 Privacy Protection MethodPrompt InjectionPoisoning 2025.09.25 2025.09.27 Literature Database