Beyond Jailbreaks: Revealing Stealthier and Broader LLM Security Risks Stemming from Alignment Failures Authors: Yukai Zhou, Sibei Yang, Wenjie Wang | Published: 2025-06-09 Cooperative Effects with LLMCyber ThreatLarge Language Model 2025.06.09 2025.06.11 Literature Database
TracLLM: A Generic Framework for Attributing Long Context LLMs Authors: Yanting Wang, Wei Zou, Runpeng Geng, Jinyuan Jia | Published: 2025-06-04 Cooperative Effects with LLMPoisoning attack on RAGEfficiency Evaluation 2025.06.04 2025.06.06 Literature Database
Bridging Expertise Gaps: The Role of LLMs in Human-AI Collaboration for Cybersecurity Authors: Shahroz Tariq, Ronal Singh, Mohan Baruwal Chhetri, Surya Nepal, Cecile Paris | Published: 2025-05-06 Cooperative Effects with LLMAlignmentParticipant Question Analysis 2025.05.06 2025.05.27 Literature Database
Knowledge-to-Jailbreak: Investigating Knowledge-driven Jailbreaking Attacks for Large Language Models Authors: Shangqing Tu, Zhuoran Pan, Wenxuan Wang, Zhexin Zhang, Yuliang Sun, Jifan Yu, Hongning Wang, Lei Hou, Juanzi Li | Published: 2024-06-17 | Updated: 2025-06-09 Cooperative Effects with LLMPrompt InjectionLarge Language Model 2024.06.17 2025.06.11 Literature Database