Beyond Jailbreaks: Revealing Stealthier and Broader LLM Security Risks Stemming from Alignment Failures Authors: Yukai Zhou, Sibei Yang, Wenjie Wang | Published: 2025-06-09 LLMとの協力効果サイバー脅威大規模言語モデル 2025.06.09 文献データベース
TracLLM: A Generic Framework for Attributing Long Context LLMs Authors: Yanting Wang, Wei Zou, Runpeng Geng, Jinyuan Jia | Published: 2025-06-04 LLMとの協力効果RAGへのポイズニング攻撃効率評価 2025.06.04 文献データベース
Bridging Expertise Gaps: The Role of LLMs in Human-AI Collaboration for Cybersecurity Authors: Shahroz Tariq, Ronal Singh, Mohan Baruwal Chhetri, Surya Nepal, Cecile Paris | Published: 2025-05-06 LLMとの協力効果アライメント参加者の質問分析 2025.05.06 文献データベース
Knowledge-to-Jailbreak: Investigating Knowledge-driven Jailbreaking Attacks for Large Language Models Authors: Shangqing Tu, Zhuoran Pan, Wenxuan Wang, Zhexin Zhang, Yuliang Sun, Jifan Yu, Hongning Wang, Lei Hou, Juanzi Li | Published: 2024-06-17 | Updated: 2025-06-09 LLMとの協力効果プロンプトインジェクション大規模言語モデル 2024.06.17 文献データベース