文献データベース

MCP Safety Training: Learning to Refuse Falsely Benign MCP Exploits using Improved Preference Alignment

Authors: John Halloran | Published: 2025-05-29
RAGへのポイズニング攻撃
アライメント
料理材料

Merge Hijacking: Backdoor Attacks to Model Merging of Large Language Models

Authors: Zenghui Yuan, Yangming Xu, Jiawen Shi, Pan Zhou, Lichao Sun | Published: 2025-05-29
LLMセキュリティ
ポイズニング攻撃
モデル保護手法

Disrupting Vision-Language Model-Driven Navigation Services via Adversarial Object Fusion

Authors: Chunlong Xie, Jialing He, Shangwei Guo, Jiacheng Wang, Shudong Zhang, Tianwei Zhang, Tao Xiang | Published: 2025-05-29
アライメント
敵対的オブジェクト生成
最適化手法

SimProcess: High Fidelity Simulation of Noisy ICS Physical Processes

Authors: Denis Donadel, Gabriele Crestanello, Giulio Morandini, Daniele Antonioli, Mauro Conti, Massimo Merro | Published: 2025-05-28
データの起源と変遷
モデル設計
動的分析手法

Transformers for Secure Hardware Systems: Applications, Challenges, and Outlook

Authors: Banafsheh Saber Latibari, Najmeh Nazari, Avesta Sasan, Houman Homayoun, Pratik Satam, Soheil Salehi, Hossein Sayadi | Published: 2025-05-28
セキュリティ分析
ハードウェアトロイの検出
バックドアモデルの検知

Does Johnny Get the Message? Evaluating Cybersecurity Notifications for Everyday Users

Authors: Victor Jüttner, Erik Buchmann | Published: 2025-05-28
パーソナライズ
プロンプトインジェクション
対策の説明

Test-Time Immunization: A Universal Defense Framework Against Jailbreaks for (Multimodal) Large Language Models

Authors: Yongcan Yu, Yanbo Wang, Ran He, Jian Liang | Published: 2025-05-28
LLMセキュリティ
プロンプトインジェクション
大規模言語モデル

Jailbreak Distillation: Renewable Safety Benchmarking

Authors: Jingyu Zhang, Ahmed Elgohary, Xiawei Wang, A S M Iftekhar, Ahmed Magooda, Benjamin Van Durme, Daniel Khashabi, Kyle Jackson | Published: 2025-05-28
プロンプトインジェクション
モデル評価
攻撃の評価

VulBinLLM: LLM-powered Vulnerability Detection for Stripped Binaries

Authors: Nasir Hussain, Haohan Chen, Chanh Tran, Philip Huang, Zhuohao Li, Pravir Chugh, William Chen, Ashish Kundu, Yuan Tian | Published: 2025-05-28
LLMセキュリティ
脆弱性分析
逆アセンブル

Breaking the Ceiling: Exploring the Potential of Jailbreak Attacks through Expanding Strategy Space

Authors: Yao Huang, Yitong Sun, Shouwei Ruan, Yichi Zhang, Yinpeng Dong, Xingxing Wei | Published: 2025-05-27
LLMの安全機構の解除
プロンプトインジェクション
攻撃の評価