AIセキュリティポータルbot

TrojanStego: Your Language Model Can Secretly Be A Steganographic Privacy Leaking Agent

Authors: Dominik Meier, Jan Philip Wahle, Paul Röttger, Terry Ruas, Bela Gipp | Published: 2025-05-26
プロンプトインジェクション
モデル抽出攻撃
透かし技術

DFIR-Metric: A Benchmark Dataset for Evaluating Large Language Models in Digital Forensics and Incident Response

Authors: Bilel Cherif, Tamas Bisztray, Richard A. Dubniczky, Aaesha Aldahmani, Saeed Alshehhi, Norbert Tihanyi | Published: 2025-05-26
ハルシネーション
モデル性能評価
評価手法

Differential Privacy Analysis of Decentralized Gossip Averaging under Varying Threat Models

Authors: Antti Koskela, Tejas Kulkarni | Published: 2025-05-26
プライバシー問題
差分プライバシー
透かし技術

Deconstructing Obfuscation: A four-dimensional framework for evaluating Large Language Models assembly code deobfuscation capabilities

Authors: Anton Tkachenko, Dmitrij Suskevic, Benjamin Adolphi | Published: 2025-05-26
モデル評価手法
大規模言語モデル
透かし技術

CPA-RAG:Covert Poisoning Attacks on Retrieval-Augmented Generation in Large Language Models

Authors: Chunyang Li, Junwei Zhang, Anda Cheng, Zhuo Ma, Xinghua Li, Jianfeng Ma | Published: 2025-05-26
RAGへのポイズニング攻撃
テキスト生成手法
ポイズニング攻撃

What Really Matters in Many-Shot Attacks? An Empirical Study of Long-Context Vulnerabilities in LLMs

Authors: Sangyeop Kim, Yohan Lee, Yongwoo Song, Kimin Lee | Published: 2025-05-26
プロンプトインジェクション
モデル性能評価
大規模言語モデル

CoTGuard: Using Chain-of-Thought Triggering for Copyright Protection in Multi-Agent LLM Systems

Authors: Yan Wen, Junfeng Guo, Heng Huang | Published: 2025-05-26
LLMセキュリティ
トリガーベースの透かし
著作権保護

VADER: A Human-Evaluated Benchmark for Vulnerability Assessment, Detection, Explanation, and Remediation

Authors: Ethan TS. Liu, Austin Wang, Spencer Mateega, Carlos Georgescu, Danny Tang | Published: 2025-05-26
ウェブサイト脆弱性
ハルシネーション
動的脆弱性管理

LLM-Driven APT Detection for 6G Wireless Networks: A Systematic Review and Taxonomy

Authors: Muhammed Golec, Yaser Khamayseh, Suhib Bani Melhem, Abdulmalik Alwarafy | Published: 2025-05-24 | Updated: 2025-06-23
LLMとAPTの関係
情報セキュリティ
研究方法論

Invisible Prompts, Visible Threats: Malicious Font Injection in External Resources for Large Language Models

Authors: Junjie Xiong, Changjia Zhu, Shuhang Lin, Chong Zhang, Yongfeng Zhang, Yao Liu, Lingyao Li | Published: 2025-05-22
LLMセキュリティ
LLMの安全機構の解除
プロンプトインジェクション