LLMセキュリティ

Securing RAG: A Risk Assessment and Mitigation Framework

Authors: Lukas Ammann, Sara Ott, Christoph R. Landolt, Marco P. Lehmann | Published: 2025-05-13
LLMセキュリティ
RAG
RAGへのポイズニング攻撃

SecReEvalBench: A Multi-turned Security Resilience Evaluation Benchmark for Large Language Models

Authors: Huining Cui, Wei Liu | Published: 2025-05-12
LLMセキュリティ
プロンプトインジェクション
プロンプトリーキング

Private LoRA Fine-tuning of Open-Source LLMs with Homomorphic Encryption

Authors: Jordan Frery, Roman Bredehoft, Jakub Klemsa, Arthur Meyre, Andrei Stoian | Published: 2025-05-12
LLMセキュリティ
暗号技術
機械学習技術

One Trigger Token Is Enough: A Defense Strategy for Balancing Safety and Usability in Large Language Models

Authors: Haoran Gu, Handing Wang, Yi Mei, Mengjie Zhang, Yaochu Jin | Published: 2025-05-12
LLMセキュリティ
LLMの安全機構の解除
プロンプトインジェクション

LLM-Text Watermarking based on Lagrange Interpolation

Authors: Jarosław Janas, Paweł Morawiecki, Josef Pieprzyk | Published: 2025-05-09 | Updated: 2025-05-12
LLMセキュリティ
プロンプトリーキング
生成AI向け電子透かし

Red Teaming the Mind of the Machine: A Systematic Evaluation of Prompt Injection and Jailbreak Vulnerabilities in LLMs

Authors: Chetan Pathade | Published: 2025-05-07 | Updated: 2025-05-13
LLMセキュリティ
LLMの安全機構の解除
プロンプトインジェクション

Weaponizing Language Models for Cybersecurity Offensive Operations: Automating Vulnerability Assessment Report Validation; A Review Paper

Authors: Abdulrahman S Almuhaidib, Azlan Mohd Zain, Zalmiyah Zakaria, Izyan Izzati Kamsani, Abdulaziz S Almuhaidib | Published: 2025-05-07
LLMセキュリティ
脆弱性分析

LLMs’ Suitability for Network Security: A Case Study of STRIDE Threat Modeling

Authors: AbdulAziz AbdulGhaffar, Ashraf Matrawy | Published: 2025-05-07
LLMセキュリティ
性能評価
脆弱性分析

LlamaFirewall: An open source guardrail system for building secure AI agents

Authors: Sahana Chennabasappa, Cyrus Nikolaidis, Daniel Song, David Molnar, Stephanie Ding, Shengye Wan, Spencer Whitman, Lauren Deason, Nicholas Doucette, Abraham Montilla, Alekhya Gampa, Beto de Paola, Dominik Gabi, James Crnkovich, Jean-Christophe Testud, Kat He, Rashnil Chaturvedi, Wu Zhou, Joshua Saxe | Published: 2025-05-06
LLMセキュリティ
アライメント
プロンプトインジェクション

Output Constraints as Attack Surface: Exploiting Structured Generation to Bypass LLM Safety Mechanisms

Authors: Shuoming Zhang, Jiacheng Zhao, Ruiyuan Xu, Xiaobing Feng, Huimin Cui | Published: 2025-03-31
LLMセキュリティ
LLMの安全機構の解除
プロンプトインジェクション