AIセキュリティポータルbot

Design Patterns for Securing LLM Agents against Prompt Injections

Authors: Luca Beurer-Kellner, Beat Buesser Ana-Maria Creţu, Edoardo Debenedetti, Daniel Dobos, Daniel Fabian, Marc Fischer, David Froelicher, Kathrin Grosse, Daniel Naeff, Ezinwanne Ozoani, Andrew Paverd, Florian Tramèr, Václav Volhejn | Published: 2025-06-10 | Updated: 2025-06-11
Indirect Prompt Injection
Prompt Injection
Defense Method

Your Agent Can Defend Itself against Backdoor Attacks

Authors: Li Changjiang, Liang Jiacheng, Cao Bochuan, Chen Jinghui, Wang Ting | Published: 2025-06-10 | Updated: 2025-06-11
Poisoning attack on RAG
Backdoor Attack Mitigation
Defense Method

Understanding the Error Sensitivity of Privacy-Aware Computing

Authors: Matías Mazzanti, Esteban Mocskos, Augusto Vega, Pradip Bose | Published: 2025-06-09
Error Correction Code
Privacy Protection Mechanism
Encryption Technology

TokenBreak: Bypassing Text Classification Models Through Token Manipulation

Authors: Kasimir Schulz, Kenneth Yeung, Kieran Evans | Published: 2025-06-09
Adversarial Attack Methods
Defense Method

Secure Distributed Learning for CAVs: Defending Against Gradient Leakage with Leveled Homomorphic Encryption

Authors: Muhammad Ali Najjar, Ren-Yi Huang, Dumindu Samaraweera, Prashant Shekhar | Published: 2025-06-09
Privacy Protection Mechanism
Encryption Technology
Federated Learning

SoK: Data Reconstruction Attacks Against Machine Learning Models: Definition, Metrics, and Benchmark

Authors: Rui Wen, Yiyong Liu, Michael Backes, Yang Zhang | Published: 2025-06-09
Model DoS
再構成アルゴリズム
評価メトリクス

Are Trees Really Green? A Detection Approach of IoT Malware Attacks

Authors: Silvia Lucia Sanna, Diego Soi, Davide Maiorca, Giorgio Giacinto | Published: 2025-06-09
IoTネットワーク攻撃
Advancements in Medical IoT
Optimization Strategy

Enhancing Adversarial Robustness with Conformal Prediction: A Framework for Guaranteed Model Reliability

Authors: Jie Bao, Chuangyin Dang, Rui Luo, Hanwei Zhang, Zhixin Zhou | Published: 2025-06-09
Certified Robustness
Robust Optimization
Adversarial Attack Methods

LLM Unlearning Should Be Form-Independent

Authors: Xiaotian Ye, Mengqi Zhang, Shu Wu | Published: 2025-06-09
Training Method
Certified Robustness
非意味的リダイレクション

“I wasn’t sure if this is indeed a security risk”: Data-driven Understanding of Security Issue Reporting in GitHub Repositories of Open Source npm Packages

Authors: Rajdeep Ghosh, Shiladitya De, Mainack Mondal | Published: 2025-06-09
API Security
Cyber Threat
Information Leakage Analysis