Adversarial Surrogate Risk Bounds for Binary Classification

Authors: Natalie S. Frank | Published: 2025-06-11

Design Patterns for Securing LLM Agents against Prompt Injections

Authors: Luca Beurer-Kellner, Beat Buesser Ana-Maria Creţu, Edoardo Debenedetti, Daniel Dobos, Daniel Fabian, Marc Fischer, David Froelicher, Kathrin Grosse, Daniel Naeff, Ezinwanne Ozoani, Andrew Paverd, Florian Tramèr, Václav Volhejn | Published: 2025-06-10 | Updated: 2025-06-11

Your Agent Can Defend Itself against Backdoor Attacks

Authors: Li Changjiang, Liang Jiacheng, Cao Bochuan, Chen Jinghui, Wang Ting | Published: 2025-06-10 | Updated: 2025-06-11

Understanding the Error Sensitivity of Privacy-Aware Computing

Authors: Matías Mazzanti, Esteban Mocskos, Augusto Vega, Pradip Bose | Published: 2025-06-09

TokenBreak: Bypassing Text Classification Models Through Token Manipulation

Authors: Kasimir Schulz, Kenneth Yeung, Kieran Evans | Published: 2025-06-09

Secure Distributed Learning for CAVs: Defending Against Gradient Leakage with Leveled Homomorphic Encryption

Authors: Muhammad Ali Najjar, Ren-Yi Huang, Dumindu Samaraweera, Prashant Shekhar | Published: 2025-06-09

SoK: Data Reconstruction Attacks Against Machine Learning Models: Definition, Metrics, and Benchmark

Authors: Rui Wen, Yiyong Liu, Michael Backes, Yang Zhang | Published: 2025-06-09

Are Trees Really Green? A Detection Approach of IoT Malware Attacks

Authors: Silvia Lucia Sanna, Diego Soi, Davide Maiorca, Giorgio Giacinto | Published: 2025-06-09

Enhancing Adversarial Robustness with Conformal Prediction: A Framework for Guaranteed Model Reliability

Authors: Jie Bao, Chuangyin Dang, Rui Luo, Hanwei Zhang, Zhixin Zhou | Published: 2025-06-09

LLM Unlearning Should Be Form-Independent

Authors: Xiaotian Ye, Mengqi Zhang, Shu Wu | Published: 2025-06-09