Robust LLM safeguarding via refusal feature adversarial training

Authors: Lei Yu, Virginie Do, Karen Hambardzumyan, Nicola Cancedda | Published: 2024-09-30 | Updated: 2025-03-20

The Early Bird Catches the Leak: Unveiling Timing Side Channels in LLM Serving Systems

Authors: Linke Song, Zixuan Pang, Wenhao Wang, Zihao Wang, XiaoFeng Wang, Hongbo Chen, Wei Song, Yier Jin, Dan Meng, Rui Hou | Published: 2024-09-30 | Updated: 2025-08-13

Infighting in the Dark: Multi-Label Backdoor Attack in Federated Learning

Authors: Ye Li, Yanchao Zhao, Chengcheng Zhu, Jiale Zhang | Published: 2024-09-29 | Updated: 2025-03-22

OnePath: Efficient and Privacy-Preserving Decision Tree Inference in the Cloud

Authors: Shuai Yuan, Hongwei Li, Xinyuan Qian, Guowen Xu | Published: 2024-09-28 | Updated: 2025-07-21

Confidential Prompting: Privacy-preserving LLM Inference on Cloud

Authors: Caihua Li, In Gim, Lin Zhong | Published: 2024-09-27 | Updated: 2025-08-25

Enhancing Robustness of Graph Neural Networks through p-Laplacian

Authors: Anuj Kumar Sirohi, Subhanu Halder, Kabir Kumar, Sandeep Kumar | Published: 2024-09-27

System-Level Defense against Indirect Prompt Injection Attacks: An Information Flow Control Perspective

Authors: Fangzhou Wu, Ethan Cecchetti, Chaowei Xiao | Published: 2024-09-27 | Updated: 2024-10-10

Evading Toxicity Detection with ASCII-art: A Benchmark of Spatial Attacks on Moderation Systems

Authors: Sergey Berezin, Reza Farahbakhsh, Noel Crespi | Published: 2024-09-27 | Updated: 2025-09-24

Code Vulnerability Repair with Large Language Model using Context-Aware Prompt Tuning

Authors: Arshiya Khan, Guannan Liu, Xing Gao | Published: 2024-09-27 | Updated: 2025-06-11

An Adversarial Perspective on Machine Unlearning for AI Safety

Authors: Jakub Łucki, Boyi Wei, Yangsibo Huang, Peter Henderson, Florian Tramèr, Javier Rando | Published: 2024-09-26 | Updated: 2025-04-10