ReCIT: Reconstructing Full Private Data from Gradient in Parameter-Efficient Fine-Tuning of Large Language Models

Authors: Jin Xie, Ruishi He, Songze Li, Xiaojun Jia, Shouling Ji | Published: 2025-04-29

Token-Efficient Prompt Injection Attack: Provoking Cessation in LLM Reasoning via Adaptive Token Compression

Authors: Yu Cui, Yujun Cai, Yiwei Wang | Published: 2025-04-29

Robustness via Referencing: Defending against Prompt Injection Attacks by Referencing the Executed Instruction

Authors: Yulin Chen, Haoran Li, Yuan Sui, Yue Liu, Yufei He, Yangqiu Song, Bryan Hooi | Published: 2025-04-29

Network Attack Traffic Detection With Hybrid Quantum-Enhanced Convolution Neural Network

Authors: Zihao Wang, Kar Wai Fok, Vrizlynn L. L. Thing | Published: 2025-04-29

Enhancing Leakage Attacks on Searchable Symmetric Encryption Using LLM-Based Synthetic Data Generation

Authors: Joshua Chiu, Partha Protim Paul, Zahin Wahab | Published: 2025-04-29

The Automation Advantage in AI Red Teaming

Authors: Rob Mulla, Will Pearce, Nick Landers, Brian Greunke, Brad Palm, Vincent Abruzzo, Ads Dawson | Published: 2025-04-28

CodeBC: A More Secure Large Language Model for Smart Contract Code Generation in Blockchain

Authors: Lingxiang Wang, Hainan Zhang, Qinnan Zhang, Ziwei Wang, Hongwei Zheng, Jin Dong, Zhiming Zheng | Published: 2025-04-28 | Updated: 2025-05-07

$\texttt{SAGE}$: A Generic Framework for LLM Safety Evaluation

Authors: Madhur Jindal, Hari Shrawgi, Parag Agrawal, Sandipan Dandapat | Published: 2025-04-28

Can Differentially Private Fine-tuning LLMs Protect Against Privacy Attacks?

Authors: Hao Du, Shang Liu, Yang Cao | Published: 2025-04-28 | Updated: 2025-05-01

BadMoE: Backdooring Mixture-of-Experts LLMs via Optimizing Routing Triggers and Infecting Dormant Experts

Authors: Qingyue Wang, Qi Pang, Xixun Lin, Shuai Wang, Daoyuan Wu | Published: 2025-04-24 | Updated: 2025-04-29