Pelta: Shielding Transformers to Mitigate Evasion Attacks in Federated Learning

Authors: Simon Queyrut, Yérom-David Bromberg, Valerio Schiavoni | Published: 2023-08-08

Accurate, Explainable, and Private Models: Providing Recourse While Minimizing Training Data Leakage

Authors: Catherine Huang, Chelse Swoopes, Christina Xiao, Jiaqi Ma, Himabindu Lakkaraju | Published: 2023-08-08

Breaking Speaker Recognition with PaddingBack

Authors: Zhe Ye, Diqun Yan, Li Dong, Kailai Shen | Published: 2023-08-08 | Updated: 2024-03-11

Backdoor Federated Learning by Poisoning Backdoor-Critical Layers

Authors: Haomin Zhuang, Mingxian Yu, Hao Wang, Yang Hua, Jian Li, Xu Yuan | Published: 2023-08-08 | Updated: 2024-04-15

Labeling without Seeing? Blind Annotation for Privacy-Preserving Entity Resolution

Authors: Yixiang Yao, Weizhao Jin, Srivatsan Ravi | Published: 2023-08-07

“Do Anything Now”: Characterizing and Evaluating In-The-Wild Jailbreak Prompts on Large Language Models

Authors: Xinyue Shen, Zeyuan Chen, Michael Backes, Yun Shen, Yang Zhang | Published: 2023-08-07 | Updated: 2024-05-15

When Federated Learning meets Watermarking: A Comprehensive Overview of Techniques for Intellectual Property Protection

Authors: Mohammed Lansari, Reda Bellafqira, Katarzyna Kapusta, Vincent Thouvenot, Olivier Bettan, Gouenou Coatrieux | Published: 2023-08-07

Mondrian: Prompt Abstraction Attack Against Large Language Models for Cheaper API Pricing

Authors: Wai Man Si, Michael Backes, Yang Zhang | Published: 2023-08-07

PURL: Safe and Effective Sanitization of Link Decoration

Authors: Shaoor Munir, Patrick Lee, Umar Iqbal, Zubair Shafiq, Sandra Siby | Published: 2023-08-07 | Updated: 2024-03-06

A reading survey on adversarial machine learning: Adversarial attacks and their understanding

Authors: Shashank Kotyan | Published: 2023-08-07