Adversarial Demonstration Attacks on Large Language Models

Authors: Jiongxiao Wang, Zichen Liu, Keun Hee Park, Zhuojun Jiang, Zhaoheng Zheng, Zhuofeng Wu, Muhao Chen, Chaowei Xiao | Published: 2023-05-24 | Updated: 2023-10-14

Query Rewriting for Retrieval-Augmented Large Language Models

Authors: Xinbei Ma, Yeyun Gong, Pengcheng He, Hai Zhao, Nan Duan | Published: 2023-05-23 | Updated: 2023-10-23

Expressive Losses for Verified Robustness via Convex Combinations

Authors: Alessandro De Palma, Rudy Bunel, Krishnamurthy Dvijotham, M. Pawan Kumar, Robert Stanforth, Alessio Lomuscio | Published: 2023-05-23 | Updated: 2024-03-18

On the Optimal Batch Size for Byzantine-Robust Distributed Learning

Authors: Yi-Rui Yang, Chang-Wei Shi, Wu-Jun Li | Published: 2023-05-23

Understanding Programs by Exploiting (Fuzzing) Test Cases

Authors: Jianyu Zhao, Yuyang Rong, Yiwen Guo, Yifeng He, Hao Chen | Published: 2023-05-23 | Updated: 2023-06-12

Differential Privacy with Random Projections and Sign Random Projections

Authors: Ping Li, Xiaoyun Li | Published: 2023-05-22 | Updated: 2023-06-13

Evaluating Privacy Leakage in Split Learning

Authors: Xinchi Qiu, Ilias Leontiadis, Luca Melis, Alex Sablayrolles, Pierre Stock | Published: 2023-05-22 | Updated: 2024-01-19

Fact-Checking Complex Claims with Program-Guided Reasoning

Authors: Liangming Pan, Xiaobao Wu, Xinyuan Lu, Anh Tuan Luu, William Yang Wang, Min-Yen Kan, Preslav Nakov | Published: 2023-05-22

Quantifying Association Capabilities of Large Language Models and Its Implications on Privacy Leakage

Authors: Hanyin Shao, Jie Huang, Shen Zheng, Kevin Chen-Chuan Chang | Published: 2023-05-22 | Updated: 2024-02-09

In the Name of Fairness: Assessing the Bias in Clinical Record De-identification

Authors: Yuxin Xiao, Shulammite Lim, Tom Joseph Pollard, Marzyeh Ghassemi | Published: 2023-05-18 | Updated: 2024-01-03