GoldCoin: Grounding Large Language Models in Privacy Laws via Contextual Integrity Theory

Authors: Wei Fan, Haoran Li, Zheye Deng, Weiqi Wang, Yangqiu Song | Published: 2024-06-17 | Updated: 2024-10-04

Threat Modelling and Risk Analysis for Large Language Model (LLM)-Powered Applications

Authors: Stephen Burabari Tete | Published: 2024-06-16

Promoting Data and Model Privacy in Federated Learning through Quantized LoRA

Authors: JianHao Zhu, Changze Lv, Xiaohua Wang, Muling Wu, Wenhao Liu, Tianlong Li, Zixuan Ling, Cenyuan Zhang, Xiaoqing Zheng, Xuanjing Huang | Published: 2024-06-16

Really Unlearned? Verifying Machine Unlearning via Influential Sample Pairs

Authors: Heng Xu, Tianqing Zhu, Lefeng Zhang, Wanlei Zhou | Published: 2024-06-16

Trading Devil: Robust backdoor attack via Stochastic investment models and Bayesian approach

Authors: Orson Mengara | Published: 2024-06-15 | Updated: 2024-09-16

Emerging Safety Attack and Defense in Federated Instruction Tuning of Large Language Models

Authors: Rui Ye, Jingyi Chai, Xiangrui Liu, Yaodong Yang, Yanfeng Wang, Siheng Chen | Published: 2024-06-15

RMF: A Risk Measurement Framework for Machine Learning Models

Authors: Jan Schröder, Jakub Breier | Published: 2024-06-15

Speed-up of Data Analysis with Kernel Trick in Encrypted Domain

Authors: Joon Soo Yoo, Baek Kyung Song, Tae Min Ahn, Ji Won Heo, Ji Won Yoon | Published: 2024-06-14

Privacy-preserving Quantification of Non-IID Degree in Federated Learning

Authors: Yuping Yan, Yizhi Wang, Yingchao Yu, Yaochu Jin | Published: 2024-06-14

DefAn: Definitive Answer Dataset for LLMs Hallucination Evaluation

Authors: A B M Ashikur Rahman, Saeed Anwar, Muhammad Usman, Ajmal Mian | Published: 2024-06-13