On Protecting the Data Privacy of Large Language Models (LLMs): A Survey

Authors: Biwei Yan, Kun Li, Minghui Xu, Yueyan Dong, Yue Zhang, Zhaochun Ren, Xiuzhen Cheng | Published: 2024-03-08 | Updated: 2024-03-14

ZTRAN: Prototyping Zero Trust Security xApps for Open Radio Access Network Deployments

Authors: Aly S. Abdalla, Joshua Moore, Nisha Adhikari, Vuk Marojevic | Published: 2024-03-06

Fuzzing BusyBox: Leveraging LLM and Crash Reuse for Embedded Bug Unearthing

Authors: Asmita, Yaroslav Oliinyk, Michael Scott, Ryan Tsang, Chongzhou Fang, Houman Homayoun | Published: 2024-03-06

Effect of Ambient-Intrinsic Dimension Gap on Adversarial Vulnerability

Authors: Rajdeep Haldar, Yue Xing, Qifan Song | Published: 2024-03-06 | Updated: 2024-03-23

Do You Trust Your Model? Emerging Malware Threats in the Deep Learning Ecosystem

Authors: Dorjan Hitaj, Giulio Pagnotta, Fabio De Gaspari, Sediola Ruko, Briland Hitaj, Luigi V. Mancini, Fernando Perez-Cruz | Published: 2024-03-06

Wildest Dreams: Reproducible Research in Privacy-preserving Neural Network Training

Authors: Tanveer Khan, Mindaugas Budzys, Khoa Nguyen, Antonis Michalas | Published: 2024-03-06

Enhancing Security in Federated Learning through Adaptive Consensus-Based Model Update Validation

Authors: Zahir Alsulaimawi | Published: 2024-03-05

Mitigating Label Flipping Attacks in Malicious URL Detectors Using Ensemble Trees

Authors: Ehsan Nowroozi, Nada Jadalla, Samaneh Ghelichkhani, Alireza Jolfaei | Published: 2024-03-05

Federated Learning Under Attack: Exposing Vulnerabilities through Data Poisoning Attacks in Computer Networks

Authors: Ehsan Nowroozi, Imran Haider, Rahim Taheri, Mauro Conti | Published: 2024-03-05

InjecAgent: Benchmarking Indirect Prompt Injections in Tool-Integrated Large Language Model Agents

Authors: Qiusi Zhan, Zhixiang Liang, Zifan Ying, Daniel Kang | Published: 2024-03-05 | Updated: 2024-08-04