A Comprehensive Study of Jailbreak Attack versus Defense for Large Language Models

Authors: Zihao Xu, Yi Liu, Gelei Deng, Yuekang Li, Stjepan Picek | Published: 2024-02-21 | Updated: 2024-05-17

The Wolf Within: Covert Injection of Malice into MLLM Societies via an MLLM Operative

Authors: Zhen Tan, Chengshuai Zhao, Raha Moraffah, Yifan Li, Yu Kong, Tianlong Chen, Huan Liu | Published: 2024-02-20 | Updated: 2024-06-03

Revisiting Differentially Private Hyper-parameter Tuning

Authors: Zihang Xiang, Tianhao Wang, Chenglong Wang, Di Wang | Published: 2024-02-20 | Updated: 2024-06-04

IT Intrusion Detection Using Statistical Learning and Testbed Measurements

Authors: Xiaoxuan Wang, Rolf Stadler | Published: 2024-02-20

TRAP: Targeted Random Adversarial Prompt Honeypot for Black-Box Identification

Authors: Martin Gubri, Dennis Ulmer, Hwaran Lee, Sangdoo Yun, Seong Joon Oh | Published: 2024-02-20 | Updated: 2024-06-06

Prompt Stealing Attacks Against Large Language Models

Authors: Zeyang Sha, Yang Zhang | Published: 2024-02-20

Bounding Reconstruction Attack Success of Adversaries Without Data Priors

Authors: Alexander Ziller, Anneliese Riess, Kristian Schwethelm, Tamara T. Mueller, Daniel Rueckert, Georgios Kaissis | Published: 2024-02-20

APT-MMF: An advanced persistent threat actor attribution method based on multimodal and multilevel feature fusion

Authors: Nan Xiao, Bo Lang, Ting Wang, Yikai Chen | Published: 2024-02-20

Indiscriminate Data Poisoning Attacks on Pre-trained Feature Extractors

Authors: Yiwei Lu, Matthew Y. R. Yang, Gautam Kamath, Yaoliang Yu | Published: 2024-02-20

An Adversarial Approach to Evaluating the Robustness of Event Identification Models

Authors: Obai Bahwal, Oliver Kosut, Lalitha Sankar | Published: 2024-02-19 | Updated: 2024-04-22