Defensive Prompt Patch: A Robust and Interpretable Defense of LLMs against Jailbreak Attacks Authors: Chen Xiong, Xiangyu Qi, Pin-Yu Chen, Tsung-Yi Ho | Published: 2024-05-30 | Updated: 2025-06-04 DPPセット生成Prompt InjectionAttack Method 2024.05.30 2025.06.06 Literature Database
Efficient Black-box Adversarial Attacks via Bayesian Optimization Guided by a Function Prior Authors: Shuyu Cheng, Yibo Miao, Yinpeng Dong, Xiao Yang, Xiao-Shan Gao, Jun Zhu | Published: 2024-05-29 AlgorithmAttack MethodOptimization Problem 2024.05.29 2025.05.27 Literature Database
Medical MLLM is Vulnerable: Cross-Modality Jailbreak and Mismatched Attacks on Medical Multimodal Large Language Models Authors: Xijie Huang, Xinyuan Wang, Hantao Zhang, Yinghao Zhu, Jiawen Xi, Jingkun An, Hao Wang, Hao Liang, Chengwei Pan | Published: 2024-05-26 | Updated: 2024-08-21 Prompt InjectionThreats of Medical AIAttack Method 2024.05.26 2025.05.27 Literature Database
Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Character Authors: Siyuan Ma, Weidi Luo, Yu Wang, Xiaogeng Liu | Published: 2024-05-25 | Updated: 2024-06-12 LLM SecurityPrompt InjectionAttack Method 2024.05.25 2025.05.27 Literature Database
A novel reliability attack of Physical Unclonable Functions Authors: Gaoxiang Li, Yu Zhuang | Published: 2024-05-21 | Updated: 2024-06-07 FPGAExperimental ValidationAttack Method 2024.05.21 2025.05.27 Literature Database
GAN-GRID: A Novel Generative Attack on Smart Grid Stability Prediction Authors: Emad Efatinasab, Alessandro Brighente, Mirco Rampazzo, Nahal Azadi, Mauro Conti | Published: 2024-05-20 Model Performance EvaluationAttack EvaluationAttack Method 2024.05.20 2025.05.27 Literature Database
Dealing Doubt: Unveiling Threat Models in Gradient Inversion Attacks under Federated Learning, A Survey and Taxonomy Authors: Yichuan Shi, Olivera Kotevska, Viktor Reshniak, Abhishek Singh, Ramesh Raskar | Published: 2024-05-16 PoisoningAttack MethodDefense Method 2024.05.16 2025.05.27 Literature Database
Adversarial Machine Learning Threats to Spacecraft Authors: Rajiv Thummala, Shristi Sharma, Matteo Calabrese, Gregory Falco | Published: 2024-05-14 CybersecurityPoisoningAttack Method 2024.05.14 2025.05.27 Literature Database
Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM Authors: Xikang Yang, Xuehai Tang, Songlin Hu, Jizhong Han | Published: 2024-05-09 LLM SecurityPrompt InjectionAttack Method 2024.05.09 2025.05.27 Literature Database
Revisiting Character-level Adversarial Attacks for Language Models Authors: Elias Abad Rocamora, Yongtao Wu, Fanghui Liu, Grigorios G. Chrysos, Volkan Cevher | Published: 2024-05-07 | Updated: 2024-09-04 WatermarkingLoss FunctionAttack Method 2024.05.07 2025.05.27 Literature Database