Attack Method

Defensive Prompt Patch: A Robust and Interpretable Defense of LLMs against Jailbreak Attacks

Authors: Chen Xiong, Xiangyu Qi, Pin-Yu Chen, Tsung-Yi Ho | Published: 2024-05-30 | Updated: 2025-06-04
DPPセット生成
Prompt Injection
Attack Method

Efficient Black-box Adversarial Attacks via Bayesian Optimization Guided by a Function Prior

Authors: Shuyu Cheng, Yibo Miao, Yinpeng Dong, Xiao Yang, Xiao-Shan Gao, Jun Zhu | Published: 2024-05-29
Algorithm
Attack Method
Optimization Problem

Medical MLLM is Vulnerable: Cross-Modality Jailbreak and Mismatched Attacks on Medical Multimodal Large Language Models

Authors: Xijie Huang, Xinyuan Wang, Hantao Zhang, Yinghao Zhu, Jiawen Xi, Jingkun An, Hao Wang, Hao Liang, Chengwei Pan | Published: 2024-05-26 | Updated: 2024-08-21
Prompt Injection
Threats of Medical AI
Attack Method

Visual-RolePlay: Universal Jailbreak Attack on MultiModal Large Language Models via Role-playing Image Character

Authors: Siyuan Ma, Weidi Luo, Yu Wang, Xiaogeng Liu | Published: 2024-05-25 | Updated: 2024-06-12
LLM Security
Prompt Injection
Attack Method

A novel reliability attack of Physical Unclonable Functions

Authors: Gaoxiang Li, Yu Zhuang | Published: 2024-05-21 | Updated: 2024-06-07
FPGA
Experimental Validation
Attack Method

GAN-GRID: A Novel Generative Attack on Smart Grid Stability Prediction

Authors: Emad Efatinasab, Alessandro Brighente, Mirco Rampazzo, Nahal Azadi, Mauro Conti | Published: 2024-05-20
Model Performance Evaluation
Attack Evaluation
Attack Method

Dealing Doubt: Unveiling Threat Models in Gradient Inversion Attacks under Federated Learning, A Survey and Taxonomy

Authors: Yichuan Shi, Olivera Kotevska, Viktor Reshniak, Abhishek Singh, Ramesh Raskar | Published: 2024-05-16
Poisoning
Attack Method
Defense Method

Adversarial Machine Learning Threats to Spacecraft

Authors: Rajiv Thummala, Shristi Sharma, Matteo Calabrese, Gregory Falco | Published: 2024-05-14
Cybersecurity
Poisoning
Attack Method

Chain of Attack: a Semantic-Driven Contextual Multi-Turn attacker for LLM

Authors: Xikang Yang, Xuehai Tang, Songlin Hu, Jizhong Han | Published: 2024-05-09
LLM Security
Prompt Injection
Attack Method

Revisiting Character-level Adversarial Attacks for Language Models

Authors: Elias Abad Rocamora, Yongtao Wu, Fanghui Liu, Grigorios G. Chrysos, Volkan Cevher | Published: 2024-05-07 | Updated: 2024-09-04
Watermarking
Loss Function
Attack Method