Robustness Improvement Method

DefenSee: Dissecting Threat from Sight and Text – A Multi-View Defensive Pipeline for Multi-modal Jailbreaks

Authors: Zihao Wang, Kar Wai Fok, Vrizlynn L. L. Thing | Published: 2025-12-01
Prompt Injection
Model DoS
Robustness Improvement Method

On the Feasibility of Hijacking MLLMs’ Decision Chain via One Perturbation

Authors: Changyue Li, Jiaying Li, Youliang Yuan, Jiaming He, Zhicong Huang, Pinjia He | Published: 2025-11-25
Robustness Improvement Method
Image Processing
Adaptive Adversarial Training

Q-MLLM: Vector Quantization for Robust Multimodal Large Language Model Security

Authors: Wei Zhao, Zhe Li, Yige Li, Jun Sun | Published: 2025-11-20
Prompt leaking
Robustness Improvement Method
Digital Watermarking for Generative AI

FlowPure: Continuous Normalizing Flows for Adversarial Purification

Authors: Elias Collaert, Abel Rodríguez, Sander Joos, Lieven Desmet, Vera Rimmer | Published: 2025-05-19
Robustness Improvement Method
Adversarial Learning
Effectiveness Analysis of Defense Methods

Addressing Neural Network Robustness with Mixup and Targeted Labeling Adversarial Training

Authors: Alfred Laugros, Alice Caplier, Matthieu Ospici | Published: 2020-08-19
Robustness Improvement Method
Adversarial Example
Vulnerability of Adversarial Examples

Provably robust deep generative models

Authors: Filipe Condessa, Zico Kolter | Published: 2020-04-22
Robustness Improvement Method
Adversarial attack
Deep Learning Method

Certifying Joint Adversarial Robustness for Model Ensembles

Authors: Mainuddin Ahmad Jonas, David Evans | Published: 2020-04-21
Model Ensemble
Robustness Improvement Method
Adversarial Example

Luring of transferable adversarial perturbations in the black-box paradigm

Authors: Rémi Bernhard, Pierre-Alain Moellic, Jean-Max Dutertre | Published: 2020-04-10 | Updated: 2021-03-03
Robustness Improvement Method
Attack Evaluation
Adversarial Example

Adversarial Robustness for Code

Authors: Pavol Bielik, Martin Vechev | Published: 2020-02-11 | Updated: 2020-08-15
Poisoning
Robustness Improvement Method
Adversarial Training

Robustness of Bayesian Neural Networks to Gradient-Based Attacks

Authors: Ginevra Carbone, Matthew Wicker, Luca Laurenti, Andrea Patane, Luca Bortolussi, Guido Sanguinetti | Published: 2020-02-11 | Updated: 2020-06-24
Robustness Evaluation
Robustness Improvement Method
Adversarial attack