防御手法

Time Traveling to Defend Against Adversarial Example Attacks in Image Classification

Authors: Anthony Etim, Jakub Szefer | Published: 2024-10-10
攻撃手法
敵対的サンプル
防御手法

Prompt Infection: LLM-to-LLM Prompt Injection within Multi-Agent Systems

Authors: Donghyun Lee, Mo Tiwari | Published: 2024-10-09
プロンプトインジェクション
攻撃手法
防御手法

SecAlign: Defending Against Prompt Injection with Preference Optimization

Authors: Sizhe Chen, Arman Zharmagambetov, Saeed Mahloujifar, Kamalika Chaudhuri, David Wagner, Chuan Guo | Published: 2024-10-07 | Updated: 2025-01-13
LLMセキュリティ
プロンプトインジェクション
防御手法

SoK: Towards Security and Safety of Edge AI

Authors: Tatjana Wingarz, Anne Lauscher, Janick Edinger, Dominik Kaaser, Stefan Schulte, Mathias Fischer | Published: 2024-10-07
バイアス
プライバシー保護
防御手法

Robustness Reprogramming for Representation Learning

Authors: Zhichao Hou, MohamadAli Torkamani, Hamid Krim, Xiaorui Liu | Published: 2024-10-06
攻撃の評価
防御手法

Enhancing Robustness of Graph Neural Networks through p-Laplacian

Authors: Anuj Kumar Sirohi, Subhanu Halder, Kabir Kumar, Sandeep Kumar | Published: 2024-09-27
最適化問題
防御手法

Obliviate: Neutralizing Task-agnostic Backdoors within the Parameter-efficient Fine-tuning Paradigm

Authors: Jaehan Kim, Minkyoo Song, Seung Ho Na, Seungwon Shin | Published: 2024-09-21 | Updated: 2024-10-06
バックドア攻撃
モデル性能評価
防御手法

Defending against Model Inversion Attacks via Random Erasing

Authors: Viet-Hung Tran, Ngoc-Bao Nguyen, Son T. Mai, Hans Vandierendonck, Ngai-man Cheung | Published: 2024-09-02
ウォーターマーキング
プライバシー保護手法
防御手法

EEG-Defender: Defending against Jailbreak through Early Exit Generation of Large Language Models

Authors: Chongwen Zhao, Zhihao Dou, Kaizhu Huang | Published: 2024-08-21
LLMセキュリティ
プロンプトインジェクション
防御手法

Robust Image Classification: Defensive Strategies against FGSM and PGD Adversarial Attacks

Authors: Hetvi Waghela, Jaydip Sen, Sneha Rakshit | Published: 2024-08-20
ポイズニング
敵対的サンプル
防御手法