AIセキュリティポータルbot

Machine Unlearning: A Survey

Authors: Heng Xu, Tianqing Zhu, Lefeng Zhang, Wanlei Zhou, Philip S. Yu | Published: 2023-06-06
ウォーターマーキング
メンバーシップ推論
モデルの設計や精度

Adversarial attacks and defenses in explainable artificial intelligence: A survey

Authors: Hubert Baniecki, Przemyslaw Biecek | Published: 2023-06-06 | Updated: 2024-02-13
メンバーシップ推論
攻撃手法
敵対的サンプル

A Survey on Federated Learning Poisoning Attacks and Defenses

Authors: Junchuan Lianga, Rong Wang, Chaosheng Feng, Chin-Chen Chang | Published: 2023-06-06
バックドア攻撃
ポイズニング
未ターゲット毒性攻撃

Evading Black-box Classifiers Without Breaking Eggs

Authors: Edoardo Debenedetti, Nicholas Carlini, Florian Tramèr | Published: 2023-06-05 | Updated: 2024-02-14
攻撃の評価
敵対的サンプル
敵対的攻撃

A Privacy-Preserving Federated Learning Approach for Kernel methods

Authors: Anika Hannemann, Ali Burak Ünal, Arjhun Swaminathan, Erik Buchmann, Mete Akgün | Published: 2023-06-05
MPCアルゴリズム
プライバシー保護技術
透かし評価

Building Resilient SMEs: Harnessing Large Language Models for Cyber Security in Australia

Authors: Benjamin Kereopa-Yorke | Published: 2023-06-05
LLMの応用
サイバー攻撃
倫理的ガイドライン遵守

Poisoning Network Flow Classifiers

Authors: Giorgio Severi, Simona Boboila, Alina Oprea, John Holodnak, Kendra Kratkiewicz, Jason Matterer | Published: 2023-06-02
バックドア攻撃
ポイズニング
敵対的攻撃

Hyperparameter Learning under Data Poisoning: Analysis of the Influence of Regularization via Multiobjective Bilevel Optimization

Authors: Javier Carnerero-Cano, Luis Muñoz-González, Phillippa Spencer, Emil C. Lupu | Published: 2023-06-02 | Updated: 2023-06-23
ポイズニング
攻撃の評価
正則化

PassGPT: Password Modeling and (Guided) Generation with Large Language Models

Authors: Javier Rando, Fernando Perez-Cruz, Briland Hitaj | Published: 2023-06-02 | Updated: 2023-06-14
LLMの応用
モデルの設計や精度
透かし評価

A Closer Look at the Adversarial Robustness of Deep Equilibrium Models

Authors: Zonghan Yang, Tianyu Pang, Yang Liu | Published: 2023-06-02
ロバスト性に関する評価
敵対的攻撃
適応型敵対的訓練