AIセキュリティポータルbot

Seeing the Forest through the Trees: Data Leakage from Partial Transformer Gradients

Authors: Weijun Li, Qiongkai Xu, Mark Dras | Published: 2024-06-03 | Updated: 2024-10-04
ウォーターマーキング
データプライバシー評価
プライバシー保護手法

BadRAG: Identifying Vulnerabilities in Retrieval Augmented Generation of Large Language Models

Authors: Jiaqi Xue, Mengxin Zheng, Yebowen Hu, Fei Liu, Xun Chen, Qian Lou | Published: 2024-06-03 | Updated: 2024-06-06
LLM性能評価
クエリの多様性
クエリ生成手法

A Synergistic Approach In Network Intrusion Detection By Neurosymbolic AI

Authors: Alice Bizzarri, Chung-En Yu, Brian Jalaian, Fabrizio Riguzzi, Nathaniel D. Bastian | Published: 2024-06-03
NSAI統合
モデルの解釈性
未知の攻撃検出

Constrained Adaptive Attack: Effective Adversarial Attack Against Deep Neural Networks for Tabular Data

Authors: Thibault Simonetto, Salah Ghamizi, Maxime Cordy | Published: 2024-06-02
CAPGDアルゴリズム
攻撃手法
敵対的訓練

Transforming Computer Security and Public Trust Through the Exploration of Fine-Tuning Large Language Models

Authors: Garrett Crumrine, Izzat Alsmadi, Jesus Guerrero, Yuvaraj Munian | Published: 2024-06-02
LLMセキュリティ
サイバーセキュリティ
倫理的ガイドライン遵守

VeriSplit: Secure and Practical Offloading of Machine Learning Inferences across IoT Devices

Authors: Han Zhang, Zifan Wang, Mihir Dhamankar, Matt Fredrikson, Yuvraj Agarwal | Published: 2024-06-02 | Updated: 2025-03-31
ウォーターマーキング
データプライバシー評価
計算効率

Exploring Vulnerabilities and Protections in Large Language Models: A Survey

Authors: Frank Weizhen Liu, Chenhui Hu | Published: 2024-06-01
LLMセキュリティ
プロンプトインジェクション
防御手法

Improved Techniques for Optimization-Based Jailbreaking on Large Language Models

Authors: Xiaojun Jia, Tianyu Pang, Chao Du, Yihao Huang, Jindong Gu, Yang Liu, Xiaochun Cao, Min Lin | Published: 2024-05-31 | Updated: 2024-06-05
LLMセキュリティ
ウォーターマーキング
プロンプトインジェクション

ACE: A Model Poisoning Attack on Contribution Evaluation Methods in Federated Learning

Authors: Zhangchen Xu, Fengqing Jiang, Luyao Niu, Jinyuan Jia, Bo Li, Radha Poovendran | Published: 2024-05-31 | Updated: 2024-06-05
ポイズニング
評価手法
防御手法

Defensive Prompt Patch: A Robust and Interpretable Defense of LLMs against Jailbreak Attacks

Authors: Chen Xiong, Xiangyu Qi, Pin-Yu Chen, Tsung-Yi Ho | Published: 2024-05-30 | Updated: 2025-06-04
DPPセット生成
プロンプトインジェクション
攻撃手法