AIセキュリティポータルbot

SOFT: Selective Data Obfuscation for Protecting LLM Fine-tuning against Membership Inference Attacks

Authors: Kaiyuan Zhang, Siyuan Cheng, Hanxi Guo, Yuetian Chen, Zian Su, Shengwei An, Yuntao Du, Charles Fleming, Ashish Kundu, Xiangyu Zhang, Ninghui Li | Published: 2025-06-12
Privacy Protection Method
Prompt Injection
Prompt leaking

FicGCN: Unveiling the Homomorphic Encryption Efficiency from Irregular Graph Convolutional Networks

Authors: Zhaoxuan Kan, Husheng Han, Shangyi Shi, Tenghui Hua, Hang Lu, Xiaowei Li, Jianan Mu, Xing Hu | Published: 2025-06-12
ノード最適化手法
Privacy Enhancing Technology
Watermarking Technology

Can We Infer Confidential Properties of Training Data from LLMs?

Authors: Penguin Huang, Chhavi Yadav, Ruihan Wu, Kamalika Chaudhuri | Published: 2025-06-12
Privacy Enhancing Technology
医療診断属性
Large Language Model

ELFuzz: Efficient Input Generation via LLM-driven Synthesis Over Fuzzer Space

Authors: Chuyang Chen, Brendan Dolan-Gavitt, Zhiqiang Lin | Published: 2025-06-12
Fuzzing
Prompt Injection
効率的入力生成

LLMail-Inject: A Dataset from a Realistic Adaptive Prompt Injection Challenge

Authors: Sahar Abdelnabi, Aideen Fay, Ahmed Salem, Egor Zverev, Kai-Chieh Liao, Chi-Huang Liu, Chun-Chih Kuo, Jannis Weigend, Danyael Manlangit, Alex Apostolov, Haris Umair, João Donato, Masayuki Kawakita, Athar Mahboob, Tran Huu Bach, Tsun-Han Chiang, Myeongjin Cho, Hajin Choi, Byeonghyeon Kim, Hyeonjin Lee, Benjamin Pannell, Conor McCauley, Mark Russinovich, Andrew Paverd, Giovanni Cherubin | Published: 2025-06-11
Indirect Prompt Injection
Prompt Injection
Defense Method

Private Aggregation for Byzantine-Resilient Heterogeneous Federated Learning

Authors: Maximilian Egger, Rawad Bitar | Published: 2025-06-11
Privacy protection framework
Robustness Evaluation
Federated Learning

Evasion Attacks Against Bayesian Predictive Models

Authors: Pablo G. Arce, Roi Naveiro, David Ríos Insua | Published: 2025-06-11
Backdoor Attack
Bayesian Adversarial Learning
Adversarial Perturbation Techniques

Learning Obfuscations Of LLM Embedding Sequences: Stained Glass Transform

Authors: Jay Roberts, Kyle Mylonakis, Sidhartha Roy, Kaan Kale | Published: 2025-06-11
Privacy Enhancing Technology
Differential Privacy
情報理論的関係

LLMs Cannot Reliably Judge (Yet?): A Comprehensive Assessment on the Robustness of LLM-as-a-Judge

Authors: Songze Li, Chuokun Xu, Jiaying Wang, Xueluan Gong, Chen Chen, Jirui Zhang, Jun Wang, Kwok-Yan Lam, Shouling Ji | Published: 2025-06-11
Disabling Safety Mechanisms of LLM
Prompt Injection
Adversarial attack

Adversarial Surrogate Risk Bounds for Binary Classification

Authors: Natalie S. Frank | Published: 2025-06-11
Certified Robustness
Convergence Analysis
Function Boundary Pair Formation