GLL: A Differentiable Graph Learning Layer for Neural Networks Authors: Jason Brown, Bohan Chen, Harris Hardiman-Mostow, Jeff Calder, Andrea L. Bertozzi | Published: 2024-12-11 PoisoningAdversarial Training 2024.12.11 2025.05.27 Literature Database
Heuristic-Induced Multimodal Risk Distribution Jailbreak Attack for Multimodal Large Language Models Authors: Ma Teng, Jia Xiaojun, Duan Ranjie, Li Xinfeng, Huang Yihao, Chu Zhixuan, Liu Yang, Ren Wenqi | Published: 2024-12-08 | Updated: 2025-01-03 Content ModerationPrompt InjectionAttack Method 2024.12.08 2025.05.27 Literature Database
ChatNVD: Advancing Cybersecurity Vulnerability Assessment with Large Language Models Authors: Shivansh Chopra, Hussain Ahmad, Diksha Goel, Claudia Szabo | Published: 2024-12-06 | Updated: 2025-05-20 Text Generation MethodPrompt InjectionComputational Efficiency 2024.12.06 2025.05.28 Literature Database
On the Lack of Robustness of Binary Function Similarity Systems Authors: Gianluca Capozzi, Tong Tang, Jie Wan, Ziqi Yang, Daniele Cono D'Elia, Giuseppe Antonio Di Luna, Lorenzo Cavallaro, Leonardo Querzoni | Published: 2024-12-05 | Updated: 2025-05-22 バイナリ分析Adversarial LearningAdversarial Learning 2024.12.05 2025.05.28 Literature Database
DP-2Stage: Adapting Language Models as Differentially Private Tabular Data Generators Authors: Tejumade Afonja, Hui-Po Wang, Raouf Kerkouche, Mario Fritz | Published: 2024-12-03 | Updated: 2025-04-29 Privacy ViolationSynthetic Data GenerationDifferential Privacy 2024.12.03 2025.05.27 Literature Database
Intermediate Outputs Are More Sensitive Than You Think Authors: Tao Huang, Qingyu Huang, Jiayang Meng | Published: 2024-12-01 Privacy ProtectionMembership Inference 2024.12.01 2025.05.27 Literature Database
VLSBench: Unveiling Visual Leakage in Multimodal Safety Authors: Xuhao Hu, Dongrui Liu, Hao Li, Xuanjing Huang, Jing Shao | Published: 2024-11-29 | Updated: 2025-01-17 Prompt InjectionSafety Alignment 2024.11.29 2025.05.27 Literature Database
LUMIA: Linear probing for Unimodal and MultiModal Membership Inference Attacks leveraging internal LLM states Authors: Luis Ibanez-Lissen, Lorena Gonzalez-Manzano, Jose Maria de Fuentes, Nicolas Anciaux, Joaquin Garcia-Alfaro | Published: 2024-11-29 | Updated: 2025-01-10 LLM Performance EvaluationMembership Inference 2024.11.29 2025.05.27 Literature Database
CantorNet: A Sandbox for Testing Geometrical and Topological Complexity Measures Authors: Michal Lewandowski, Hamid Eghbalzadeh, Bernhard A. Moser | Published: 2024-11-29 | Updated: 2025-01-28 Framework 2024.11.29 2025.05.27 Literature Database
Immune: Improving Safety Against Jailbreaks in Multi-modal LLMs via Inference-Time Alignment Authors: Soumya Suvra Ghosal, Souradip Chakraborty, Vaibhav Singh, Tianrui Guan, Mengdi Wang, Ahmad Beirami, Furong Huang, Alvaro Velasquez, Dinesh Manocha, Amrit Singh Bedi | Published: 2024-11-27 | Updated: 2025-03-20 Prompt InjectionSafety AlignmentAdversarial attack 2024.11.27 2025.05.27 Literature Database