Literature Database

BadToken: Token-level Backdoor Attacks to Multi-modal Large Language Models

Authors: Zenghui Yuan, Jiawen Shi, Pan Zhou, Neil Zhenqiang Gong, Lichao Sun | Published: 2025-03-20
Backdoor Attack
Prompt Injection
Large Language Model

TVineSynth: A Truncated C-Vine Copula Generator of Synthetic Tabular Data to Balance Privacy and Utility

Authors: Elisabeth Griesbauer, Claudia Czado, Arnoldo Frigessi, Ingrid Hobæk Haff | Published: 2025-03-20
Data Generation
Privacy-Preserving Machine Learning
Membership Inference

Digital Asset Data Lakehouse. The concept based on a blockchain research center

Authors: Raul Cristian Bag | Published: 2025-03-20
ETL Process
Cloud Computing
Data-Driven Clustering

DroidTTP: Mapping Android Applications with TTP for Cyber Threat Intelligence

Authors: Dincy R Arikkat, Vinod P., Rafidha Rehiman K. A., Serena Nicolazzo, Marco Arazzi, Antonino Nocera, Mauro Conti | Published: 2025-03-20
Poisoning attack on RAG
Performance Evaluation Method
Feature Extraction Method

Detecting LLM-Generated Peer Reviews

Authors: Vishisht Rao, Aounon Kumar, Himabindu Lakkaraju, Nihar B. Shah | Published: 2025-03-20 | Updated: 2025-05-19
Prompt Injection
Digital Watermarking for Generative AI
Watermark Design

AutoRedTeamer: Autonomous Red Teaming with Lifelong Attack Integration

Authors: Andy Zhou, Kevin Wu, Francesco Pinto, Zhaorun Chen, Yi Zeng, Yu Yang, Shuang Yang, Sanmi Koyejo, James Zou, Bo Li | Published: 2025-03-20
Error Handling
Text Generation Method
Test Case Generation

Towards Understanding the Safety Boundaries of DeepSeek Models: Evaluation and Findings

Authors: Zonghao Ying, Guangyi Zheng, Yongxin Huang, Deyue Zhang, Wenxin Zhang, Quanchen Zou, Aishan Liu, Xianglong Liu, Dacheng Tao | Published: 2025-03-19
Prompt Injection
Large Language Model
Attack Method

Prada: Black-Box LLM Adaptation with Private Data on Resource-Constrained Devices

Authors: Ziyao Wang, Yexiao He, Zheyu Shen, Yu Li, Guoheng Sun, Myungjin Lee, Ang Li | Published: 2025-03-19
Privacy Protection Method
Prompt leaking
Deep Learning

Temporal Context Awareness: A Defense Framework Against Multi-turn Manipulation Attacks on Large Language Models

Authors: Prashant Kulkarni, Assaf Namer | Published: 2025-03-18
Prompt Injection
Prompt leaking
Attack Method

Personalized Attacks of Social Engineering in Multi-turn Conversations — LLM Agents for Simulation and Detection

Authors: Tharindu Kumarage, Cameron Johnson, Jadie Adams, Lin Ai, Matthias Kirchner, Anthony Hoogs, Joshua Garland, Julia Hirschberg, Arslan Basharat, Huan Liu | Published: 2025-03-18
Alignment
Social Engineering Attack
Attack Method