Prompt Injection

Attacks and Defenses Against LLM Fingerprinting

Authors: Kevin Kurian, Ethan Holland, Sean Oesch | Published: 2025-08-12
Prompt Injection
Reinforcement Learning
Watermark Design

Oblivionis: A Lightweight Learning and Unlearning Framework for Federated Large Language Models

Authors: Fuyao Zhang, Xinyu Yan, Tiantong Wu, Wenjie Li, Tianxiang Chen, Yang Cao, Ran Yan, Longtao Huang, Wei Yang Bryan Lim, Qiang Yang | Published: 2025-08-12
Data Management System
Framework
Prompt Injection

Robust Anomaly Detection in O-RAN: Leveraging LLMs against Data Manipulation Attacks

Authors: Thusitha Dayaratne, Ngoc Duy Pham, Viet Vo, Shangqi Lai, Sharif Abuadbba, Hajime Suzuki, Xingliang Yuan, Carsten Rudolph | Published: 2025-08-11
Framework
Prompt Injection
Performance Evaluation Method

JPS: Jailbreak Multimodal Large Language Models with Collaborative Visual Perturbation and Textual Steering

Authors: Renmiao Chen, Shiyao Cui, Xuancheng Huang, Chengwei Pan, Victor Shea-Jay Huang, QingLin Zhang, Xuan Ouyang, Zhexin Zhang, Hongning Wang, Minlie Huang | Published: 2025-08-07
Prompt Injection
Inappropriate Content Generation
攻撃戦略分析

When Good Sounds Go Adversarial: Jailbreaking Audio-Language Models with Benign Inputs

Authors: Bodam Kim, Hiskias Dingeto, Taeyoun Kwon, Dasol Choi, DongGeon Lee, Haon Park, JaeHoon Lee, Jongho Shin | Published: 2025-08-05
Prompt Injection
Attack Evaluation
音声モデルの脆弱性

VFLAIR-LLM: A Comprehensive Framework and Benchmark for Split Learning of LLMs

Authors: Zixuan Gu, Qiufeng Fan, Long Sun, Yang Liu, Xiaojun Ye | Published: 2025-08-05
Prompt Injection
Prompt leaking
Watermark

PhishParrot: LLM-Driven Adaptive Crawling to Unveil Cloaked Phishing Sites

Authors: Hiroki Nakano, Takashi Koide, Daiki Chiba | Published: 2025-08-04
Indirect Prompt Injection
Prompt Injection
Malicious Website Detection

Breaking Obfuscation: Cluster-Aware Graph with LLM-Aided Recovery for Malicious JavaScript Detection

Authors: Zhihong Liang, Xin Wang, Zhenhuang Hu, Liangliang Song, Lin Chen, Jingjing Guo, Yanbin Wang, Ye Tian | Published: 2025-07-30
Program Verification
Prompt Injection
Robustness of Watermarking Techniques

Can We End the Cat-and-Mouse Game? Simulating Self-Evolving Phishing Attacks with LLMs and Genetic Algorithms

Authors: Seiji Sato, Tetsushi Ohki, Masakatsu Nishigaki | Published: 2025-07-29
Prompt Injection
Prompt leaking
心理学理論

Repairing vulnerabilities without invisible hands. A differentiated replication study on LLMs

Authors: Maria Camporese, Fabio Massacci | Published: 2025-07-28
Prompt Injection
Large Language Model
Vulnerability Management