Prompt Injection

When Good Sounds Go Adversarial: Jailbreaking Audio-Language Models with Benign Inputs

Authors: Bodam Kim, Hiskias Dingeto, Taeyoun Kwon, Dasol Choi, DongGeon Lee, Haon Park, JaeHoon Lee, Jongho Shin | Published: 2025-08-05
Prompt Injection
Attack Evaluation
音声モデルの脆弱性

VFLAIR-LLM: A Comprehensive Framework and Benchmark for Split Learning of LLMs

Authors: Zixuan Gu, Qiufeng Fan, Long Sun, Yang Liu, Xiaojun Ye | Published: 2025-08-05
Prompt Injection
Prompt leaking
Watermark

PhishParrot: LLM-Driven Adaptive Crawling to Unveil Cloaked Phishing Sites

Authors: Hiroki Nakano, Takashi Koide, Daiki Chiba | Published: 2025-08-04
Indirect Prompt Injection
Prompt Injection
Malicious Website Detection

Breaking Obfuscation: Cluster-Aware Graph with LLM-Aided Recovery for Malicious JavaScript Detection

Authors: Zhihong Liang, Xin Wang, Zhenhuang Hu, Liangliang Song, Lin Chen, Jingjing Guo, Yanbin Wang, Ye Tian | Published: 2025-07-30
Program Verification
Prompt Injection
Robustness of Watermarking Techniques

Can We End the Cat-and-Mouse Game? Simulating Self-Evolving Phishing Attacks with LLMs and Genetic Algorithms

Authors: Seiji Sato, Tetsushi Ohki, Masakatsu Nishigaki | Published: 2025-07-29
Prompt Injection
Prompt leaking
心理学理論

Repairing vulnerabilities without invisible hands. A differentiated replication study on LLMs

Authors: Maria Camporese, Fabio Massacci | Published: 2025-07-28
Prompt Injection
Large Language Model
Vulnerability Management

Information Security Based on LLM Approaches: A Review

Authors: Chang Gong, Zhongwen Li, Xiaoqi Li | Published: 2025-07-24
Network Traffic Analysis
Prompt Injection
Prompt leaking

Tab-MIA: A Benchmark Dataset for Membership Inference Attacks on Tabular Data in LLMs

Authors: Eyal German, Sagiv Antebi, Daniel Samira, Asaf Shabtai, Yuval Elovici | Published: 2025-07-23
Relationship of AI Systems
Property Inference Attack
Prompt Injection

Depth Gives a False Sense of Privacy: LLM Internal States Inversion

Authors: Tian Dong, Yan Meng, Shaofeng Li, Guoxing Chen, Zhen Liu, Haojin Zhu | Published: 2025-07-22
Prompt Injection
Prompt leaking
Attack Method

Attacking interpretable NLP systems

Authors: Eldor Abdukhamidov, Tamer Abuhmed, Joanna C. S. Santos, Mohammed Abuhamad | Published: 2025-07-22
Prompt Injection
Prompt validation
Adversarial Attack Methods