Prompt leaking

When LLMs Copy to Think: Uncovering Copy-Guided Attacks in Reasoning LLMs

Authors: Yue Li, Xiao Li, Hao Wu, Yue Zhang, Fengyuan Xu, Xiuzhen Cheng, Sheng Zhong | Published: 2025-07-22
Prompt leaking
Model DoS
Attack Method

LLMxCPG: Context-Aware Vulnerability Detection Through Code Property Graph-Guided Large Language Models

Authors: Ahmed Lekssays, Hamza Mouhcine, Khang Tran, Ting Yu, Issa Khalil | Published: 2025-07-22
Dataset Analysis
Prompt leaking
脆弱性検出手法

Depth Gives a False Sense of Privacy: LLM Internal States Inversion

Authors: Tian Dong, Yan Meng, Shaofeng Li, Guoxing Chen, Zhen Liu, Haojin Zhu | Published: 2025-07-22
Prompt Injection
Prompt leaking
Attack Method

Talking Like a Phisher: LLM-Based Attacks on Voice Phishing Classifiers

Authors: Wenhao Li, Selvakumar Manickam, Yung-wey Chong, Shankar Karuppayah | Published: 2025-07-22
Prompt leaking
Psychological Manipulation
Adversarial Attack Methods

Distilled Large Language Model in Confidential Computing Environment for System-on-Chip Design

Authors: Dong Ben, Hui Feng, Qian Wang | Published: 2025-07-22
Prompt leaking
Model Extraction Attack
Factors of Performance Degradation

PhishIntentionLLM: Uncovering Phishing Website Intentions through Multi-Agent Retrieval-Augmented Generation

Authors: Wenhao Li, Selvakumar Manickam, Yung-wey Chong, Shankar Karuppayah | Published: 2025-07-21
Poisoning attack on RAG
フィッシング攻撃の意図
Prompt leaking

MAD-Spear: A Conformity-Driven Prompt Injection Attack on Multi-Agent Debate Systems

Authors: Yu Cui, Hongyang Du | Published: 2025-07-17
Security Strategy Generation
Prompt leaking
Attacker Behavior Analysis

Defending Against Prompt Injection With a Few DefensiveTokens

Authors: Sizhe Chen, Yizhu Wang, Nicholas Carlini, Chawin Sitawarin, David Wagner | Published: 2025-07-10
Indirect Prompt Injection
Prompt leaking
Defense Method

PenTest2.0: Towards Autonomous Privilege Escalation Using GenAI

Authors: Haitham S. Al-Sinani, Chris J. Mitchell | Published: 2025-07-09
Indirect Prompt Injection
Prompt validation
Prompt leaking

The Landscape of Memorization in LLMs: Mechanisms, Measurement, and Mitigation

Authors: Alexander Xiong, Xuandong Zhao, Aneesh Pappu, Dawn Song | Published: 2025-07-08
Prompt leaking
メモリ化メカニズム
Large Language Model