Prompt Injection

PromptGuard: Soft Prompt-Guided Unsafe Content Moderation for Text-to-Image Models

Authors: Lingzhi Yuan, Xinfeng Li, Chejian Xu, Guanhong Tao, Xiaojun Jia, Yihao Huang, Wei Dong, Yang Liu, XiaoFeng Wang, Bo Li | Published: 2025-01-07
Content Moderation
Soft Prompt Optimization
Prompt Injection

RTLMarker: Protecting LLM-Generated RTL Copyright via a Hardware Watermarking Framework

Authors: Kun Wang, Kaiyan Chang, Mengdi Wang, Xinqi Zou, Haobo Xu, Yinhe Han, Ying Wang | Published: 2025-01-05
Prompt Injection
Watermark Robustness
Watermark Evaluation

GNSS/GPS Spoofing and Jamming Identification Using Machine Learning and Deep Learning

Authors: Ali Ghanbarzade, Hossein Soleimani | Published: 2025-01-04
GNSS Security
Prompt Injection
Label

Auto-RT: Automatic Jailbreak Strategy Exploration for Red-Teaming Large Language Models

Authors: Yanjiang Liu, Shuhen Zhou, Yaojie Lu, Huijia Zhu, Weiqiang Wang, Hongyu Lin, Ben He, Xianpei Han, Le Sun | Published: 2025-01-03
Framework
Prompt Injection
Attack Method

CySecBench: Generative AI-based CyberSecurity-focused Prompt Dataset for Benchmarking Large Language Models

Authors: Johan Wahréus, Ahmed Mohamed Hussain, Panos Papadimitratos | Published: 2025-01-02
LLM Performance Evaluation
Cybersecurity
Prompt Injection

Safeguarding Large Language Models in Real-time with Tunable Safety-Performance Trade-offs

Authors: Joao Fonseca, Andrew Bell, Julia Stoyanovich | Published: 2025-01-02
Framework
Prompt Injection
Safety Alignment

SPADE: Enhancing Adaptive Cyber Deception Strategies with Generative AI and Structured Prompt Engineering

Authors: Shihab Ahmed, A B M Mohaimenur Rahman, Md Morshed Alam, Md Sajidul Islam Sajid | Published: 2025-01-01
Cybersecurity
Prompt Injection
Prompt Engineering

SecBench: A Comprehensive Multi-Dimensional Benchmarking Dataset for LLMs in Cybersecurity

Authors: Pengfei Jing, Mengyun Tang, Xiaorong Shi, Xing Zheng, Sen Nie, Shi Wu, Yong Yang, Xiapu Luo | Published: 2024-12-30 | Updated: 2025-01-06
LLM Performance Evaluation
Cybersecurity
Prompt Injection

From Vulnerabilities to Remediation: A Systematic Literature Review of LLMs in Code Security

Authors: Enna Basic, Alberto Giaretta | Published: 2024-12-19 | Updated: 2025-04-14
Prompt Injection
Prompt leaking
Vulnerability detection

SATA: A Paradigm for LLM Jailbreak via Simple Assistive Task Linkage

Authors: Xiaoning Dong, Wenbo Hu, Wei Xu, Tianxing He | Published: 2024-12-19 | Updated: 2025-03-21
Prompt Injection
Large Language Model
Adversarial Learning