Prompt Injection

ModelShield: Adaptive and Robust Watermark against Model Extraction Attack

Authors: Kaiyi Pang, Tao Qi, Chuhan Wu, Minhao Bai, Minghu Jiang, Yongfeng Huang | Published: 2024-05-03 | Updated: 2025-01-12
Watermarking
Prompt Injection
Watermark Evaluation

Generative AI in Cybersecurity

Authors: Shivani Metta, Isaac Chang, Jack Parker, Michael P. Roman, Arturo F. Ehuan | Published: 2024-05-02
Evolution of AI
Cybersecurity
Prompt Injection

WitheredLeaf: Finding Entity-Inconsistency Bugs with LLMs

Authors: Hongbo Chen, Yifan Zhang, Xing Han, Huanyao Rong, Yuheng Zhang, Tianhao Mao, Hang Zhang, XiaoFeng Wang, Luyi Xing, Xun Chen | Published: 2024-05-02
LLM Performance Evaluation
Code Generation
Prompt Injection

Boosting Jailbreak Attack with Momentum

Authors: Yihao Zhang, Zeming Wei | Published: 2024-05-02
Watermarking
Prompt Injection
Attack Method

DLAP: A Deep Learning Augmented Large Language Model Prompting Framework for Software Vulnerability Detection

Authors: Yanjing Yang, Xin Zhou, Runfeng Mao, Jinwei Xu, Lanxin Yang, Yu Zhangm, Haifeng Shen, He Zhang | Published: 2024-05-02
Prompt Injection
Prompt Engineering
Vulnerability Management

LLM Security Guard for Code

Authors: Arya Kavian, Mohammad Mehdi Pourhashem Kallehbasti, Sajjad Kazemi, Ehsan Firouzi, Mohammad Ghafari | Published: 2024-05-02 | Updated: 2024-05-03
LLM Security
Security Analysis
Prompt Injection

Unleashing the Power of LLM to Infer State Machine from the Protocol Implementation

Authors: Haiyang Wei, Ligeng Chen, Zhengjie Du, Yuhan Wu, Haohui Huang, Yue Liu, Guang Cheng, Fengyuan Xu, Linzhang Wang, Bing Mao | Published: 2024-05-01 | Updated: 2025-03-27
LLM Performance Evaluation
Prompt Injection
State Transition Model

TuBA: Cross-Lingual Transferability of Backdoor Attacks in LLMs with Instruction Tuning

Authors: Xuanli He, Jun Wang, Qiongkai Xu, Pasquale Minervini, Pontus Stenetorp, Benjamin I. P. Rubinstein, Trevor Cohn | Published: 2024-04-30 | Updated: 2025-03-17
Content Moderation
Backdoor Attack
Prompt Injection

Evaluating and Mitigating Linguistic Discrimination in Large Language Models

Authors: Guoliang Dong, Haoyu Wang, Jun Sun, Xinyu Wang | Published: 2024-04-29 | Updated: 2024-05-10
LLM Performance Evaluation
Bias
Prompt Injection

Attacks on Third-Party APIs of Large Language Models

Authors: Wanru Zhao, Vidit Khazanchi, Haodi Xing, Xuanli He, Qiongkai Xu, Nicholas Donald Lane | Published: 2024-04-24
LLM Security
Prompt Injection
Attack Method