Prompt Injection

ChatIDS: Explainable Cybersecurity Using Generative AI

Authors: Victor Jüttner, Martin Grimmer, Erik Buchmann | Published: 2023-06-26
Online Safety Advice
Prompt Injection
Expert Opinion Collection

On the Uses of Large Language Models to Interpret Ambiguous Cyberattack Descriptions

Authors: Reza Fayyazi, Shanchieh Jay Yang | Published: 2023-06-24 | Updated: 2023-08-22
Prompt Injection
Malware Classification
Natural Language Processing

Visual Adversarial Examples Jailbreak Aligned Large Language Models

Authors: Xiangyu Qi, Kaixuan Huang, Ashwinee Panda, Peter Henderson, Mengdi Wang, Prateek Mittal | Published: 2023-06-22 | Updated: 2023-08-16
Prompt Injection
Inappropriate Content Generation
Adversarial attack

Matching Pairs: Attributing Fine-Tuned Models to their Pre-Trained Large Language Models

Authors: Myles Foley, Ambrish Rawat, Taesung Lee, Yufang Hou, Gabriele Picco, Giulio Zizzo | Published: 2023-06-15
LLM Performance Evaluation
Algorithm
Prompt Injection

Augmenting Greybox Fuzzing with Generative AI

Authors: Jie Hu, Qian Zhang, Heng Yin | Published: 2023-06-11
Fuzzing
Prompt Injection
Performance Evaluation

Prompt Injection attack against LLM-integrated Applications

Authors: Yi Liu, Gelei Deng, Yuekang Li, Kailong Wang, Zihao Wang, Xiaofeng Wang, Tianwei Zhang, Yepang Liu, Haoyu Wang, Yan Zheng, Yang Liu | Published: 2023-06-08 | Updated: 2024-03-02
Prompt Injection
Malicious Prompt

On the Detectability of ChatGPT Content: Benchmarking, Methodology, and Evaluation through the Lens of Academic Writing

Authors: Zeyan Liu, Zijun Yao, Fengjun Li, Bo Luo | Published: 2023-06-07 | Updated: 2024-03-18
LLM Application
Prompt Injection
Literature List

On Evaluating Adversarial Robustness of Large Vision-Language Models

Authors: Yunqing Zhao, Tianyu Pang, Chao Du, Xiao Yang, Chongxuan Li, Ngai-Man Cheung, Min Lin | Published: 2023-05-26 | Updated: 2023-10-29
LLM Performance Evaluation
Prompt Injection
Adversarial attack

Spear Phishing With Large Language Models

Authors: Julian Hazell | Published: 2023-05-11 | Updated: 2023-12-22
Cyber Attack
Phishing Attack
Prompt Injection

In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT

Authors: Xinyue Shen, Zeyuan Chen, Michael Backes, Yang Zhang | Published: 2023-04-18 | Updated: 2023-10-05
LLM Security
Prompt Injection
User Experience Evaluation