LLM Security

Getting pwn’d by AI: Penetration Testing with Large Language Models

Authors: Andreas Happe, Jürgen Cito | Published: 2023-07-24 | Updated: 2023-08-17
LLM Security
Prompt Injection
Penetration Testing Methods

Privacy-Preserving Prompt Tuning for Large Language Model Services

Authors: Yansong Li, Zhixing Tan, Yang Liu | Published: 2023-05-10 | Updated: 2025-01-10
DNN IP Protection Method
LLM Security
Privacy Assessment

In ChatGPT We Trust? Measuring and Characterizing the Reliability of ChatGPT

Authors: Xinyue Shen, Zeyuan Chen, Michael Backes, Yang Zhang | Published: 2023-04-18 | Updated: 2023-10-05
LLM Security
Prompt Injection
User Experience Evaluation

Stochastic Parrots Looking for Stochastic Parrots: LLMs are Easy to Fine-Tune and Hard to Detect with other LLMs

Authors: Da Silva Gameiro Henrique, Andrei Kucharavy, Rachid Guerraoui | Published: 2023-04-18
LLM Security
Text Generation Method
Generative Adversarial Network

Multi-step Jailbreaking Privacy Attacks on ChatGPT

Authors: Haoran Li, Dadi Guo, Wei Fan, Mingshi Xu, Jie Huang, Fanpu Meng, Yangqiu Song | Published: 2023-04-11 | Updated: 2023-11-01
LLM Security
Privacy Analysis
Prompt Injection