Prompt Injection

Large Language Model-Powered Smart Contract Vulnerability Detection: New Perspectives

Authors: Sihao Hu, Tiansheng Huang, Fatih İlhan, Selim Furkan Tekin, Ling Liu | Published: 2023-10-02 | Updated: 2023-10-16
Security Analysis
Prompt Injection
Vulnerability Prediction

Watch Your Language: Investigating Content Moderation with Large Language Models

Authors: Deepak Kumar, Yousef AbuHashem, Zakir Durumeric | Published: 2023-09-25 | Updated: 2024-01-17
LLM Performance Evaluation
Prompt Injection
Inappropriate Content Generation

Can LLM-Generated Misinformation Be Detected?

Authors: Canyu Chen, Kai Shu | Published: 2023-09-25 | Updated: 2024-04-23
LLM Performance Evaluation
Prompt Injection
Inappropriate Content Generation

Defending Against Alignment-Breaking Attacks via Robustly Aligned LLM

Authors: Bochuan Cao, Yuanpu Cao, Lu Lin, Jinghui Chen | Published: 2023-09-18 | Updated: 2024-06-12
Prompt Injection
Safety Alignment
Defense Method

FuzzLLM: A Novel and Universal Fuzzing Framework for Proactively Discovering Jailbreak Vulnerabilities in Large Language Models

Authors: Dongyu Yao, Jianshu Zhang, Ian G. Harris, Marcel Carlsson | Published: 2023-09-11 | Updated: 2024-04-14
LLM Security
Watermarking
Prompt Injection

Vulnerability of Machine Learning Approaches Applied in IoT-based Smart Grid: A Review

Authors: Zhenyong Zhang, Mengxiang Liu, Mingyang Sun, Ruilong Deng, Peng Cheng, Dusit Niyato, Mo-Yuen Chow, Jiming Chen | Published: 2023-08-30 | Updated: 2023-12-25
Energy Management
Prompt Injection
Adversarial Training

Detecting Language Model Attacks with Perplexity

Authors: Gabriel Alon, Michael Kamfonas | Published: 2023-08-27 | Updated: 2023-11-07
LLM Security
Prompt Injection
Malicious Prompt

Use of LLMs for Illicit Purposes: Threats, Prevention Measures, and Vulnerabilities

Authors: Maximilian Mozes, Xuanli He, Bennett Kleinberg, Lewis D. Griffin | Published: 2023-08-24
Prompt Injection
Malicious Content Generation
Adversarial Example

Devising and Detecting Phishing: Large Language Models vs. Smaller Human Models

Authors: Fredrik Heiding, Bruce Schneier, Arun Vishwanath, Jeremy Bernstein, Peter S. Park | Published: 2023-08-23 | Updated: 2023-11-30
Phishing
Phishing Attack
Prompt Injection

Time Travel in LLMs: Tracing Data Contamination in Large Language Models

Authors: Shahriar Golchin, Mihai Surdeanu | Published: 2023-08-16 | Updated: 2024-02-21
Data Contamination Detection
Prompt Injection
Natural Language Processing