JailbreakLens: Visual Analysis of Jailbreak Attacks Against Large Language Models
Authors: Yingchaojie Feng, Zhizhang Chen, Zhining Kang, Sijia Wang, Minfeng Zhu, Wei Zhang, Wei Chen | Published: 2024-04-12
LLM Performance Evaluation
Prompt Injection
Evaluation Method