These labels were automatically added by AI and may be inaccurate. For details, see About Literature Database.
Abstract
Current LLMs are generally aligned to follow safety requirements and tend to
refuse toxic prompts. However, LLMs can fail to refuse toxic prompts or be
overcautious and refuse benign examples. In addition, state-of-the-art toxicity
detectors have low TPRs at low FPR, incurring high costs in real-world
applications where toxic examples are rare. In this paper, we introduce
Moderation Using LLM Introspection (MULI), which detects toxic prompts using
the information extracted directly from LLMs themselves. We found we can
distinguish between benign and toxic prompts from the distribution of the first
response token's logits. Using this idea, we build a robust detector of toxic
prompts using a sparse logistic regression model on the first response token
logits. Our scheme outperforms SOTA detectors under multiple metrics.