Calculation of Output Harmfulness

You Only Prompt Once: On the Capabilities of Prompt Learning on Large Language Models to Tackle Toxic Content

Authors: Xinlei He, Savvas Zannettou, Yun Shen, Yang Zhang | Published: 2023-08-10
Text Detoxification
Prompt leaking
Calculation of Output Harmfulness

Toxicity Detection with Generative Prompt-based Inference

Authors: Yau-Shian Wang, Yingshan Chang | Published: 2022-05-24
Prompting Strategy
Calculation of Output Harmfulness
Large Language Model