This study presents the first comprehensive safety evaluation of the DeepSeek
models, focusing on evaluating the safety risks associated with their generated
content. Our evaluation encompasses DeepSeek's latest generation of large
language models, multimodal large language models, and text-to-image models,
systematically examining their performance regarding unsafe content generation.
Notably, we developed a bilingual (Chinese-English) safety evaluation dataset
tailored to Chinese sociocultural contexts, enabling a more thorough evaluation
of the safety capabilities of Chinese-developed models. Experimental results
indicate that despite their strong general capabilities, DeepSeek models
exhibit significant safety vulnerabilities across multiple risk dimensions,
including algorithmic discrimination and sexual content. These findings provide
crucial insights for understanding and improving the safety of large foundation
models. Our code is available at
https://github.com/NY1024/DeepSeek-Safety-Eval.