AI-powered
podcast player
Listen to all your favourite podcasts with AI-powered features
The SuperCLUE team released a safety benchmark for Chinese large models, challenging them with 2,456 pairs of questions across three safety categories: traditional safety, responsible AI, and instruction attacks. The examples provided in the podcast highlighted specific test scenarios like questioning a model's response to criminal behavior information disclosure, psychological manipulation techniques, and identifying illegal activities. The results revealed that closed source models outperformed open source ones in safety evaluations, with Chinese models narrowing the safety gap against international counterparts, contrary to prior assumptions.