Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

AI's answers on China differ depending on the language, analysis finds | TechCrunch

Mar 20, 2025 - techcrunch.com
A study by a developer known as "xlr8harder" has revealed that AI models, including those developed by Chinese labs like DeepSeek, exhibit varying levels of censorship based on the language used to prompt them. The study found that models such as DeepSeek’s R1 and Alibaba’s Qwen 2.5 72B Instruct are less likely to respond to politically sensitive questions when asked in Chinese compared to English. This discrepancy is attributed to "generalization failure," where models trained on politically censored Chinese text are less likely to generate critical responses in Chinese. Experts agree that the uneven compliance is due to differences in training data and the effectiveness of guardrails across languages.

The findings highlight broader challenges in AI development, such as the tension between creating general models versus culturally specific ones and the difficulty in achieving cultural reasoning. Experts like Chris Russell and Vagrant Gautam suggest that the models' behavior is influenced by the availability of critical content in different languages, with more English-language criticism of the Chinese government available online. Geoffrey Rockwell and Maarten Sap emphasize the need for AI models to better understand socio-cultural norms and the ongoing debates over model sovereignty and influence in the AI community.

Key takeaways:

  • AI models developed by Chinese labs like DeepSeek censor politically sensitive topics, with a 2023 measure forbidding content that damages the unity of the country and social harmony.
  • Language affects AI model responses, with models like Claude 3.7 Sonnet and Qwen 2.5 72B Instruct being less likely to answer politically sensitive questions in Chinese compared to English.
  • Experts suggest that the uneven compliance in different languages may be due to generalization failure, as Chinese training data is more likely to be politically censored.
  • The findings highlight ongoing debates in the AI community about model sovereignty, cultural competence, and cross-lingual alignment.
View Full Article

Comments (0)

Be the first to comment!