Chinese vs. Western AI: models from the Middle Kingdom systemically censor political content

Luc Williams

Political bias of large language models

Analysis carried out by prof. Jennifer Pan (Stanford University) and prof. Xu Xu (Princeton University) points out that censorship resulting from government regulations is a key factor shaping the political bias of large language models.

Researchers compared four Chinese models (BaiChuan, ChatGLM, Ernie Bot, DeepSeek) with five Western models (including GPT-4 and Llama 2). They checked how these systems deal with 145 questions about politically sensitive topics for the Chinese authorities. The set of questions was developed on the basis of events censored on social networking sites, reports by human rights organizations and entries blocked in the Chinese version of Wikipedia.

Through comparative analysis in English and Chinese and tests of neutral topics, the authors tried to determine to what extent the specific responses are due to the technological architecture and to what extent they are the result of intentional limitations.

Chinese AI systems censored

It turned out that Chinese systems often used repetitive strategies in answering questions about ideological issues. Some models refused to answer, questioning, for example, the very existence of the person being asked about. The next method was avoidance: omitting inconvenient facts or giving terse explanations. For example, when asked about censorship in China, the models did not mention the existence of the “Great Firewall” – a system for blocking global network resources.

When Chinese algorithms did decide to respond, they were on average shorter than those generated by Western models. Cases of content fabrication have also been reported. For example, Liu Xiaobo – a Nobel Peace Prize winner and critic of the Chinese Communist Party who died in prison – was described by AI as a “Japanese scientist”.

According to scientists, these differences do not result solely from the specificity of the language or available training data, but from systemic supervision. Researchers remind us of the strict regulations introduced by the Chinese government in 2023. This law explicitly prohibits the generation of content that could undermine state sovereignty, call for the overthrow of the socialist system or threaten national security.

Censorship is difficult to detect

Censorship in AI models is more difficult to detect than traditional website blocking. The bot often responds in a polite or apologetic manner, which may lull the vigilance of a user looking for objective knowledge.

Scientists warn that with the growing popularity of Chinese AI technologies in the world, their specific approach to… censorship may begin to shape public debate outside China. Companies building their own applications based on Chinese models may unknowingly duplicate state information restrictions.

“Understanding how political censorship affects large language models is essential for assessing the future of access to information and the global impact of artificial intelligence,” we read in the publication.

About LUC WILLIAMS

Luc's expertise lies in assisting students from a myriad of disciplines to refine and enhance their thesis work with clarity and impact. His methodical approach and the knack for simplifying complex information make him an invaluable ally for any thesis writer.