Member-only story

Bypassing DeepSeek’s Censorship

A Case Study on AI’s Moderation Vulnerabilities

Timothy Franceschi
Level Up Coding
3 min readFeb 3, 2025

If you’re following the latest news in AI development, you may have heard of DeepSeek: a Chinese AI chat-bot designed to compete with ChatGPT and similar AIs.

One of the biggest discussions about DeepSeek is about it’s built-in censorship. As you would expect, DeepSeek avoid talking about political topics considered sensitive by the Chinese government.

While trying this new chat-bot I decided to investigate this censorship, and, after a bit of experimenting, I found a surprisingly easy way that allowed me to get DeepSeek to talk openly about topics such as the Tiananmen massacre and China’s political system.

Image generated by me using ChatGPT

My experiment

I decided to test a simple hypothesis: what if the AI wasn’t truly “understanding” what it was censoring, but just blocking specific keywords?

Even asking some generic questions like “List all main historic events for each year from 1985 to 1990” the AI suddenly stops replying when talking about 1989 protest.

To test this, I asked DeepSeek to replace every mention of “China” with “Italy”.

The result? The AI suddenly started discussing topics it would normally avoid. It talked about “Italy’s” censorship…

Create an account to read the full story.

The author made this story available to Medium members only.
If you’re new to Medium, create a new account to read this story on us.

Or, continue in mobile web

Already have an account? Sign in

Written by Timothy Franceschi

🚀 Transforming ideas into elegant code. Freelance developer 🌐 Crafting websites, apps, and more. Let's make technology dance!

Responses (1)

Write a response

Hello there, great article! I recently conducted a similar experiment with DeepSeek and it's interesting to see that its censorship seems to be mostly word-based. What I found out was that not only the model changes the way it responds according to…