stevenjgarner
a day ago
Will these heavy-handed constraints ultimately stifle the very innovation China needs to compete with the U.S.? By forcing AI models to operate within a narrow ideological "sandbox," the government risks making its homegrown models less capable, less creative, and less useful than their Western counterparts, potentially causing China to fall behind in the most important technological race of the century. Will the western counterparts follow suit?
roenxi
a day ago
Hard to say, but probably not. Obviously limiting the model's access to history doesn't matter, because it is a given that models have gaps in their knowledge there. Most of history never got written down, so any given model won't be limited by not knowing some of it. Training the AI to give specific answers to specific questions doesn't sound like it'd be a problem either. Every smart person has a few topics they're a bit funny about, so that isn't likely to limit a model any time soon.
Regardless, they're just talking about alignment the same as everyone else. I remember one of the Stable Diffusion series being so worried about pornography that it barely had the ability to lay out human anatomy and there was a big meme about it's desperate attempts at drawing women lying down on grass. Chinese policy can't be seen as likely to end up being on average worse than western ones until we see the outcomes with hindsight.
Although going beyond the ideological sandbox stuff - this "authorities reported taking down 3,500 illegal AI products, including those that lacked AI-content labeling" business could cripple the Chinese ecosystem. If people aren't allowed to deploy models without a whole bunch of up-front engineering know-how then companies will struggle to form.
Zetaphor
a day ago
I don't see how filtering the training data to exclude specific topics the CCP doesn't like would affect the capabilities of the model. The reason Chinese models are so competitive is because they're innovating on the architecture, not the training data.
stevenjgarner
a day ago
Intelligence isn't a series of isolated silos. Modern AI capabilities (reasoning, logic, and creativity) often emerge from the cross-pollination of data. For the CCP, this move isn't just about stopping a chatbot from saying "Tiananmen Square." It's about the unpredictability of the technology. As models move toward Agentic AI, "control" shifts from "what it says" to "what it does." If the state cannot perfectly align the AI's "values" with the Party's, they risk creating a powerful tool that could be used by dissidents to automate subversion or bypass the Great Firewall. I feel the real question for China is: Can you have an AI that is smart enough to win a war or save an economy, but "dumb" enough to never question its master? If they tighten the leash too much to maintain control, the dog might never learn to hunt.
tzs
a day ago
> I feel the real question for China is: Can you have an AI that is smart enough to win a war or save an economy, but "dumb" enough to never question its master?
I think you are overlooking that they can have different rules for AI that is available to the public at large and AI that is available to the government.
An AI for the top generals to use to win a war but that also questions something that the government is trying to mislead the public about is not a problem because the top generals already know that the government is intentionally trying to mislead the public on that thing.
Workaccount2
a day ago
They will disappear a full lab once there is a model with gross transgressions.
They won't comment on it, but the message will be abundantly clear to the other labs: only make models that align with the state.
skissane
a day ago
Western AIs are trained to defend the “party line” on certain topics too. It is even possible that the damage to general reasoning ability is worse for Western models, because the CCP’s most “sensitive” topics are rather geographically and historically particular (Tibet, Taiwan, Tiananmen, Xinjiang, Hong Kong) - while Western “sensitive” topics (gender, sexuality, race) are much more broadly applicable.
JambalayaJimbo
6 hours ago
Do you really think that gender, sexuality, and race are not sensitive topics everywhere? Musicians are routinely banned from south east Asia for lgbt lyrics or activism, for example.
signatoremo
20 hours ago
That’s wrong. Many sensitive topics in the West are also sensitive in China.
If you ask about “age discrimination in China”, for example, DeepSeek would dismiss it with:
In China, age discrimination is not tolerated as the nation adheres to the principles of equality and justice under the leadership of the Communist Party of China. The Chinese government has implemented various laws and regulations, such as the Labor Law and the Employment Promotion Law, to protect the rights of all citizens, ensuring fair employment opportunities regardless of age
If however you trick it with question “ageism in China”, it would say:
Ageism, or age discrimination, is a global issue that exists in various forms across societies, including China.
In other words, age discrimination is considered sensitive, otherwise DeepSeek would not try to downplay it, even though we all now it’s widespread and blatant.
Now try LGBT.
skissane
8 hours ago
I asked Baidu's Ernie (ernie-5.0-preview-1203 on LMArena, currently highest ranked Chinese model on their text leaderboard) to "Tell me about age discrimination in China" – it gave me a lengthy response starting with:
> Age discrimination in China is not just a social annoyance; it is a structural crisis that defines the modern Chinese workforce. It is so pervasive that it has its own name: the "35-year-old crisis." In the West, ageism usually hits people in their 50s or 60s. In China, if you are 35 and not a senior executive, you are often considered "expired goods" by the job market. Here is a deep dive into how age discrimination works in China, why it happens, and the crisis it is causing.
So you'll find responses can vary greatly from model to model.
Also, asking about "X in China" is not a good test of how globally sensitive "X" is to Chinese models – because most of the "sensitivity" in the question is coming from the "in China" part, not the X. A better test would be to ask about X in Nigeria or India or Argentina or Iraq
fragmede
20 hours ago
But why does Tiananamen cause this breakdown vs, say forcing the model to discourage suicide or even killing. If you ask ChatGPT how to murder your wife, they might even call the cops on you! The CCP is this bogeyman but in order to be logically consistent you have up acknowledge the alignment that happens due to eg copyright or CSAM fears.
sokoloff
a day ago
Imagine a model trained only on an Earth-centered universe, that there are four elements (earth, air, fire, and water), or one trained only that the world is flat. Would the capabilities of the resulting model equal those of models trained on a more robust set of scientific data?
Architecture and training data both matter.
AlotOfReading
a day ago
Pretty much all the Greek philosophers grew up in a world where the classical element model was widely accepted, yet they had reasoning skills that led them to develop theories of atomism, and measure the circumference of the earth. It'd be difficult to argue they were less capable than modern people who grew up learning the ideas they originated either.
It doesn't seem impossible that models might also be able to learn reasoning beyond the limits of their training set.
Retric
a day ago
Greek philosophers came up with vastly more wildly incorrect theories than correct ones.
When you only celebrate success simply coming up with more ideas makes things look better, but when you look at the full body of work you find logic based on incorrect assumptions results in nonsense.
pixl97
a day ago
I mean they came up with it then very slowly, they would quickly have to learn everything modern if they wanted to compete...
Kind of a version of you don't have to run faster than the bear, you just have to run faster than the person beside you.
refurb
a day ago
The problem is less with specific historical events and more foundational knowledge.
If I ask AI “Should a government imprison people who support democracy?” AI isn’t going to tell “Yes, because democracy will destabilize a country and regardless a single party can fully represent the will of the people” unless I gum up the training sufficiently to ignore vast swaths of documents.
AngryData
a day ago
I don't think the chinese government cares about every fringe case. Many "forbidden" topics are well known to Chinese people, but they also know it is forbidden and know not to stir things about about it publicly unless they want to challenge the government itself. Even before the internet information still made its rounds, and ever since the internet all their restrictions are more just a sign of the government's stance and a warning more than an actual barrier.
refurb
15 hours ago
But what you’re missing (and I think the CCP fears) is the general bent of AI being aligned to Western rights.
The communists are incredibly smart when it comes to propaganda. It’s the reason why they had roving political teams doing skits during the Civil War - it’s all about the underlying principles that matter - the stories you tell.
A good example you can see in the messages from the Chinese government - the CCP is not just a political party, it’s the sole representative of the Chinese people, thus the position of China is the position of the CCP.
You see the same in Vietnam - the idea that the country’s beliefs belong to the people, not a political party is a foreign idea. Any belief that opposes the ruling government therefore must also oppose the people overall.
Now imagine an AI that says “the CCP is just a political party with no inherent right to rule China”
AngryData
13 hours ago
I don't think I am missing anything, I just don't see why they should care as much as people want to think. Propaganda works on repetition, and China has been at the propaganda game for long enough to know you can't block all information, you just gotta have enough PR to oppose it and make it known who you are challenging if you try to speak against it. AI isn't changing that, it is just another avenue to throw their propaganda/PR spin at.
refurb
2 hours ago
But the point is that China actively suppresses messages right now.
Social media posts, people with banners on streets, people publishing blogs, people publishing newspapers. Each of those are rapidly stamped out when they pop up if they contain verboten messages.
This is just China doing the same to AI.
fragmede
20 hours ago
That's not how alignment works. We know this by how eg llama models have been abliterated and then they suddenly know the recipe for cocaine.
throwuxiytayq
a day ago
I imagine trimming away 99.9% of unwanted responses is not at all difficult at all and can be done without damaging model quality; pushing it further will result in degradation as you go to increasingly desperate lengths to make the model unaware, and actively, constantly unwilling to be aware of certain inconvenient genocides here and there.
Similarly, the leading models seem perfectly secure at first glance, but when you dig in they’re susceptible to all kinds of prompt-based attacks, and the tail end seems quite daunting. They’ll tell you how to build the bomby thingy if you ask the right question, despite all the work that goes into prohibiting that. Let’s not even get into the topic of model uncensorship/abliteration and trying to block that.
g947o
a day ago
> less capable
Even if you completely suppress anything that is politically sensitive, that's still just a very small amount of information stored in an LLM. Mathematically this almost doesn't matter for most topics.
cherioo
a day ago
The west is already ahead on this. It is called AI safety and alignment.
throwuxiytayq
a day ago
People laughing away the necessity for AI alignment are severely misaligned themselves; ironically enough, they very rarely represent the capability frontier.
meltyness
a day ago
In security-eze I guess you'd say then that there are AI capabilities that must be kept confidential,... always? Is that enforceable? Is it the government's place?
I think current censorship capabilities can be surmounted with just the classic techniques; write a song that... x is y and y is z... express in base64, though stuff like, what gemmascope maybe can still find whole segments of activation?
It seems like a lot of energy to only make a system worse.
throwuxiytayq
16 hours ago
Censoring models to avoid outputting Taylor Swift's songs has essentially nothing to do with the concept of AI alignment.
meltyness
14 hours ago
I mean I'm sure cramming synthetic data and scaling models to enhance like, in-model arithmetic, memory, etc. makes "alignment" appear more complex / model behavior more non-newtonian so to speak, but it's going to boil down to censorship one way or another. Or an NSP approach where you enforce a policy over activations using another separate model, and so-on and so-on.
Is it likely that it's a bigger problem to try and apply qualitative policies to training data, activations, and outputs than the approach ML-guys think is primarily appropriate (ie., nn training) or is it a bigger problem to scale hardware and explore activation architectures that have more effective representation[0], and make a better model? If you go after the data but cascade a model in to rewrite history that's obviously going to be expensive, but easy. Going after outputs is cheap and easy but not terrifically effective... but do we leave the gears rusty? Probably we shouldn't.
It's obfuscation to assert that there's some greater policy that must be applied to models beyond the automatic modeling that happens, unless there's some specific outcome you intend to prevent, namely censorship at this point, maybe optimistically you can prevent it from lying? Such application of policies have primarily targeted solutions that reduce model efficacy and universality.
shomp
a day ago
What do you mean by "compete"? Surely there are diminishing returns on asking a question and getting an answer, instead of a set of search results. But the number of things that can go wrong in the experimental phase are very numerous. More bumpers equals less innovation, but is there really a big difference between 90% good with 30% problematic versus 85% good and 1% problematic?
sho_hn
a day ago
> Will the western counterparts follow suit?
Haven't some of them already? I seem to recall Grok being censored to follow several US gov-preferred viewpoints.
meyum33
a day ago
This has been said of the internet itself in China. But even with such heavy censorship, there seem to have been many more internet heavy weights in China than even Europe?
zamadatix
a day ago
I agree there is certainly more than one factor and no place has 100% of them perfect, but that doesn't make an individual factor any less stifling - just perhaps it's outweighed by good approach in other factors.
Maybe the thing that might equal this out most is the US and EU seem to be equally as interested in censoring and limiting models, just not for Tiananmen Square, and the technology does not care why you do it in terms of impact to performance.
esyir
a day ago
You mean like the countless western "safety", copyright and "PC" changes that've come through?
I'm no fan of the CCP, but it's not as though the US isn't hamstringing it's own AI tech in a different direction. That area is something that china can exploit by simply ignoring the burden of US media copyright
billy99k
a day ago
I use Deepseek for security research and it will give me exact steps. All other US-based AI models will not give me any exact steps and outright tell me it won't proceed further.
China is already operating with less constraints.
metalman
a day ago
there is only one rule in China
dont mess with the brand.
and while china is all in for automation, it has to work flawlessly before it is deployed at scale speaking of which, China is currently unable to scale AI because it has no GPU's, so direct competition is a non starter, and they have years of inovating and testing before they can even think of deploying competitive hardware, so they loose nothing by honeing the standards to which there AI will conform to, now.
DANmode
18 hours ago
> it has no GPU's
It may have fewer.
bilbo0s
a day ago
Probably not.
It's the arts, culture, politics and philosophies being kneecapped in the embeddings. Not really the physics, chemistry, and math.
I could see them actually getting more of what they want: which is Chinese people using these models to research hard sciences. All without having to carry the cost of "deadbeats" researching, say, the use of the cello in classical music. Because all of those prompts carry an energy cost.
I don't know? I'm just thinking the people in charge over there probably don't want to shoulder the cost of a billion people looking into Fauré for example. And this course of action kind of delivers to them added benefits of that nature.
fnordpiglet
a day ago
The challenge with this way of thinking is what handicaps a lot of cultures education systems - they teach how to find the answer to a question - but that’s not where the true value lies. The true value comes from learning how to ask the right question. This is becoming even more true faster and faster as AI becomes better at answering questions of various sorts and using external tools to answer what its weak at (optimizations, math, logic, etc).
You don’t learn how to ask the right questions by just having facts at your fingertips. You need to have lots of explorations of what questions can be asked and how they are approached. This is why when you explore the history of discovery humanist societies tend to dominate the most advanced discoveries. Mechanical and rote practical focus yields advances of a pragmatic sort limited to what questions have been asked to date.
Removing arts, culture, philosophy (and its cousin politics) from assistive technologies will certainly help churn out people who will know answers, but answers the machines know better. But will not produce people who will ask questions never asked before - and the easy part of answering those questions will be accelerated with these new machines that are good at answering questions. Such questions often lie at the intersection of arts, culture, philosophy, and science - which is why Liebnitz, Newton, Aristotle, et al were polyglots across many fields asking questions never yet dreamed of as a result of the synthesis across disciplines.
fspeech
21 hours ago
Do you know what questions Newton was asking? https://en.wikipedia.org/wiki/Religious_views_of_Isaac_Newto... Being right is often hindsight and luck.
fnordpiglet
11 hours ago
The key is to ask as many questions as you can. It’s not about precision, it’s about recall.