What are good safety standards for open source AIs from China?
post by ChristianKl · 2025-04-12T13:06:16.663Z · LW · GW · 2 commentsContents
2 comments
It looks like the compromising between those who want to ban Chinese AI models in the United States in the c rrent administration will be to allow the models but require some safety standards to prevent China from harming the United States.
While we generally would want more AI regulation than the current administration is willing to create, this is a window where the AI safety community potentially can affect safety policy.
Within the existing political constraints, what standards for the Chinese models should we wish for?
2 comments
Comments sorted by top scores.
comment by StanislavKrym · 2025-04-12T13:35:28.039Z · LW(p) · GW(p)
If they are open-source, then doesn't it mean that anyone can check how the models' alignment is influenced by training or adding noise? Or does it mean that anyone can repeat the training methods?
Replies from: robert-k↑ comment by Mis-Understandings (robert-k) · 2025-04-12T13:50:03.329Z · LW(p) · GW(p)
Generally the releases of the "open source" models release the inference code and the weights, but not the exact training data, and often not information about training setup. (for instance, Deepseek has done a pile of hacking on how to get the most out of their H800s, which is private)