That’s the great thing about open models. Censorship? All it takes is one person and a bit of cash to get rid of it.
Perplexity open sources R1 1776, a version of the DeepSeek R1 model that CEO Aravind Srinivas says has been “post-trained to remove the China censorship”.
Submitted 1 month ago by Cat@ponder.cat to technology@lemmy.world
https://www.perplexity.ai/hub/blog/open-sourcing-r1-1776
Comments
brucethemoose@lemmy.world 1 month ago
Even_Adder@lemmy.dbzer0.com 1 month ago
Can’t wait to try a distillation. The full model is huge.
brucethemoose@lemmy.world 1 month ago
In the 32B range? I think we have plenty of uncensored thinking models there, maybe try fusion 32B.
I’m not an expert though, as models trained from base Qwen have been sufficient for that, for me.
thefluffiest@feddit.nl 1 month ago
Great. Has it also removed American censorship and propaganda?
DarkSpectrum@lemmy.world 1 month ago
I believe this is what was added
FrankLaskey@lemmy.ml 1 month ago
I think we can all agree that modifications to these models which remove censorship and propaganda on behalf of one particular country or party is valuable for the sake of accuracy and impartiality, but reading some of the example responses for the new model I honestly find myself wondering if they haven’t gone a bit further than that by replacing some of the old non-responses and positive portrayals of China and the CPC with a highly critical perspective typified by western governments which are hostile to China (in particular the US).
ZILtoid1991@lemmy.world 1 month ago
LLMs are gigantic bias reproduction machines, so it will never be perfect.
iopq@lemmy.world 1 month ago
What part is highly critical of China? Facts can’t be critical
fruitycoder@sh.itjust.works 1 month ago
Listen, I’m highly critical of the CCP, but LLMs aren’t facts machines, they are make text like what they are trained on machines.
They have no grasp of truth, and we can only get some sense of truth of what the average collective text response of its dataset (at best!).
brucethemoose@lemmy.world 1 month ago
Well you can merge it with the original model, to any degree, to get any sliding scale of “bias” you want.
Practically, though, I guess that’s not super practical, as very few have the hardware or cash to deploy a custom fill R1 themselves.
vrighter@discuss.tchncs.de 1 month ago
not remove ic replace.
Also, stop calling releasing binary blobs of weights as open source
brucethemoose@lemmy.world 1 month ago
It’s honestly not that big a deal, as it’s not like knowing anything about how it was trained (beyond the config) would help you modify it. It’s still highly modifiable. It’s not like anyone can afford to replicate it.
It would be nice to publish the hyperparameters for research purposes, but… shrug.
I think a subset of the exact training data/hyperparameters would help with quantization-aware-training, maybe, but that’s all I got.
GissaMittJobb@lemmy.ml 1 month ago
My god, could he have chosen a name more cringe than this? Fuck.
MunkyNutts@lemmy.world 1 month ago
An attempt to appeal to U.S. conservatives maybe? You know all that 1776, 3%, ΜΟΛΩΝ ΛΑΒΕ etc. sticker suckers.
GissaMittJobb@lemmy.ml 1 month ago
From what I gather from the Perplexity CEO, he is just that type of Musk-tier cringelord, so yes, probably
Beldarofremulak@discuss.online 1 month ago
“Pat’rats”
ZILtoid1991@lemmy.world 1 month ago
IDK, but this seems like wankery to me. Just google it if you want to know about it, the AI isn’t an “all knowing being” nor “the arbitrer of truth”.
I have a feeling that a new logical fallacy will soon emerge (if it isn’t already widespread on certain places of the internet), that will be “X is true because the LLM said so”.
fruitycoder@sh.itjust.works 1 month ago
It’s really an extension of “Would some really do that? Just lie on the Internet?” But now “Would AI, which is built to create content like what people post on the Internet, really just lie?”
KeenFlame@feddit.nu 1 month ago
Seems like almost everyone understands that it hallucinates.
ivanafterall@lemmy.world 1 month ago
Ctrl + F
Find: Chinese
Replace: God-damned Chinese
New model’s ready!
biofaust@lemmy.world 1 month ago
I run an uncensored version on my PC since weeks, there are multiple ones on HuggingFace.
brucethemoose@lemmy.world 1 month ago
Not full R1, which is developed differently than any of the distllations.
Korkki@lemmy.ml 1 month ago
Now that it has been “freedomized” it instead says that Israel is actually 3000+ years old and Palestinians are invaders, and Israel has the right to defend itself like chatgpt and that American style liberal democracy is the peak of human development and civilization. Don’t kid yourself that this has anything to do with truth or making it more “accurate”. Yeah just replace one set of official truths, half thrush with other set of official positions, half truths and outright lies to plug the gaps. Again, who fact checks the factcheckers? Even just out of spite I would not use any model that advertises itself as being trained to answer to the sensibilities of a western techbro liberals so that they can once more outsource their thinking to outside party, because at-least it isn’t the CCP.
I just don’t get it man. What kind of non thinking cretin purposefully uses this kind of model?
Squizzy@lemmy.world 1 month ago
My chatgpt will list the questionable human righfs record of Israel ad agree with the conclusion that they are, under the same standards and logic, likely guilty of their accused atrocities as thenlikes of China or Russia.