Comment on Perplexity open sources R1 1776, a version of the DeepSeek R1 model that CEO Aravind Srinivas says has been “post-trained to remove the China censorship”.

<- View Parent
brucethemoose@lemmy.world ⁨2⁩ ⁨days⁩ ago

In the 32B range? I think we have plenty of uncensored thinking models there, maybe try fusion 32B.

I’m not an expert though, as models trained from base Qwen have been sufficient for that, for me.

source
Sort:hotnewtop