Comment on Elon Musk wants to rewrite "the entire corpus of human knowledge" with Grok

<- View Parent
brucethemoose@lemmy.world ⁨20⁩ ⁨hours⁩ ago

There’s some nuance.

Using LLMs to augment data, especially for fine tuning (not training the base model), is a sound method. The Deepseek paper using, for instance, generated reasoning traces is famous for it.

Another is using LLMs to generate logprobs of text, and train not just on the text itself but on the *probability a frontier LLM sees in every ‘word.’ This is called distillation, though there’s some variation and complication.

But yes, the “dumb” way, aka putting data into a text box and asking an LLM to correct it, is dumb and dumber, because:

In other words, Musk has no idea WTF he’s talking about.

source
Sort:hotnewtop