Comment on We need to stop pretending AI is intelligent
tmpod@lemmy.pt 19 hours agoThat is not really true. Yes, there are jump instructions being executed when you run interference on a model, but they are in no way related to the model itself. There’s no translation of weights to jumps in transformers and the underlying attention mechanisms.
I suggest reading …wikipedia.org/…/Transformer_(deep_learning_archi…
Blue_Morpho@lemmy.world 18 hours ago
The model is data. It needs to be operated on to get information out. That means lots of JMPs.
If someone said viewing a gif is just a bunch of if-else’s, that’s also true. That the data in the gif isn’t itself a bunch of if-else’s isn’t relevant.
Executing LLM’S is particularly JMP heavy. It’s why you need massive fast ram because caching doesn’t help them.
tmpod@lemmy.pt 16 hours ago
You’re correct, but that’s like saying along the lines of manufacturing a car is just bolting and soldering a bunch of stuff. It’s technically true to some degree, but it’s very disingenuous to make such a statement without being ironic. If you’re making these claims, you’re either incompetent or acting in bad faith.
I think there is a lot wrong with LLMs and how the public at large uses them, and even more so with how companies are developing and promoting them. But to spread misinformation and polute an already overcrowded space with junk is irresponsible at best.