Comment on I'm looking for an article showing that LLMs don't know how they work internally

<- View Parent
Voldemort@lemmy.world ⁨2⁩ ⁨days⁩ ago

We’re on the same page about consciousness then. My original comment only pointed out that current AI have problems that we have because they replicate how we work and it seems that people don’t like recognising that very obvious fact that we have the exact problems that LLMs have. LLMs aren’t rational because we inherently are not rational. That was the only point I was originally trying to make.

For AGI or UGI to exist, massive hurdles will need to be made, likely an entire restructuring of it. I think LLMs will continue to get smarter and likely exceed us but it will not be perfect without a massive rework.

Personally and this is pure speculation, I wouldn’t be surprised if AGI or UGI is only possible with the help of a highly advanced AI. Similar to how microbiologist are only now starting to unravel protein synthesis with the help of AI. I think the shear volume of data that needs processing requires something like a highly evolved AI to understand, and that current technology is purely a stepping stone for something more.

source
Sort:hotnewtop