Comment on I'm looking for an article showing that LLMs don't know how they work internally

<- View Parent
Excrubulent@slrpnk.net ⁨3⁩ ⁨days⁩ ago

If all you’re saying is that neural networks could develop consciousness one day,, sure, and nothing I said contradicts that. Our brains are neural networks, so it stands to reason they could do what our brains can do. But the technical hurdles are huge.

You need at least two things to get there:

  1. Enough computing power to support it.
  2. Insight into how consciousness is structured.

1 is hard because a single brain alone is about as powerful as a significant chunk of worldwide computing, the gulf between our current power and what we would need is about… 100% of what we would need. We are so woefully under resourced for that. You also need to solve how to power the computers without cooking the planet, which is not something we’re even close to solving currently.

2 means that we can’t just throw more power or training at the problem. Modern NN modules have an underlying theory that makes them work. They’re essentially statistical curve-fitting machines. We don’t currently have a good theoretical model that would allow us to structure the NN to create a consciousness. It’s not even on the horizon yet.

Those are two enormous hurdles. I think saying modern NN design can create consciousness is like Jules Verne in 1867 saying we can get to the Moon with a cannon because of “what progress artillery science has made in the last few years”.

Moon rockets are essentially artillery science in many ways, yes, but Jules Verne was still a century away in terms of supporting technologies, raw power, and essential insights into how to do it.

source
Sort:hotnewtop