Comment on Ex-OpenAI star Sutskever shoots for superintelligent AI with new company
Alphane_Moon@lemmy.world 4 months agoI don’t consider tech company boardroom drama to be an indicator of anything (in of itself). This is not some complex dilemma around morality and “doing the right thing”.
Is my take on their PR copytext unreasonable? Is my interpretation purely a matter of subjectivity?
Why should I buy into this “AI god-mommy” and “skynet” stuff? Guy can’t even provide a definition of “superintelligence”. Seems very suspicious for a “top mind in AI” (paraphrasing your description).
Don’t get me wrong, I am not saying he acts like a movie antagonist IRL, but that doesn’t mean we have any reason to trust his motives or ignore the long history of similar proclamations.
webghost0101@sopuli.xyz 4 months ago
No i applaud a healthy dose of skepticism.
I am everything but in favor of idolizing silicon valley gurus and tech leaders but from Sutskeva i have seen enough to know he is one of the few to pay attention to
Artificial Super intelligence or ASI is the step beyond AGI (artificial general intelligence)
The later is equal or better in capacity to a real human being in almost all fields.
Artificial Super intelligence is defined (long before openai was a thing) as transcending human intelligence in every conceivable way. At which point its a fully independent entity that can no longer be controlled or shutdown.
Alphane_Moon@lemmy.world 4 months ago
Thank you for the clarification regarding ASI. That still leaves the question of the definition of “safe ASI”; a key point that is emphasized in their manifesto.
To use your example it’s like an early mass market car industry professional (say in 1890) discussing road safety and ethical dilemmas in roads dominated by regular drivers and a large share of L4/L5 cars (with some of them being used as part-time taxis). I just don’t buy it.
Mind you I am not anti-ML/AI. I am an avid user of “AI” (ML?) upscaling (specifically video) and to lesser extent stable diffusion. While AI video upscaling is very fiddly and good results can be hard to get right, it is clearly on another level with respect to quality compared to “classical” upscaling algorithms. I was truly impressed when I was able to run by own SD upscale with good results.
What I am opposed to is oligarchs, oligarch-wanabees, shallow sounding proclamations of grandiose this or that. As far as I am concerned it’s all bullshit and they are all to one degree or another soulless ghouls that will eat your children alive for the right price and the correct mental excuse model (I am only partially exaggerating, happy to clarify if needed) .
If one has all these grand plans for safe ASI, concern for humanity and whatnot, setup a public repo and release all your code under GPL (and all relevant documentation, patent indemnification, no trademark tricks etc.). Considering Sutskever’s status as AI royalty who is also allegedly concerned about humanity, he would be the ideal person to pull this off.
If you can’t do that, then chances are you’re lying about your true motives. It’s really as simple as that.
webghost0101@sopuli.xyz 4 months ago
No need to clarify what you meant with the oligarchs theres barely any exaggeration there. Ghouls is quite accurate.
Considering the context of a worst case possible scenario (hostile takeover by an artificial superior) which honestly is indistinguishable from general end of the world doomerism prophecies but very alive in the circles of Sutskeva i believe safe ai consistent of the very low bar of
“humanity survives wile agi improves the standards of living worldwide” of course for this i am reading between the lines based on previously aquired information.
One could argue that If ASI is created the possibilities become very black and white:
ASI is indifferent about human beings and pursues its own goals, regardless of consequences for the human race. It could even find a way off the planet and just abandon us.
ASI is malaligned with humanity and we become but a. Resource, treating us no different then we have historically treated animals and plants.
ASI is aligned with humanity and it has the best intentions for our future.
For either scenario it would be impossible to calculate its intentions because by definition its more intelligent then all of us. Its possible that some things that we understand as moral may be immoral from a better informed perspective, and vice versa.
The scary thing is we wont be able to tell wether its malicious and pretending to be good. Or benevolent and trying to fix us. Would it respect consent if say a racist refuses therapy?
Of course we can just as likely hit a roadblock next week and the whole hype dies out for another 10 years.