Comment on Google Researchers’ Attack Prompts ChatGPT to Reveal Its Training Data
Capricorn_Geriatric@lemm.ee 11 months agoBut if it knows everything, it knows nothing. You cannot discern a lie from the truth. It’ll spit something out and it may seem true, but is it really?
Immersive_Matthew@sh.itjust.works 11 months ago
What do you mean if it knows everything it knows nothing? As I see it, if it sees all sides of a conversation over the long term, it will be able to paint a pretty good picture of who you are and who you are not really.
CileTheSane@lemmy.ca 11 months ago
Your friend tells you about his new job:
He sits at a computer and a bunch of nonsense symbols are shown on the screen. He has to guess which symbol comes next. At first he was really bad at it, but over time he started noticing patterns; the symbol that looks like 2 x’s connected together is usually followed by the symbol that looks like a staff.
Once he started guessing accurately on a regular basis they started having him guess more symbols that follow. Now he’s got the hang of it and they no longer tell him if he’s right or not. He has no idea why, it’s just the job they have him.
He shows you his work one day and you tell him those symbols are Chinese. He looks at you like you’re an idiot and says “nah man, it’s just nonsense. It does follow a pattern though: this one is next.”
That is what LLM are doing.
Immersive_Matthew@sh.itjust.works 11 months ago
I would disagree that AI knows nothing. I use ChatGPT plus near daily to code and it went from a hallucinating mess to what feels like a pretty competent and surprisingly insightful service in the months I have been using it. With the rumblings of Q* it only looks like it is getting better. AI knows a lot and very much seems to understand, albeit far from perfect but it surprises me all the time. It is almost like a child who is beyond their years in reading and writing but does not yet have enough life experience to really understand what it is reading and writing…yet.
JohnEdwa@sopuli.xyz 11 months ago
Because language learning models don’t actually understand what is truth or what is real, they just know how humans usually string words together so they can conjure plausible text. If your training data contains falsehoods, it will learn to write them.
To get something that would benefit from knowing both sides, we’d need to first create a proper gai, artificial general intelligence, with the ability to actually think.
Immersive_Matthew@sh.itjust.works 11 months ago
I sort of agree. They do have some level of right and wrong already, it is just very spotty and inconsistent in the current models. As you said we need AGI level AI to really address the shortcomings which sounds like it is just a matter of time. Maybe sooner than we are all expecting.