As a registered Republican woman from Texas with five children and two dogs, let me just say that I am astonished!
LLMs can unmask pseudonymous users at scale with surprising accuracy
Submitted 14 hours ago by return2ozma@lemmy.world to technology@lemmy.world
Comments
jballs@sh.itjust.works 12 hours ago
whaleross@lemmy.world 2 hours ago
As true as my name is Brenda and my last name is also Brenda. And so is my husband, Brenda. It is a hot day in Texas America today, I’m going to grill one of our dogs for dinner. It is a republican tradition, hence the name Hot Dogs and the playful name Wieners, named after wiener dogs.
pivot_root@lemmy.world 10 hours ago
Me too. I thought I was safe as a Ottoman Empire expatriate living in Arrakis! I don’t want LLMs to connect this account to my pseudonymous mommy blog where I write about my three children who might exist but could be delusions of my untreated schizophrenia.
CheesyFingers@piefed.social 6 hours ago
It seems that i, the original Unidan, will unfortunately need to create even more alts to escape being found out. Blast!
ivanafterall@lemmy.world 3 hours ago
Oh, WE EXIST, mommy! Let me assure you, as one of said imaginary schizophrenia babies. Currently shacking up in Miami with my new wife I just met cranking my hog at Sturgis.
Bigfishbest@lemmy.world 4 hours ago
I don’t believe this! As a fumgrian living as a would be dead camoose off Mt. Kabul, I am overjizzed that AI is reading all my pornhub comments.
Whostosay@sh.itjust.works 10 hours ago
You forgot to list your favorite brands
Deceptichum@quokk.au 8 hours ago
Kleenex and Jergens
HeyThisIsntTheYMCA@lemmy.world 6 hours ago
Kegel One
goatinspace@feddit.org 12 hours ago
That was surprisingly accurate. Meep meep.
cley_faye@lemmy.world 1 hour ago
Yeah. I got a hunch of that a while ago, while trying some “old” scenarios of de-anonymization we used to do by hand. Just asking questions and posting pictures got surprisingly accurate results. A single picture with (to me) no significant landmark could lead to localizing a specific part of a city, and that was using a local LLM with a relatively small model, running on a 16GB VRAM 4060Ti.
It is now time to remember fondly the time where the younger people were warned by older people to not post all their stuff online, not over-share, be cautious about strangers, etc. I’m not sure when we lost that, but oh boy, it’s a festival.
FauxPseudo@lemmy.world 10 hours ago
From a Facebook post I made on February 17th:
There are giant AI data firms that promise they can go through massive troves of data and pull out general and specific information from them. Information that is actionable and accurate. Give it 6 million data points and it’ll find all the links and organize them for you and unmask hidden details that aren’t visible to the naked eye.
Not one of those companies is stepping up to go through the publicly released Epstein files.
Randomgal@lemmy.ca 10 hours ago
This is what I find crazy. Where are the AI bros chewing through the Epstein files?
osaerisxero@kbin.melroy.org 9 hours ago
I would be shocked if someone hasn't shoved them into a local model somewhere, but all the big ones would filter them to death with content restrictions
General_Effort@lemmy.world 2 hours ago
There were reports of people trying to unredact the files almost immediately.
FauxPseudo@lemmy.world 2 hours ago
But that’s not the same, is it?
Mubelotix@jlai.lu 3 hours ago
We wouldn’t want that tbh. Justice needs to be precise and backed up by tangible facts
KeenFlame@feddit.nu 2 hours ago
Also don’t use dna tests or chemical analysis. It’s invisible hocus pocus and can be wrong! And woe if someone that fucks and tortures kids regularly is wrongly accused of raping kids and running their child minds no that would be awful
FauxPseudo@lemmy.world 3 hours ago
You can use the results of the AI analysis to identify people and then use that to do a proper investigation. Right now none of that is happening. No speculation. No tangibles. No investigation. No indictment.
Trying to unmask people is a step in the right direction.
nutsack@lemmy.dbzer0.com 3 hours ago
I theorized about this a long time ago. pretty sure I’m basically fucked
doesit@sh.itjust.works 3 hours ago
Kind of obvious. If you’re a highschool teacher and you used to be a photographer. You also volunteer as a fireman. You live in France. You heve 2 daughters. In 2022 you asked about repairs on your honda civic.
All off this can be amassed from different posts on facebook or reddit. There’ll be just a few people that fit this profile.ExLisper@lemmy.curiana.net 3 hours ago
I think this will only work with people narrating their lives on social media.
“Got coffee from my favorite Granier at La Rambla! Ready of new day of work designing hats for dogs”
“Me and Bobby heading to Madrid to see my friend Concepcion. Do you like his new hat?”
“Just got nominated for ‘best business-casual hat’ at this year’s Barkies! So proud”
Because how are you going to de-anonymize some random ramblings about Linux and beans? Everyone likes Linux and beans.
KeenFlame@feddit.nu 2 hours ago
Nope. It’s in special tiny ways we author text. I think.
tal@lemmy.today 13 hours ago
Of course, another option is for people to dramatically curb their use of social media, or at a minimum, regularly delete posts after a set time threshold.
Deletion won’t deal with someone seriously-interested in harvesting stuff, because they can log it as it becomes available. And curbing use isn’t ideal.
I mentioned before the possibility of poisoning data, like, sporadically adding some incorrect information about oneself into one’s comments. Ideally something that doesn’t impact the meaning of the comments, but would cause a computer to associate one with someone else.
There are some other issues. My guess is that it’s probably possible to fingerprint someone to a substantial degree by the phrasing that they use. One mole in the counterintelligence portion of the FBI, Robert Hanssen, was found because on two occasions he used the unusual phrase “the purple-pissing Japanese”.
FBI investigators later made progress during an operation where they paid disaffected Russian intelligence officers to deliver information on moles. They paid $7 million to KGB agent Aleksander Shcherbakov[48] who had access to a file on “B”. While it did not contain Hanssen’s name, among the information was an audiotape of a July 21, 1986, conversation between “B” and KGB agent Aleksander Fefelov.[49] FBI agent Michael Waguespack recognized the voice in the tape, but could not remember who it was from. Rifling through the rest of the files, they found notes of the mole using a quote from George S. Patton’s speech to the Third Army about “the purple-pissing Japanese”.[50] FBI analyst Bob King remembered Hanssen using that same quote. Waguespack listened to the tape again and recognized the voice as Hanssen’s. With the mole finally identified, locations, dates, and cases were matched with Hanssen’s activities during the period. Two fingerprints collected from a trash bag in the file were analyzed and proved to be Hanssen’s.[51][52][53]
That might be defeated by passing text through something like an LLM to rewrite it. So, for example, to take a snippet of my above comment:
Respond with the following text rephrased sentence by sentence, concisely written as a British computer scientist might write it:
Deletion won’t deal with someone seriously-interested in harvesting stuff, because they can log it as it becomes available. And curbing use isn’t ideal.
I mentioned before the possibility of poisoning data, like, sporadically adding some incorrect information about oneself into one’s comments. Ideally something that doesn’t impact the meaning of the comments, but would cause a computer to associate one with someone else.
I get:
The deletion of data alone will not prevent a determined party from gathering information, as they may simply record the information as it becomes available prior to its deletion. Moreover, restricting usage is not an ideal solution to the problem at hand.
I previously mentioned the possibility of introducing deliberate errors or misinformation into one’s own data, such as periodically inserting inaccurate details about oneself within comments. The goal would be to include information that does not significantly alter the meaning of the comment, but which would cause automated systems to incorrectly associate that individual with another person.
That might work. One would have to check the comment to make sure that it doesn’t mangle the thing to the point that it is incorrect, but it might defeat profiling based on phrasing peculiarities of a given person, especially if many users used a similar “profile” for comment re-writing.
A second problem is that one’s interests are probably something of a fingerprint. It might be possible to use separate accounts related to separate interests — for example, instead of having one account, having an account per community or similar. That does undermine the ability to use reputation generated elsewhere (“Oh, user X has been providing helpful information for five years over in community X, so they’re likely to also be doing so in community Y”), which kind of degrades online communities, but it’s better than just dropping pseudonymity and going 4chan-style fully anonymous and completely losing reputation.
zerofk@lemmy.zip 3 hours ago
Your above average use of the word “one” and variations like “one’s” could be quite telling.
As could my correction of “it’s” in the above sentence.
Yliaster@lemmy.world 7 hours ago
Why is curbing use unideal?
KeenFlame@feddit.nu 2 hours ago
We like internet
HyperfocusSurfer@lemmy.dbzer0.com 12 hours ago
Regarding the last point: it’s more of a bias, tho, so it may even be a good thing. E.g. asking Kent Overstreet’s opinion on your bcachefs setup is probably useful, while getting relationship advice from him is ill-advised.
regenwetter@piefed.social 6 hours ago
Advice being right or wrong isn’t necessarily the big issue for online communities (unless most other users are also wrong). What really degrades them is users acting like assholes, and someone who acts like that in a tech community is fairly likely to also do that in a political or relationship community.
DarkCloud@lemmy.world 13 hours ago
Great, we’re at a point where “researchers” are helping tech bros hurt the public interest. Could they just NOT publish this shit? Stop giving helpful tips to tyrannical oligarchs!
Academics can be stupid idiots sometimes.
zerofk@lemmy.zip 3 hours ago
Researchers’ work has always been abused by others. The advancement and free distribution of knowledge should not be curtailed for fear of malicious parties.
maplesaga@lemmy.world 12 hours ago
Average people download gamed and apps and their phone is loaded to the tilt with bloatware. You think they care?
SupraMario@lemmy.world 10 hours ago
The average person puts their entire lives on Facebook or linkedin with their real names…they don’t give a shit.
EndlessNightmare@reddthat.com 2 hours ago
LinkedIn, if used properly, should just be professional/career related content. If you put anything overly personal or controversial, you are using it wrong.
I’m not saying that people don’t do that though.
ToTheGraveMyLove@sh.itjust.works 13 hours ago
Who am I? No forreal, WHO AM I? Last I remember I was on a cruise around the Caribbean. I blacked out one night while at the casino and when I came to I was on a beach in the middle of nowhere with a toothless man who spoke a language I couldn’t comprehend. Thankfully he still has a dial up connection somehow in the year of our lord 2026, but I’ve been on this island for two years now. SOMEONE COME GET ME!
FenrirIII@lemmy.world 9 hours ago
Your wife is much happier with me now and the children are already calling me dad. It’s time to move on.
merde@sh.itjust.works 13 hours ago
somebody should inform EU that they no longer need chatControl
:/
workgood@lemmy.dbzer0.com 11 hours ago
no it cant
corsicanguppy@lemmy.ca 6 hours ago
I kinda think I want it to try. I make little effort to hide my location or identity, and I think I’d kike to see the results.
…just without saying who I am before I get those results. And my desire to stay anonymous-ish and not give it a chance to cheat means I can’t satisfy I have the right to the identify of myself if it finds who I am.
Quite seriously, I cannot prove I have the right to make it search for me, for myself, without giving it too much information or without risking the leak of private info to a so-far unidentified stranger if it finds anything.
Catch-22
ne0phyte@feddit.org 2 hours ago
I am so grateful for already having been paranoid about sharing anything identifying about me starting 15+ years ago.
I never uploaded a picture of myself. Never used my real name anywhere. I used different nicks for different branches of the Internet. A plethora of different email addresses etc.
People thought I was being overly careful and I probably missed a lot of things due to not using Whatsapp, Facebook, Instagram, Twitter, Snapchat but I can’t say I regretted it at any point.