Allegedly
Warl0k3@lemmy.world 11 hours ago
For clarity, it’s only being summarized for the users that wrote it, it’s not leaking them to everyone. A comedically inept bug to allow though, holy shit.
horn_e4_beaver@discuss.tchncs.de 6 hours ago
Warl0k3@lemmy.world 6 hours ago
In this case there’s no evidence showing that it’s being spread widely - the bug reports are entirely about users being shown their own content. If you have something to dispute that I’m all ears.
horn_e4_beaver@discuss.tchncs.de 6 hours ago
I was being a bit difficult tbh.
But it is absolutely true that we can’t know for sure that it isn’t being leaked elsewhere.
Reygle@lemmy.world 11 hours ago
AITA for understanding that as meaning in order to “summarize” the data the AI read it entirely and will never be instructed to “forget” that data
TRBoom@lemmy.zip 10 hours ago
Unless someone has released something new while I haven’t been paying attention, all the gen AIs are essentially frozen. Your use of them can’t impact the actual weights inside of the model.
If it seems like it’s remember things is because of the actual input of the LLM is larger than the input you will usually give it.
For instance lets say the max input for a particular LLM is 9096 tokens. The first part of that will be instructions from the owners of the LLM to prevent their model from being used for things they don’t like. Lets say the first 2000 tokens. That leaves 7k or so for a conversation that will be ‘remembered’.
Now if someone was really savvy, they’d have the model generate summaries of the conversation and stick them into another chunk of memory, maybe another 2000 tokens worth, that way it will seem to remember more than just the current thread. That would leave you with 5000 tokens to have a running conversation.
dgdft@lemmy.world 10 hours ago
Microsoft is almost certainly recording these summarization requests for QA and future training runs; that’s where the leakage would happen.
TRBoom@lemmy.zip 9 hours ago
100% agree. At this point I am assuming everything sent through their servers is actively being collected for LLM training.
SirHaxalot@nord.pub 8 hours ago
That is kind of assuming the worst case scenario though. You wouldn’t assume that QA can read every email you send through their mail servers ”just because ”
This article sounds a bit like engagement bait based on the idea that any use of LLMs is inherently a privacy violation. I don’t see how pushing the text through a specific class of software is worse than storing confidential data in the mailbox though.
That is assuming that they don’t leak data for training but the article doesn’t mention that.
fuckwit_mcbumcrumble@lemmy.dbzer0.com 9 hours ago
Why would that make you an asshole?
Reygle@lemmy.world 9 hours ago
I’ve noticed growing opposition to critical thoughts about the sick and twisted nature of ai and the people who are in the cult.
VeganCheesecake@lemmy.blahaj.zone 7 hours ago
LLMs are stateless. The model itself stays the same. Doesn’t mean they’re not saving the data elsewhere, but the LLM does not retain interactions.