Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.
Any AI model that uses publically available information for training should be open source by law.
This business where corporations (that includes authors, who are published by huge corporations) fight over who “owns” ideas is assinine.
OpenAI, Google, all these fucks should be forced to open source their models, end of story.
Gibdos@feddit.de 1 year ago
I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.
adriaan@sh.itjust.works 1 year ago
That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.
Shurimal@kbin.social 1 year ago
...like the naturally occuring neural networks are.
Hackerman_uwu@lemmy.world 1 year ago
More to the point: they replicate patterns of words.
newthrowaway20@lemmy.world 1 year ago
That’s an interesting take, I didn’t know software could be inspired by other people’s works. Here. I thought software just did exactly as it’s instructed to do.
FaceDeer@kbin.social 1 year ago
Well, now you know; software can be inspired by other people's works. That's what AIs are instructed to do during their training phase.
PsychedSy@sh.itjust.works 1 year ago
They weren’t given data. They were shown data then the company spent tens of millions of dollars on cpu time to do statistical analysis of the data shown.
lloram239@feddit.de 1 year ago
AI isn’t software. Everything the AI knows is from the books. There is no human instructing the AI what to do. All the human does is build the scaffolding to let the AI learn, everything else is in the data.
Wander@kbin.social 1 year ago
Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?
Grimy@lemmy.world 1 year ago
Clearly not. He’s saying that other authors have done the same as the software does. The software creators implemented the same principle into their llm. You are being daft on purpose.
elbarto777@lemmy.world 1 year ago
These are machines, though, not human beings.
I guess I’d have to be an author to find out how I’d feel about it, to be fair.
Touching_Grass@lemmy.world 1 year ago
Machines that aren’t reproducing or distributing works
Shurimal@kbin.social 1 year ago
What's the difference? On the most fundamental level it's all the same.
kromem@lemmy.world 1 year ago
Did you write a comment on Reddit before 2015? If so, your copyrighted content was used without your permission to train today’s LLMs, so you absolutely get to feel one way or another about it.
The idea that these authors were somehow the backbone of the models when any individual contribution was like spitting in the ocean and model weights would have considered 100 pages of Twilight fan fiction equivalent to 100 pages from Twilight is honestly one of the negative impacts of the extensive coverage these suits are getting.
Pretty much everyone who has ever written anything indexed online is a tiny part of today’s LLMs.
dutchkimble@lemy.lol 1 year ago
But terminator said neural networks
sab@lemmy.world 1 year ago
I don’t think anyone is faulting the machines for this, just the people who instruct the machines to do it.
kromem@lemmy.world 1 year ago
Generally they probably bought the books they read though.
If George RR Martin torrented Tolkien, wouldn’t he be infringing on the copyright no matter how he subsequently incorporated it into future output?
I completely agree that the training as infringement argument is ludicrous.
But OpenAI exposed themselves to IP infringement by sailing the high seas in how they obtained the works in the first place.
I hate that the world we live in is one where so much data is gated behind paywalls, but the law is what it is, and if the government was going to come down hard on Aaron Swartz for trying to bypass paywalls for massive amounts of written text, it’s not exactly fair if there’s a double standard for OpenAI doing the same thing in an even more closed fashion.
But yes, the degree of entitled focus on the premise of training an AI as equivalent of infringing is weird as heck to see from authors drawing quite clearly from earlier works in their own output.
st0v@lemmy.zip 1 year ago
I have to assume that openAI also paid for the books. if yes then i consider it the same as me reciting passages from memory or coming up with derivative text.
if no, then by all means, go after them and any model trainer for the cost of one book.
Asking an LLM to recite an entire novel isn’t even vaguely a thing yet.
Omniraptor@lemm.ee 1 year ago
God that Aaron/jstor thing makes me see red every time. He was scraping jstor to publish it for the benefit of everyone, openai is doing it to make billions of dollars. Don’t forget who the bad guys are (and donate to sci-hub) everyone