Tell you what, you get a landmark legal decision classifying LLM as people and then we’ll talk.
Until then it’s software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.
Comment on Authors Are Furious After Finding Their Works on List of Books Used To Train AI
Shurimal@kbin.social 1 year agojust reinforcement learning models
...like the naturally occuring neural networks are.
Tell you what, you get a landmark legal decision classifying LLM as people and then we’ll talk.
Until then it’s software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.
What exactly was not permitted by the license
Using it to (create a tool to) create derivatives of the work on a massive scale.
An AI model is not a derivative work. It does not contain a copyrighted expression, just information about the copyrighted expression.
Wikipedia: In copyright law, a derivative work is an expressive creation that includes major copyrightable elements of a first, previously created original work.
I think you may be off a bit on what a derivative work is. I don’t see LLMs spouting out major copyrightable elements of books. They can give a summary sure, but Cliff Notes would like to have a word if you think that’s copyright infringement.
Better tell that Google and their search index, book scanning project and knowledge graph.
Well when that happens we have laws. So no problems
Khalic@kbin.social 1 year ago
The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.
canihasaccount@lemmy.world 1 year ago
Yeah, accurately simulating a single pyramidal neuron requires an eight-layer deep neural network:
www.cell.com/neuron/…/S0896-6273(21)00501-8.pdf
demonsword@lemmy.world 1 year ago
that was an interesting read, thank you
FaceDeer@kbin.social 1 year ago
Yet you confidently state that the brain doesn't work the way LLMs do?
Obviously it doesn't work exactly the same way that LLMs do, if only because of the completely different substrates. But when you get to more nebulous concepts like "creativity" and "inspiration" it's not so clear.
lloram239@feddit.de 1 year ago
The part where brain and neural net differ is in the learning via backpropagation, that seem to be done different in the brain, as there is no mechanism to go backwards through the network and jiggle the weights.
That aside, they seem to work very similar once they are trained, as the knowledge they are able to extract from data ends up being basically the same that a human would be able to extract. There is surprisingly little weirdness in AI and a surprising amount of human-like capabilities.
originalucifer@moist.catsweat.com 1 year ago
people have a definite fear of being defined as machines... not sure why we think were so special..
originalucifer@moist.catsweat.com 1 year ago
so its barely understood, but this definitely is not it. got it.
FontMasterFlex@lemmy.world 1 year ago
But you, random stranger on the internet, knows better than the guy that literally works in the field. Got it.
originalucifer@moist.catsweat.com 1 year ago
i do? where did i claim that?