this post was submitted on 29 Sep 2023
436 points (93.4% liked)

Technology

59689 readers
3167 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

top 50 comments
sorted by: hot top controversial new old
[–] [email protected] 51 points 1 year ago* (last edited 1 year ago) (1 children)

Here's an idea, legally force companies like OpenAI to rely on opt-in data, rather then build their entire company on stealing massive amounts of data. That includes requiring to retrain from scratch. Sam Altman was crying for regulations for scary AI, right?

[–] [email protected] 17 points 1 year ago (2 children)

Would search engines only be allowed to show search results for sources that had opted in? They "train" their search engine on public data too, after all.

[–] [email protected] 17 points 1 year ago (14 children)

They aren't reselling their information, they're linking you to the source which then the website decides what to do with your traffic. Which they usually want your traffic, that's the point of a public site.

That's like trying to say it's bad to point to where a book store is so someone can buy from it. Whereas the LLM is stealing from that bookstore and selling it to you in a back alley.

load more comments (14 replies)
load more comments (1 replies)
[–] [email protected] 49 points 1 year ago (1 children)

This is no different than every other capitalist enterprise. The whole system works on taking a public resource, claiming private ownership of it, and then selling it back to the public for profit.

First it was farmland, then coal and minerals, oil, seafood, and now ideas. Its how the system works and is the whole reason people have been trying to stop it for the past 150 years.

The people making the laws are there because they and/or their parents and/or grandparents did the exact same thing. As despicable and corrupt as it is you won't change it by complaining and no-one is going to make a law to stop it.

[–] [email protected] 13 points 1 year ago

God damned right. Every "new" thing tends to be stolen. In more event history, its stolen from other capital, or from innovation with a free license, rather than artwork. Publishers might actually be able to make a problem out of this.

[–] [email protected] 28 points 1 year ago (8 children)

I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.

[–] adriaan 48 points 1 year ago (2 children)

That would be a much better comparison if it was artificial intelligence, but these are just reinforcement learning models. They do not get inspired.

[–] [email protected] 10 points 1 year ago (1 children)

More to the point: they replicate patterns of words.

[–] [email protected] 8 points 1 year ago (2 children)

just reinforcement learning models

...like the naturally occuring neural networks are.

[–] [email protected] 30 points 1 year ago (5 children)

The brain does not work the way you think… (I work in the field, bio-informatics). What you call “neural networks” come from an early misunderstanding of how the brain stores information. It’s a LOT more complicated and frankly, barely understood.

[–] [email protected] 12 points 1 year ago (1 children)

Yeah, accurately simulating a single pyramidal neuron requires an eight-layer deep neural network:

https://www.cell.com/neuron/pdf/S0896-6273(21)00501-8.pdf

load more comments (1 replies)
[–] [email protected] 5 points 1 year ago (2 children)

It’s a LOT more complicated and frankly, barely understood.

Yet you confidently state that the brain doesn't work the way LLMs do?

Obviously it doesn't work exactly the same way that LLMs do, if only because of the completely different substrates. But when you get to more nebulous concepts like "creativity" and "inspiration" it's not so clear.

load more comments (2 replies)
load more comments (3 replies)
[–] [email protected] 10 points 1 year ago (1 children)

Tell you what, you get a landmark legal decision classifying LLM as people and then we'll talk.

Until then it's software being fed content in a way not permitted by its license i.e. the makers of that software committing copyright infringement.

[–] [email protected] 8 points 1 year ago* (last edited 1 year ago) (1 children)

What exactly was not permitted by the license? Reading?

[–] [email protected] 13 points 1 year ago (9 children)

Using it to (create a tool to) create derivatives of the work on a massive scale.

[–] [email protected] 8 points 1 year ago* (last edited 1 year ago)

An AI model is not a derivative work. It does not contain the copyrighted expression, just information about the copyrighted expression.

[–] [email protected] 7 points 1 year ago

Wikipedia: In copyright law, a derivative work is an expressive creation that includes major copyrightable elements of a first, previously created original work.

I think you may be off a bit on what a derivative work is. I don't see LLMs spouting out major copyrightable elements of books. They can give a summary sure, but Cliff Notes would like to have a word if you think that's copyright infringement.

load more comments (7 replies)
[–] [email protected] 31 points 1 year ago* (last edited 1 year ago) (6 children)

That's an interesting take, I didn't know software could be inspired by other people's works. And here I thought software just did exactly as it's instructed to do. These are language models. They were given data to train those models. Did they pay for the data that they used to train for it, or did they scrub the internet and steal all these books along with everything everyone else has said?

[–] [email protected] 5 points 1 year ago (8 children)

Well, now you know; software can be inspired by other people's works. That's what AIs are instructed to do during their training phase.

load more comments (8 replies)
load more comments (5 replies)
[–] [email protected] 20 points 1 year ago (18 children)

These are machines, though, not human beings.

I guess I'd have to be an author to find out how I'd feel about it, to be fair.

[–] [email protected] 9 points 1 year ago (1 children)

Machines that aren't reproducing or distributing works

[–] [email protected] 5 points 1 year ago

If an AI "reproduces" a work it was trained on it is a failure of an AI. Why would anyone want to spend millions of dollars and devote oodles of computing power to build something that just does what a simple copy/paste operation can accomplish?

When an AI spits out something that's too close to one of the original training set that's called "overfitting" and it is considered an error to be corrected. Most overfitting that's been detected has been a result of duplication in the training set - when you hammer an AI image generator in training with thousands of copies of the Mona Lisa it eventually goes "alright, I get it already, when you say 'Mona Lisa' you want that exact pattern!" And will try its best to replicate that pattern when you ask it to later. That's why training sets need to be de-duplicated.

AIs are meant to produce new things.

load more comments (17 replies)
[–] [email protected] 19 points 1 year ago (4 children)

Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?

load more comments (4 replies)
load more comments (4 replies)
[–] [email protected] 25 points 1 year ago (8 children)

I hope they can at least get compensated.

[–] [email protected] 7 points 1 year ago (3 children)

So where can I check to see if my book was used? I published a book.

load more comments (3 replies)
load more comments (7 replies)
[–] [email protected] 9 points 1 year ago (1 children)

Ok so it’s been stealing art now it’s coming for authors. At what point do we hold the coalition who started this shit culpable for numerous accounts of plagiarism?

load more comments (1 replies)
[–] [email protected] 8 points 1 year ago (1 children)

Curious if the AI company actually bought those books or if they just came across them by pirating.

load more comments (1 replies)
[–] [email protected] 6 points 1 year ago

There's an idea by Barath Raghavan about an AI dividend that companies pay each netizen a share for the data they use to train these models.

I am into this idea if companies can't even do a simple opt-in mechanism.

[–] mindbleach 5 points 1 year ago

Of course books were used - and why not?

These are English-language models. They're trained on an English-language corpus. Ideally, the entire history of published works. They're in public, for public consumption. That's what "published" means. Nobody broke into an author's house to steal their secrets. The machine that writes is skimming the whole library to find out how to write gooder.

If they used a literal library - shelves full of dead trees - would people still insist the authors were never paid? Are we supposed to pretend a robot scanning a book a thousand times is fundamentally worse than a hundred people borrowing it? I guarantee you each human got more out of it. These stupid robots have to plow through an entire shelf just to figure out what "fantasy" means.

The more works we shovel into these things, the less each individual work counts. The model does not get bigger. Each work's unique impact shrinks as statistical patterns get generalized.

load more comments
view more: next ›