this post was submitted on 29 Sep 2023
396 points (93.4% liked)

Technology

70916 readers
3152 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related news or articles.
  3. Be excellent to each other!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, this includes using AI responses and summaries. To ask if your bot can be added please contact a mod.
  9. Check for duplicates before posting, duplicates may be removed
  10. Accounts 7 days and younger will have their posts automatically removed.

Approved Bots


founded 2 years ago
MODERATORS
 

Authors using a new tool to search a list of 183,000 books used to train AI are furious to find their works on the list.

top 50 comments
sorted by: hot top controversial new old
[–] 0ddysseus@lemmy.world 47 points 2 years ago (1 children)

This is no different than every other capitalist enterprise. The whole system works on taking a public resource, claiming private ownership of it, and then selling it back to the public for profit.

First it was farmland, then coal and minerals, oil, seafood, and now ideas. Its how the system works and is the whole reason people have been trying to stop it for the past 150 years.

The people making the laws are there because they and/or their parents and/or grandparents did the exact same thing. As despicable and corrupt as it is you won't change it by complaining and no-one is going to make a law to stop it.

[–] Franzia@lemmy.blahaj.zone 11 points 2 years ago

God damned right. Every "new" thing tends to be stolen. In more event history, its stolen from other capital, or from innovation with a free license, rather than artwork. Publishers might actually be able to make a problem out of this.

[–] mojo@lemm.ee 46 points 2 years ago* (last edited 2 years ago) (1 children)

Here's an idea, legally force companies like OpenAI to rely on opt-in data, rather then build their entire company on stealing massive amounts of data. That includes requiring to retrain from scratch. Sam Altman was crying for regulations for scary AI, right?

[–] FaceDeer@kbin.social 15 points 2 years ago (2 children)

Would search engines only be allowed to show search results for sources that had opted in? They "train" their search engine on public data too, after all.

[–] mojo@lemm.ee 16 points 2 years ago

They aren't reselling their information, they're linking you to the source which then the website decides what to do with your traffic. Which they usually want your traffic, that's the point of a public site.

That's like trying to say it's bad to point to where a book store is so someone can buy from it. Whereas the LLM is stealing from that bookstore and selling it to you in a back alley.

load more comments (1 replies)
[–] pavnilschanda@lemmy.world 23 points 2 years ago (8 children)

I hope they can at least get compensated.

[–] Fredselfish@lemmy.world 6 points 2 years ago (3 children)

So where can I check to see if my book was used? I published a book.

load more comments (3 replies)
load more comments (7 replies)
[–] Gibdos@feddit.de 19 points 2 years ago (4 children)

I certainly hope that none of these authors have ever read a book before or have been inspired by something written by another author.

[–] newthrowaway20@lemmy.world 34 points 2 years ago* (last edited 2 years ago) (11 children)

That's an interesting take, I didn't know software could be inspired by other people's works. And here I thought software just did exactly as it's instructed to do. These are language models. They were given data to train those models. Did they pay for the data that they used to train for it, or did they scrub the internet and steal all these books along with everything everyone else has said?

load more comments (11 replies)
[–] elbarto777@lemmy.world 19 points 2 years ago (15 children)

These are machines, though, not human beings.

I guess I'd have to be an author to find out how I'd feel about it, to be fair.

[–] Touching_Grass@lemmy.world 8 points 2 years ago (1 children)

Machines that aren't reproducing or distributing works

load more comments (1 replies)
[–] kromem@lemmy.world 3 points 2 years ago (3 children)

Did you write a comment on Reddit before 2015? If so, your copyrighted content was used without your permission to train today's LLMs, so you absolutely get to feel one way or another about it.

The idea that these authors were somehow the backbone of the models when any individual contribution was like spitting in the ocean and model weights would have considered 100 pages of Twilight fan fiction equivalent to 100 pages from Twilight is honestly one of the negative impacts of the extensive coverage these suits are getting.

Pretty much everyone who has ever written anything indexed online is a tiny part of today's LLMs.

load more comments (3 replies)
load more comments (13 replies)
[–] Wander@kbin.social 18 points 2 years ago (4 children)

Are you saying the writers of these programs have read all these books, and were inspired by them so much they wrote millions of books? And all this software is doing is outputting the result of someone being inspired by other books?

load more comments (4 replies)
[–] kromem@lemmy.world 5 points 2 years ago (3 children)

Generally they probably bought the books they read though.

If George RR Martin torrented Tolkien, wouldn't he be infringing on the copyright no matter how he subsequently incorporated it into future output?

I completely agree that the training as infringement argument is ludicrous.

But OpenAI exposed themselves to IP infringement by sailing the high seas in how they obtained the works in the first place.

I hate that the world we live in is one where so much data is gated behind paywalls, but the law is what it is, and if the government was going to come down hard on Aaron Swartz for trying to bypass paywalls for massive amounts of written text, it's not exactly fair if there's a double standard for OpenAI doing the same thing in an even more closed fashion.

But yes, the degree of entitled focus on the premise of training an AI as equivalent of infringing is weird as heck to see from authors drawing quite clearly from earlier works in their own output.

load more comments (3 replies)
[–] Smoogs@lemmy.world 11 points 2 years ago (1 children)

Ok so it’s been stealing art now it’s coming for authors. At what point do we hold the coalition who started this shit culpable for numerous accounts of plagiarism?

[–] pazukaza@lemmy.ml 4 points 2 years ago

TIL "culpable" is an English word too. Culpable means guilty in Spanish and I thought you were a Spanish speaker doing spanglish. Now I know you're just a man of culture.

[–] leaky_shower_thought@feddit.nl 5 points 2 years ago

There's an idea by Barath Raghavan about an AI dividend that companies pay each netizen a share for the data they use to train these models.

I am into this idea if companies can't even do a simple opt-in mechanism.

[–] Pyr_Pressure@lemmy.ca 4 points 2 years ago (1 children)

Curious if the AI company actually bought those books or if they just came across them by pirating.

[–] threadloose@midwest.social 5 points 2 years ago

Oh, they're 100% pirated. Sorry this isn't open, but the preview should give you enough information. The database is available elsewhere, IIRC. https://www.theatlantic.com/technology/archive/2023/09/books3-database-generative-ai-training-copyright-infringement/675363/

load more comments
view more: next ›