this post was submitted on 04 Dec 2023
888 points (97.9% liked)
Technology
59719 readers
2738 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's kind of odd that they could just take random information from the internet without asking and are now treating it like a trade secret.
This is why some of us have been ringing the alarm on these companies stealing data from users without consent. They know the data is valuable yet refuse to pay for the rights to use said data.
Yup. And instead, they make us pay them for it. 🤡
According to most sites TOS, when we write our posts we give them basically full access to do whatever they like including make derivative works. Here is the reddit one (not sure how Lemmy handles this):
2 points:
1 - I'm generally talking about companies extracting data from other websites, such as OpenAI scraping posts from reddit or other such postings. Companies that use their own collection of data are a very different thing.
2 - Terms of Service and Intellectual Property are not the same thing and a ToS is not guaranteed to be a fully legally binding document (the last part is the important part.) This is why services that have dealt with user created data that are used to licensing issues (think deviant art or other art hosting services) usually require the user to specify the license that they wish to distribute their content under (cc0, for example, would be fully permissible in this context.) This also means that most fan art is fair game as licensing that content is dubious at best, but raises the question around whether said content can be used to train an AI (again, intellectual property is generally different from a ToS).
It's no different from how Github's Copilot has to respect the license of your code regardless of whether you've agreed to the terms of service or not. Granted, this is legally disputable and I'm sure this will come up at some point with how these AI companies operate -- This is a brave new world. Having said that, services like Twitter might want to give second thought of claiming ownership over every post on their site as it essentially means they are liable for the content that they host. This is something they've wanted to avoid in the past because it gives them good coverage for user submitted content that they think is harmful.
If I was a company, I wouldn't want to be hinging my entire business on my terms of service being a legally binding document -- they generally aren't and can frequently be found to be unbinding. And, again, this is different from OpenAI as much of their data is based on data they've scraped from websites which they haven't agreed to take data from (finders-keepers is generally not how ownership works and is more akin to piracy. I wouldn't want to base a multinational business off of piracy.)
There was personal information included in the data. Did no one actually read the article?
Tbf it's behind a soft paywall
Well firstly the article is paywalled but secondly the example that they gave in this short bit you can read looks like contact information that you put at the end of an email.
That would still be personal information.
They do not have permission to pass it on. It might be an issue if they didn't stop it.
As if they had permission to take it in the first place
In a lot of cases, they don't have permission to not pass it along. Some of that training data was copyleft!
You don't want to let people manipulate your tools outside your expectations. It could be abused to produce content that is damaging to your brand, and in the case of GPT, damaging in general. I imagine OpenAI really doesn't want people figuring out how to weaponize the model for propaganda and/or deceit, or worse (I dunno, bomb instructions?)