this post was submitted on 09 Jul 2023
500 points (97.0% liked)

Technology

59197 readers
3191 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

Two authors sued OpenAI, accusing the company of violating copyright law. They say OpenAI used their work to train ChatGPT without their consent.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 48 points 1 year ago (4 children)

Or is it that the OpenAI computer isn’t going a good enough job of following the “three references” rule to avoid plagiarism?

This is exactly the problem, months ago I read that AI could have free access to all public source codes on GitHub without respecting their licenses.

So many developers have decided to abandon GitHub for other alternatives not realizing that in the end AI training can safely access their public repos on other platforms as well.

What should be done is to regulate this training, which however is not convenient for companies because the more data the AI ingests, the more its knowledge expands and "helps" the people who ask for information.

[–] [email protected] 41 points 1 year ago

It's incredibly convenient for companies.

Big companies like open AI can easily afford to download big data sets from companies like Reddit and deviantArt who already have the permission to freely use whatever work you upload to their website.

Individual creators do not have that ability and the act of doing this regulation will only force AI into the domain of these big companies even more than it already is.

Regulation would be a hideously bad idea that would lock these powerful tools behind the shitty web APIs that nobody has control over but the company in question.

Imagine the world is the future, magical new age technology, and Facebook owns all of it.

Do not allow that to happen.

[–] [email protected] 17 points 1 year ago (1 children)

Is it practically feasible to regulate the training? Is it even necessary? Perhaps it would be better to regulate the output instead.

It will be hard to know that any particular GET request is ultimately used to train an AI or to train a human. It's currently easy to see if a particular output is plagiarized. https://plagiarismdetector.net/ It's also much easier to enforce. We don't need to care if or how any particular model plagiarized work. We can just check if plagiarized work was produced.

That could be implemented directly in the software, so it didn't even output plagiarized material. The legal framework around it is also clear and fairly established. Instead of creating regulations around training we can use the existing regulations around the human who tries to disseminate copyrighted work.

That's also consistent with how we enforce copyright in humans. There's no law against looking at other people's work and memorizing entire sections. It's also generally legal to reproduce other people's work (eg for backups). It only potentially becomes illegal if someone distributes it and it's only plagiarism if they claim it as their own.

[–] [email protected] 2 points 1 year ago (1 children)

This makes perfect sense. Why aren’t they going about it this way then?

My best guess is that maybe they just see openAI being very successful and wanting a piece of that pie? Cause if someone produces something via chatGPT (let’s say for a book) and uses it, what are they chances they made any significant amount of money that you can sue for?

[–] [email protected] 1 points 1 year ago

It's hard to guess what the internal motivation is for these particular people.

Right now it's hard to know who is disseminating AI-generated material. Some people are explicit when they post it but others aren't. The AI companies are easily identified and there's at least the perception that regulating them can solve the problem, of copyright infringement at the source. I doubt that's true. More and more actors are able to train AI models and some of them aren't even under US jurisdiction.

I predict that we'll eventually have people vying to get their work used as training data. Think about what that means. If you write something and an AI is trained on it, the AI considers it "true". Going forward when people send prompts to that model it will return a response based on what it considers "true". Clever people can and will use that to influence public opinion. Consider how effective it's been to manipulate public thought with existing information technologies. Now imagine large segments of the population relying on AIs as trusted advisors for their daily lives and how effective it would be to influence the training of those AIs.

[–] [email protected] 7 points 1 year ago (1 children)

Plus, any regulation to limit this now means that anyone not already in the game will never breakthrough. It's going to be the domain of the current players for years, if not decades. So, not sure what's better, the current wild west where everyone can make something, or it being exclusive to the already big players and them closing the door behind

[–] [email protected] 4 points 1 year ago

My concern here is that OpenAI didn't have to share gpt with the world. These lawsuits are going to discourage companies from doing that in the future, which means well funded companies will just keep it under wraps. Once one of them eventually figures out AGI, they'll just use it internally until they dominate everything. Suddenly, Mark Zuckerberg is supreme leader and we all have to pledge allegiance to Facebook.

[–] [email protected] 2 points 1 year ago

AI could have free access to all public source codes on GitHub without respecting their licenses.

IANAL, but aren't their licenses are being respected up until they are put into a codebase? At least insomuch as Google is allowed to display code snippets in the preview when you look up a file in a GitHub repo, or you are allowed to copy a snippet to a StackOverflow discussion or ticket comment.

I do agree regulation is a very good idea, in more ways than just citation given the potential economic impacts that we seem clearly unprepared for.