this post was submitted on 29 Nov 2023
434 points (97.4% liked)
Technology
58303 readers
13 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
This is interesting in terms of copyright law. So far the lawsuits from Sarah Silverman and others haven't gone anywhere on the theory that the models do not contain a copies of books. Copyright law hinges on whether you have a right to make copies of a work. So the theory has been the models learned from the books but didn't retain exact copies, like how a human reads a book and learns it's contents but does not store an exact copy in their head. If the models "memorized" training data, including copyrighten works, OpenAI and others may have a problem (note the researchers said they did this same thing on other models).
For the silicone valley drama addicts, I find it curious that the researchers apparently didn't do this test on Bard of Anthropic's Claude, at least the article didn't mention them. Curious.
"Copyrighten" is an interesting grammatical construction that I've never seen before. I'd assume it would come from a second language speaker.
It looks like a mix of "written" and "righted".
"Copywritten" isn't a word I've ever heard, but it would be a past tense form of "copywriting", which is usually about writing text for advertisements. It's a pretty niche concept.
"Copyrighted" is the typical form for works that have copyright.
I'm not a grammar nazi - what's right & wrong is about what gets used which is why I talk about the "usual" form and not the "correct" form - but "copyrighted" is the clearest way to express that idea.
Copyrighten is just how they say it out in the country.
"I dun been copyrighten all damn day"
"Copyrightened" could mean explicit consent to use your material.
The paper suggests it was because of cost. The paper mainly focused on open models with public datasets as its basis, then attempted it on gpt3.5. They note that they didn't generate the full 1B tokens with 3.5 because it would have been too expensive. I assume they didn't test other proprietary models for the same reason. For Claude's cheapest model it would be over $5000, and bard api access isn't widely available yet.
So their angle should be plagiarism rather than copyright?