Politicians, they try to crack as fewer eggs as possible, telling us they are our friends, and we believe them. Now then.. some do more good than bad, some do more bad than good. But on the other hand something that is _good for me_ is _bad for you_ and vice versa. Politicians are just the means to move the needle juuuuuuust a little bit, so show a change, but never make a drastic one. The cost of drastic changes is re-election. And this is the bread and butter of politicians (yes, I am over-over-simplifying but this is human history and a lot will be left out in a comment).
If a savant has perfect recall, remembers text perfectly and rearranges that text to create a marginally new text, he'd be sued for breach of copyright.
Only large corporations get away with it.
> If an AI reads 100 scientific papers and churns out a new one, it is plagiarism.
That is a specific claim that is being directly addressed and pretty clearly qualifies as "good faith".
"To steal ideas from one person is plagiarism; to steal from many is research."
Any suits would be based on the degree the marginally new copy was fair use. You wouldn't be able to sue the savant for reading and remembering the text.
Using AI to creat marginally new copies of copyrighted work is ALREADY a violation. We don't need a dramatic expansion of copyright law that says that just giving the savant the book to real is a copyright violation.
Plagarism and copyright are two entirely different things. Plagarism is about citations and intellectual integrity. Copyright is a about protecting economic interests, has nothing to to with intellectual integrity, and isn't resolved by citing the original work. In fact most of the contexts where you would be accused of plagarism, would be places like reporting, criticism, education or research goals make fair use arguments much easier.
https://www.copyright.gov/ai/Copyright-and-Artificial-Intell...
And for the future, here's one heuristic: if there is a profound violation of the law anywhere that (relatively speaking) is ignored or severely downplayed, it is likely that interested parties have arrived at an understanding. Or in other words, a conspiracy.
[1] There are tons of legal arguments on both sides, but for me it is enough to ask: if this is not illegal and is totally fair use (maybe even because, oh no look at what China's doing, etc.), why did they have to resort to & foster piracy in order to obtain this?
European here, but why do you think this is so clear cut? There are other jurisdictions where training on copyrighted data has already been allowed by law/caselaw (Germany and Japan). Why do you need a conspiracy in the US?
AFAICT the US copyright law deals with direct reproductions of a copyrighted piece of content (and also carves out some leeway with direct reproduction, like fair use). I think we can all agree by now that LLMs don't fully reproduce "letter perfect" content, right? What then is the "spirit" of the law that you think was broken here? Isn't this the definition of "transformative work"?
Of note is also the other big case involving books - the one where google was allowed to process mountains of books, they were sued and allowed to continue. How is scanning & indexing tons of books different than scanning & "training" an LLM?
Contrast that with AI companies:
They don't necessarily want to assert fair use, the results aren't necessarily publicly accessible, the work used isn't cited, users aren't directed to typical sales channels, and many common usages do meaningfully reduce the market for the original content (e.g. AI summaries for paywalled pages).
It's not obvious to me as a non-lawyer that these situations are analogous, even if there's some superficial similarity.
https://drewdevault.com/2020/08/24/Alice-in-Wonderland.html
https://drewdevault.com/2021/12/23/Sustainable-creativity-po...
If an artist produces a work they should have the rights to that work. If I self-publish a novel and then penguin decides that novel is really good and they want to publish it, without copyright they'd just do that, totally swamping me with their clout and punishing my ever putting the work out. That's a bad thing.
Assuming you agree with the idea of inheritance, which is another topic, then it is unfair to deny inheritance of intellectual property. For example if your father has built a house, it will be yours when he dies, it won't become a public house. So why would a book your father wrote just before he died become public domain the moment he dies. It is unfair to those doing who are doing intellectual work, especially older people.
If you want short copyright, is would make more sense to make it 20 years, human or corporate, like patents.
* https://www.copyright.gov/ai/Copyright-and-Artificial-Intell...
Yes please.
Delete it for everyone, not just these ridiculous autocrats. It's only helping them in the first place!
andy99•1h ago
kklisura•1h ago
They acknowledge the issue is before courts:
> These issues are the subject of intense debate. Dozens of lawsuits are pending in the United States, focusing on the application of copyright’s fair use doctrine. Legislators around the world have proposed or enacted laws regarding the use of copyrighted works in AI training, whether to remove barriers or impose restrictions
Why did they write the finding: I assume it's because it's their responsibility:
> Pursuant to the Register of Copyrights’ statutory responsibility to “[c]onduct studies” and “[a]dvise Congress on national and international issues relating to copyright,”...
All excerpts are from https://www.copyright.gov/ai/Copyright-and-Artificial-Intell...
_heimdall•1h ago
Sure the courts may find its out of their jurisdiction, but they should act as they see fit and let the courts settle that later.
bgwalter•1h ago
Why could a copyright office not advise the congress/senate to enact a law that forbids copyrighted material to be used in AI training? This is literally the politicians' job.
9283409232•18m ago