If you trained a model on a single copyrighted work, then that would be a copyright violation because it would inevitably produce output similar to that single work.
But if you train it on hundreds of thousands of copyrighted works, that’s no longer a copyright violation, because output won’t closely match any single work.
How is something a crime if you do it once, but not if you do it a million times?
A basic fundamental of copyright law and fair use is if the result is transformative. People literally do stuff like make collages with copyright works and it's fine in many cases.
Turning pictures into an AI model (and that's being really generous in my phrasing as if the pictures have anything to do with the math) is just about one of the most transformative things you can do with a picture.
This is like copyright 101 and if you're shocked you don't understand what you're talking about in regards to copyright.
to add to this, no art exists in a bubble - someone that's educated enough on the relevant subject matter could take any piece of work, and break it down into the individual influences that coalesced into that work; from the content itself, to the techniques used
Except it's not really transformative because the end product is not the model itself. The product is a service that writes code or draws pictures. It is literally the exact same as the input and it is intended specifically to avoid having to buy the inputs.
Training the AI isn’t a copyright violation though. Producing content from a single source of training information is intuitively different from producing content from a litany of sources. Is there a distinction I’m not understanding that you are pointing out?
"AI" models are, essentially, solvers for mathematical system that we, humans, cannot describe and create solvers for ourselves.
For example, a calculator for pure numbers is a pretty simple device all the logic of which can be designed by a human directly. A language, thought? Or an image classifier? That is not possible to create by hand.
With "AI" instead of designing all the logic manually, we create a system which can end up in a number of finite, yet still near infinite states, each of which defines behavior different from the other. By slowly tuning the model using existing data and checking its performance we (ideally) end up with a solver for some incredibly complex system.
If we were to try to make a regular calculator that way and all we were giving the model was "2+2=4" it would memorize the equation without understanding it. That's called "overfitting" and that's something people being AI are trying their best to prevent from happening. It happens if the training data contains too many repeats of the same thing.
However, if there is no repetition in the training set, the model is forced to actually learn the patterns in the data, instead of data itself.
Essentially: if you're training a model on single copyrighted work, you're making a copy of that work via overfitting. If you're using terabytes of diverse data, overfitting is minimized. Instead, the resulting model has actual understanding of the system you're training it on.
How is something a crime if you do it once, but not if you do it a million times?
Because doing it a million times seriously dilutes the harm to any single content creator (assuming those million sources are from many, many different content creators, of course). Potential harm plays a major role in how copyright cases are determined, and in cases involving such a huge amount of sources, harm can be immeasurably small.
In addition to right and wrong, the practicality of regulation and enforcement is often a part of groundbreaking decisions like these, and I’m not certain this particular issue is something our legal system is equipped to handle.
I’m not sure I agree with the reasoning here, but I see their thinking.
An AI trained on a single image would also probably be fine if it was somehow a generalist AI that didn't overfit on that single image. The quantity really doesn't matter.
Imagine this situation if a human replaced the AI.
Imagine a human who wants to write a book. They've read hundreds of other books already, and lots of other things besides books. Then they write a book. The final work probably contains an amalgamation of all the other things they've read--similar characters, themes, plot points, etc.--but it's a unique combination, so it's distinct from those other works. No copyright violation.
Now imagine that same human has only ever read one book. Over and over. They know only the one book. The human wants to write a new book. But they only have experience with the one they've read again and again. So the book they write is almost exactly the same as the one book they read. That's a copyright violation.
Training an AI model is not a crime, any more than reading a book is a crime. You're not making "copies" or profiting directly from that single work.
How is something a crime if you do it once, but not if you do it a million times?
Companies get to steal from people all the time without repercussions through erroneous fees, 'mistakes' in billing, denying coverage, and even outright fraud only gets a slap on the wrist fine at best. But an average person steals $5 and they are thrown in jail.
How is something a crime if you do it once, but not if you do it a million times?
You can dream up other examples of this.
If you're a DJ performing for a large audience and yell "I want to see you shake it for me!", that is legal.
If you walk up to one specific woman on the street and pull her aside and say "I want to see you shake it for me", that's sexual harassment.
If the government announces that the median income of Detroit residents has gone up by 3%, that's normal.
If the government public announces that John Fuckface, 36.2 years old, living at 123 Fake Street, had his income increase by 5% in the previous year, that's a privacy violation.
The whole point of training the AI is to build a model that can't reproduce a single work.
It may seem superficially strange, but the more works you include, the less capable it is of reproducing one work.
How is something a crime if you do it once, but not if you do it a million times?
Because we are talking about a generalized knowledge base vs a specific one? Is it not obvious from the explanation you quoted that instructing an AI to respond off of millions of sources means that it isn't biased off of one person's work?
this will be the end of the open internet. Expect login walls and subscriptions everywhere.
Rising interest rates are doing that, not AI.
The open Internet is based on a fundamental principal that people like you forget over and over.
Information should be free and plentiful, and making it free and plentiful benefits the common person. Data and scraping are essential parts of that common good.
The Internet will survive. The one you think exists - where you get to mooch and demand payment - never existed.
Copyright is, at its heart, about the right to make copies. If no direct connection can be made to another work then it is clearly not a copy and therefore...
Your fears don't seem plausible, either. A person or company doing AI training only needs 1 single copy. It's hard to see how that would translate to more than a few extra copies sold; at best, maybe a few dozen or a few hundred in the long run. I can see how going to court over a single copy of each item in their catalog is worth it for the larger corporations but what you fear just doesn't make financial sense to me.
You as a human being learn by stealing lessons and ideas from literally EVERYTHING you come in contact with. What makes something plagiarism over inspiration is numbers, so yes, HUNDREDS or THOUSANDS of sources stops something from being plagiarism or theft and instead makes it transformative.