Tech companies argued in comments on the website that the way their models ingested creative content was innovative and legal. The venture capital firm Andreessen Horowitz, which has several investments in A.I. start-ups, warned in its comments that any slowdown for A.I. companies in consuming content “would upset at least a decade’s worth of investment-backed expectations that were premised on the current understanding of the scope of copyright protection in this country.”
underneath the screenshot is the "Oh no! Anyway" meme, featuring two pictures of Jeremy Clarkson saying "Oh no!" and "Anyway"
We do need copyright reform, but also fuck "AI." I couldn't care less about them infringing on proprietary works, but they're also infringing on copyleft works and for that they deserve to be shut the fuck down.
Either that, or all the output of their "AI" needs to be copyleft.
Not just the output. One could construct that training your model on GPL content which would have it create GPL content means that the model itself is now also GPL.
It's why my company calls GPL parasitic, use it once and it's everywhere.
This is something I consider to be one of the main benefits of this license.
If this is what it takes to get copyright reform, just granting tech companies unlimited power to hoover up whatever they want and put it in their models, it's not going to be the egalitarian sort of copyright reform that we need. Instead, we will just getting a carve out just for this, which is ridiculous.
There are small creators who do need at least some sort of copyright control, because ultimately people should be paid for the work they do. Artists who work on commission are the people in the direct firing line of generative AI, both in commissions and in their day jobs. This will harm them more than any particular company. I don't think models will suffer if they can only include works in the public domain, if the public domain starts in 2003, but that's not the kind of copyright protection that Amazon, Google, Facebook, etc. want, and that's not what they're going to ask for.
Tech illiterate guy here. All these Ml models require training data, right? So all these AI companies that develop new ML based chat/video/image apps require data. So where exactly do they? It can't be that their entire dataset is licensed, isn't it?
If so, are there any firms that are using these orgs for data theft? How to know if the model has been trained on your data? Sorry if this is not the right place to ask.
I saw that! Brick killed a guy! Did you throw a trident?
Yeah, there were horses, and a man on fire, and I killed a guy with a trident!
Brick, I've been meaning to talk to you about that. You should find yourself a safehouse or a relative close by. Lay low for a while, because you're probably wanted for murder.
I fear this is a giant power grab. What this will lead to is that IP holders, those that own the content that AI needs to train will dictate prices. So all the social media content you kindly gave reddit, facebook, twitter, pictures, all that stuff means you won't be able to have any free AI software.
No free / open source AI software means there is a massive power imbalance because now only those who can afford to buy this training data and do it, any they are forced to maximize profits (and naturally inclined anyway).
Basically they will own the "means of generation" while we won't.
Current large model would all be sued to death, no license with IP owner yet, would kill all existing commercial large models. Except all IP owner are named and license granted retroactive, but sound unlikely.
Hundred of IP owner company and billion of individual IP owner setting prices will probably behave like streaming: price increase and endless fragmentation. Need a license for every IP owner, paperwork will be extremely massive. License might change, expire, same problem as streaming but every time license expire need to retrain entire model (or you infringe because model keep using data).
And in the EU you have right to be forgotten, so excluded from models (because in this case not transformative enough, ianal but sound like it count as storing), so every time someone want to be excluded, retrain entire model.
Do not see where it possible to create large model like this with any amount of money, time, electricity. Maybe some smaller models. Maybe just more specific for one task.
Also piracy exists, do not care about copyright, will just train and maybe even open source (torrent). Might get caught, might not, might become dark market, idk. Will exist though, like deepfakes.
I mean, I won't deny that small bit of skill it took to construct a plausible sounding explanation for why the public should support your investment, because it's "not illegal (yet)".
They have chosen to think that if it runs through AI, it is then a derivative, it is not. If I put Disney and Amazon together as a prompt, things come out very similar to their logos and it's obviously a copyright infringement. The worst part of this, they'll still steal all of the small artists and avoid the larger ones.
Don't forget that they will then switch sides and try to copyright "their work", preventing others from even thinking about their work without paying the toll.
Nah. Even in its current stupid state, copyright has to recognize that sifting through the entire internet to get a gigabyte of linear algebra is pretty goddamn transformative.
No kidding the machine that speaks English read every book in the library. Fuck else was it gonna do?
You know, I'd be OK with all the AI tosspots breaking copyright if I didn't risk getting a nastygram every time I download Jason Statham Beats People Up For 90 Minutes VII without a VPN.
Either we can all do what we want online or no cunt can.
A scammer made unreasonable promises to investors and is now warning everyone that their victims/investors are going to lose money when the process of making fair laws takes the typical amount of time that it always takes.