Is that just a wild assumption, or...? One phenomena that has already been witnessed with AI is that it does in fact get worse if it trains upon it's own output.
Given that I have locally-run AIs sitting on my home computer that I have no plan to delete (until something better comes along), then yeah, it's never going to get worse. If all else fails I can just use the existing AI for as long as I want. It doesn't "wear out."
The physical components will, and compatible components for older systems keep getting harder to come across. Computers are not immortal entities. Maintenance of older machines will continually become more labour and cost intensive over time.
The models are digital, making copies for safekeeping is easy.
The hardware is a computer, and computers are general-purpose. The kind that run AI models well at infrastructure scale are rather high end, but are still available off-the-shelf.
Computers are general-purpose machines. You can run a computer program on any computer, it may just be faster or slower depending on the computer's capabilities.
The AIs I run locally are also open-source, so if future computers lose compatibility with existing programs they can be recompiled for the new architecture.
I suppose we could lose the ability to build computers entirely, but that strikes me as a much bigger and more general issue than just this AI thing.
Incorrect. Certain programs require certain standards for how the hardware is designed. There are already lots of old programs which can't be run natively on modern machines, and using software to emulate a compatible environment can impact performance in more ways than just speed.
You're wildly wrong about the fundamentals of computer science here. I'd be starting from first principles trying to explain further. I recommend reading up on Turing machines, or perhaps getting ChatGPT to explain it to you.
I actually happen to know a lot about computers, and can even build them from raw materials. They are not eternal existences no matter how they're treated in popular culture, and data retention as hardware/standards evolve is actually a serious concern that is getting a fair bit of attention in research. One of the more interesting avenues being explored is encoding data in DNA, because humans will always have a reason to want to be capable of reading DNA, but that's still just theory at this point.
You’re rooting for a revolutionary new technology to fail rather than get better
As long as the oligarchs who run and own these AI systems are at the helm, yes I'm rooting for it to fail. Better is in the eyes of the beholder. Because come on, we all know better is going to be defined as better for the oligarchs, not you or me.
There's nothing 'revolutionary' about a mass theft machine until EVERYONE IT'S STEALING FROM is getting paid out of the thieves' pockets for what was stolen from them; and the people that run it make no profit from it. Til then, it's just business as usual out of the west's necrocapitalists; and your business makes me vomit.