Are graphics cards going to keep getting bigger forever?
Damn, the things used to be these thin little, well, cards. Nowadays they are reaching the size of entire consoles and can more accurately be called graphics bricks. Is the tech so stagnant that they won't be getting smaller again in the future?
The high end ones are so huge, power hungry, and fucking expensive that I'm starting to think they might as well just come with an integrated CPU and system RAM (in addition to the VRAM) on the same board.
What is the general industry expectation of what GPUs are going to be like in the mid term future, maybe 20 to 30 years from now? I expect if AI continues to grow in scope and ubiquity, then a previously unprecedented amount of effort and funding is going to be thrown at R&D for these PC components that were once primarily relegated to being toys for gamers.
The card in the picture is of a kind that no longer exists: the basic, office computer GPU.
It got entirely displaced by integrated graphics.
So in a way they did get smaller, so small that they share a piece of silicon with the CPU. The only cards that remain are those that are so power hungry they can't share power and cooling with the CPU.
I found a dusty old office desktop in my parents garage a few months ago from circa 2008, and I noticed it had both an AMD and Nvidia sticker on it, so I decided to open it up to take a look at what card it had. When I got the thing open I was perplexed to find there was no card in it at all, and that single core Sempron wasn't an APU, so what gives?
Found out that not only did integrated graphics exist back then on AMD platforms, but that they were built right into the motherboard and Nvidia manufactured them. It had no fan at all or really any obvious indication of its existence lol. Just an innocuous little piece of aluminum for a heat sink that was stuck on good to the chip with some kind of strong adhesive. Took a while to figure out that's what that was. Didn't know this was a thing, but that must've been a lot lower end than a 9500 GT.
Yes, it was the cheapest graphics card that could decode 1080p H.264 video in real time (and the acceleration worked in the Flash player). The 8500 GT could also do it but it was never popular. It made a huge difference when youtube became a thing.
this type of "expanding to fit the new constraints" is the central reason for human suffering, ecological damage, and everything else
The tech was mastered, so humans naturally started making bigger beefier resource hungry tech that consumed 3000% more resources to make their games look 30% better
If someone's not okay with playing their games on 2007-era TF2 graphics I don't consider them a real socialist. I'm not even saying such sacrifices would be necessary under socialism, but being unwilling to make that sacrifice is probably indicative that you won't be compatible with it
If someone's not okay with playing their games on 2007-era TF2 graphics I don't consider them a real socialist.
I’m going to go a step forward and say we need to ban personal electronics eventually, phones, personal computers, tvs, game consoles. This entertainment is contributing to the alienation of the working class. If we force people to go into more communal spaces to access this media, like arcades, movie theaters, computer cafes. It would be more environmentally friendly and as an added bonus we can censor all reactionary content.
Of course, party officials shouldn’t be forced to obey these fucking laws. Fuck the peasants, it’s going to take a long time for the transition to communism to happen so it’s in the interests of the terminally online left such as ourselves to collaborate and share the spoils of corruption with eachother, we gotta live by a do as I say not as I do motto.
i don't think its getting much bigger than 4090. even 4090 has sagging issues. sure, there are brackets and vertical mounting but vertical mounts on most cases have three slots.
there are ways to get more gpu power, just have more GPUs. obviously not suitable for gaming (microstutters and all the issues SLI/Crossfires had) but for AI itll work just fine and is exactly whats going on
PC components that were once primarily relegated to being toys for gamers.
i've to disagree, GPUs always had uses other than gaming like video rendering and shit. Gamers were just the most marketed to.
My guess is we're going to start seeing case and motherboard designs that privilege GPUs and relegate regular CPU function to a small system on a chip.
GPUs can get bigger if you build the system around them instead of plugging them in as an accessory.
In 30 years they'll find the remnants of computers and assume graphics cards were just funky portable heating units. Back then the planet still got cold after all, so people needed a way to stay warm.
I wouldn't rule that out actually. NVIDIA did buy Mellanox so they could integrate Infiniband with NVIDIA GPUs more directly so you can have fancier GPU clusters. At that point, having a powerful CPU and associated infrastructure is kind of pointless. I guess they'll end up looking like whatever SANs (storage area network, i.e. all the hard drives are separate from the compute servers) look like.
For consumer stuff, obviously idk. I think the giant GPUs are maybe because people are willing to buy them. The actual GPU itself provided by NVIDIA or AMD or Intel is actually a tiny little thing that looks like a CPU.
(GPU board without fans and plastic cover, you can see the GPU chip in the middle taking up a small part of the board)
And when they're on a laptop, it's just that actual GPU that's soldered onto the motherboard.
So GPUs themselves aren't actually getting that much bigger. I don't know if it's the memory, power, or the fans that are the cause of the massive GPU board sizes.
Higher power consumption requires more cooling, and the easiest way to do that is to add more surface area.
I almost wonder if custom-built desktop computers will eventually go away. I feel like it would be a lot more efficient to just have the CPU and GPU soldered onto the same motherboard the same way a laptop would. Then they could both share a common cooling and power infrastructure.
And maybe liquid cooling will become more popular.
Maybe some day there will be a massive breakthrough in computer part cooling. But refrigerants are already a thing I guess, and that's what liquid cooling is.
That's only if it's x86_64 and not ARM or RISC-V and only if NVIDIA is designing the CPU themselves which they wouldn't do. Desktop/industrial motherboards just have a socket that fits a certain CPU pinout.
A lot of it's heat dispersion, which goes back to how much electricity these things are using.
But the components are bigger, the first PC I assembled as a kid had 128 stream processors but the 40 series GPUs go up to 16384 cores. CUDA cores are a lot bigger too, despite die shrinks. More memory and associated components.
Buying an 8k movie will be a massive waste. Most are shot in 6k or less, and then mastered in as low as 2, especially if its has a lot of CG. If its shot on 35mm film, there is technically the latitude for an 8k master (or at least 6k, iirc), but you're not going to get much more detail due to the nature of it. Larger format film has even more latitude, but if its news when someone shoots on 70mm, its not enough to warrant an entire new format for.
the graphic card will grow to become the size of small capybara while the home computer will shrink down until it is no longer visible to the human eye. the size will pose a problem as a capybara is bigger than a spec of dust. such is the contradiction of technological development under capitalism
My understanding is that the primary reason for the size is heat limitations. The vast majority of the size on modern cards are heatsink radiator grills and fans.
All office computers need to have Wolfenstein 3D installed via Windows 3.1, so that modern dads can continue to engage with Take Your Child to Work Day, without having to do any actual parenting.
Lol, not sure if you're even aware but: vertical mounting has become a big trend in PC enthusiasts due to this factor. I have one of these on my 4090 rig:
I wasn’t, haven’t really kept up too much with current trends bc tbh I can’t afford anything so I’m just vibing with what I got, but I’m totally not surprised to see that, it felt pretty dicey at the time and it’s my understanding that the 40 series are even larger lol.
Probably a good chance. Take the Steam Deck and these console "slim" versions. They take the lessons they learned, refined some aspects but basically shunk the die down 1-2 nm. allowing more space for the same power or better along with better cooling.
We're at an interesting tech time where die processes are shrinking rapidly, 6nm, 5nm...soon 3nm. We're also seeing GaN making it's debut. Faster, more power efficent, smaller space, better thermals.
I saved up to get an OC 7900 XTX and AMD is on a roll with refreshing the cards with new tech. I can run a game at 1080 or 2k and upscale it to 4k with littlee performance hit. This is negating the necessities of massive cards making do more with less. Smarter not harder. They just dropped AI frame generation which works so well I have to tune it back due to not needing 350fps ultra settings and turning my card into 100% radiant space heater. That an theres some screen tearing at that point without vertical refresh. So if it works that well with their flagship card, I'd imagine it works fairly well on their lower end cards.
So it's getting a 1-2 punch from better smarter ways of doing things and smaller more efficent chips.
Yes. The same way cpu coolers got bigger and bigger until multi core cpus became the norm and cpus could just say “oh fuck it’s too hot, turn off 90% of the silicon”.
The crazy thing is that if you're in 3D graphics, VFX, or some video editing applications (Resolve I think especially leans on the GPU), a 4090 is an economically practical purchase. The initial investment is high, but so is the cost of render farms. And if you're a pro, render time is literally money. Can't use your PC if it's rendering. And in Blender at least, 4090s score much higher than even other RTX 4000 cards. As much as 2x higher. And if you're doing animation, you have to multiply the time it takes to render 1 frame by 24/30/60 frames per second. Now imagine a 24fps video that lasts ten minutes. 10x60x24=14,400 frames, so if you can shave off 1 second per frame, you've saved 4h of render time.
For gaming it's a waste. Who cares about a crappy AAA game having 10% more pixels to see ugly, "realistic" characters jankily mumbling their way through the worst dialogue committed to the page. Games should be games, and trying to force them to be stories ends up with the worst of both worlds
Generally speaking, I agree. The draw for those garden variety triple AAA games really should be the game play itself and not necessarily the graphics considering how many them are released every year. More brute force processing power to continuously increase the number of pixels with ray tracing far beyond practical limits is never going to make up for a lack of care, attention to detail, and artistry in the actual game assets.
On the other hand, I'm a sucker for simulation games, particularly flight sims, and I can't help but think it'd be sick as hell if I could have an RTX 5090 Ti and an R9 8950X3D to make that shit look indistinguishable from real life.
Also it's my dream for modders to make KSP look identical to MSFS, and honestly it's already halfway there. Unfortunately modding a game in that fashion, especially one that came out in the early 2010s, is always going to result in a horrific unoptimized mess. A nice realism overhaul suite of gameplay and graphics mods can easily push into requiring 64gb of RAM these days, and over 16gb of VRAM. But wow would it be worth it if I could ever afford such a system. For now I'll need to settle on the frankly depressing experience my laptop with an intel iGPU can provide.