this post was submitted on 31 Aug 2023
11 points (100.0% liked)
PC Gaming
4 readers
1 users here now
Discuss Games, Hardware and News on PC Gaming **Discord** https://discord.gg/4bxJgkY **Mastodon** https://cupoftea.social **Donate** https://ko-fi.com/cupofteasocial **Wiki** https://www.pcgamingwiki.com
founded 1 year ago
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
It's pretty sizeable if you're running it 24/7 without checks.
A recently common example might be someone interested in running Stable Diffusion locally. Run the program overnight that's drawing ~300-400 watts for 12 hours. For a comparison, an electric heater can run up to 1,200 watts and those are known to absolutely raise the electric bill gone unchecked (like putting it on a timer vs. running over night).
For gaming 20 hours a week? Probably not too much. For gaming 20 hours a week and running AI a few times a week overnight (40 hours) it's noticeable for sure.
However there's also the factor of ambient heating, so there's technically some offset cost if you need heating... And it's also not something to ignore, I had a room consistently 68F with bad insulation up to 85F with a window open, hotter in the summers of course.
Overall, yes but also no. Like with most things it's really about the use case and consistency. NVIDIA GPU in a media server? Higher energy costs than something like an intel quicksync for limited realistic gains but somewhat noticeable cost increase. Gaming GPU running high idle all the time just browsing and watching videos? Definitely more expensive than just a laptop, but with proper checks like the gaming computer put in eco mode they're more equatable.
Righto, thanks for the detailed reply.
The CEO might not be far wrong in that case, the average user probably doesn't run their GPU long enough to notice efficiency gains. And given their preferred market are the ones with money to burn, it makes sense they'd target improved performance over efficiency.
I'm inclined to agree as well, although I do think energy efficiency is environmentally important and the solution shouldn't be to throw more power at the hardware. For that reason I do appreciate some middleground between the two.
Realistically, my friends 7900XTX compared to my 3080 are within the same power consumption under load but he has 24GB of VRAM where I do not. To get that there with NVIDIA needs an extra 150 watts for the 4090 or 3090. Regardless of performance elsewhere, that's pretty sizeable, so it would be a shame to potentially lose that in place of something like a 30GB VRAM card pushing 450 watts from AMD.
Most people aren't running their computer 24/7 doing crazy tasks tho. I almost never turn my PC off but if I'm not using it it's generally in sleep mode, some times I just leave it on and running like if I know I have to go out for a few hours but will be right back on the PC where I left off when I return home, but generally it's asleep and I can't imagine it's using much electricity at that point.
It's all relative. To add some more context with your description,
As mentioned under load my PC (5800x3D + 3080 10gb) draws between 350 and 575 watts (depends on if I have monitors plugged into my UPS and the GPU power draw, some programs draw more than others).
Idling my PC draws about 175 watts.
In sleep my PC draws about 68-80 watts.
Like the NVIDIA GPU server example, even though it's not a lot of power, in comparison to more efficient computers doing the same task it's exorbitant.
You're right that most people with a GPU won't even be running something under load for a few hours - if they don't do rendering and they don't use AI then gaming is the only thing left that can really put a GPU to use.
So then it becomes about efficiency deciding how to optimize those tasks. If AMD can push out performance relative to NVIDIA but for 100 watts less, that's the difference between a PC in idle and a PC in sleep. That's pretty sizeable to ignore, even if you just leave the PC on 24/7 as a gaming PC+ready to use web browser. Similarly, if I'm deciding to put my GPU to use at all, it seems reasonable to consider long term cost efficiency. It's weird to think about since we don't push it much, but 20 hours a week gaming even 5 years ago vs. today is a huge power difference. Just look at the 1080Ti, a beast back then and still holds up today. Draws only 300 watts under load, and the 980ti can get 250 watts.
In terms of performance, 450w for even the 3090 let alone the 4090 absolutely blows these out of the water, but in terms of long term idle they are also, relatively, much more expensive.
All in all, most people aren't putting their PC under load 24/7 but most people also aren't only turning it on as needed. While it's true that they're not consistently drawing 300+ watts all the time, they are still likely idling (on just not being used actively) at higher levels than previous generations. My idle is quite close to the 980ti under load which is pretty insane.
Yeah, it also matters the specific setups and cards. There's still some issues with idle power on newer AMD cards in multi-monitor configs where they draw >50w doing nothing and some configs draw 100+w idling. It's been an issue since the 7k cards released I believe. They've had a few updates which has helped for some people according to various reddit threads, but not everyone. I think the Nvidia ones by comparison only pull like 8w-20w idling.
This isn't major for most users utilizing sleep most of the day, but it's also add up over time.