this post was submitted on 20 Sep 2023
42 points (92.0% liked)

Games

32617 readers
2220 users here now

Welcome to the largest gaming community on Lemmy! Discussion for all kinds of games. Video games, tabletop games, card games etc.

Weekly Threads:

What Are You Playing?

The Weekly Discussion Topic

Rules:

  1. Submissions have to be related to games

  2. No bigotry or harassment, be civil

  3. No excessive self-promotion

  4. Stay on-topic; no memes, funny videos, giveaways, reposts, or low-effort posts

  5. Mark Spoilers and NSFW

  6. No linking to piracy

More information about the community rules can be found here.

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 7 points 1 year ago (3 children)

This only got bad with the most recent generation of CPUs. AMD 5xxx series is very efficient as demonstrated by Gamers Nexus. The Intel CPUs from 2500k to idk, 8xxx series? were efficient until they started slapping more cores and then cranking the power on them.

[–] [email protected] 2 points 1 year ago (1 children)

Yes the second thing about cranking power and cores is what I'm talking about.

Also, as far as gpus, the 2000 series was ridiculously power hungry at the time, and it looks downright reasonable now. It's like the Overton window of power consumption lol.

Image

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago) (1 children)

I dunno, I ran a 2080 on the same PSU that I used on a 2013 build, a 650W seasonic. Got some graphs? Power consumption didn’t seem to jump that bad until the latest gen.

My current 3090 is a power hog though, that’s when I’d say it started for Nvidia (3000 series). For AMD, 7000 series CPUs, and I’m not really sure for Intel. 9900k was the last Intel CPU I ran, it seemed fine. I was running a 9900k/2080 on the same PSU as the 2500k/570 build.

[–] [email protected] 1 points 1 year ago* (last edited 1 year ago)

As for as the 2080 goes, like I said, it was big FOR THE TIME, and power hungry FOR THE TIME. It's still reasonable especially for today's standards

As for as the last two gens, 3000 and 4000 series, they are known to draw more than their rated power requirements, which, for their min recommended psu wattage, 3080 was 50 watts more than the 2080 (750w), and 4080 was 100 w more than that (850w)

To add to that, both of these gens of cards, when doing graphics intensive things like gaming, can overdraw power and have been known to cause hard shutdowns in pcs with PSUs that are even slightly higher rated than their min rec. Before these last two gens you could get away with a slightly lower than rated wattage PSU and sacrifice a little performance but that is definitely no longer the case.

And sure, the performance to watts used is better in the 3080, but they also run 10+ degrees hotter and the 4000 series even moreso.

I just hope the 5000 series goes the way of power consumption refinement rather than smashing more chips onto a board or vram fuckery like with the 4060, like I'd be happy with similar performance on the 5000 series if it was less power hungry

[–] [email protected] 2 points 1 year ago

The 7 series are more efficient than the 5 series. They just are programmed to go as fast as thermals allow. So the reviewers that had really powerful coolers on the cpus saw really high power draw. If instead you set a power cap, you get higher performance per watt than the previous generations.

Having the clocks scale to a thermal limit is a nice feature to have, but I don't think it should have been the default mode.

[–] [email protected] 1 points 1 year ago

Intel became less efficient because of how long they were stuck on 14nm. In order to compensate to beat amd in performance mindshare, they needed to push the clocks hard.

Overtime, cpus have been sitting closer to max clock, defeating the purpose of overclocking to many, where adding 1GHz was not out of the ordinary. Now getting 0.5GHz is an acheivement.