this post was submitted on 11 Nov 2024
34 points (97.2% liked)
PC Gaming
8576 readers
211 users here now
For PC gaming news and discussion. PCGamingWiki
Rules:
- Be Respectful.
- No Spam or Porn.
- No Advertising.
- No Memes.
- No Tech Support.
- No questions about buying/building computers.
- No game suggestions, friend requests, surveys, or begging.
- No Let's Plays, streams, highlight reels/montages, random videos or shorts.
- No off-topic posts/comments, within reason.
- Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
As I understand it, the assertion is that the 1080p FPS is the same as 2k/4k FPS, assuming that you have an infinitely powerful GPU. So the 1080p FPS is your max potential FPS at any resolution with the CPU, and then you need to look at a GPU 2k/4k chart to see how much FPS it can achieve from that target. HWUnboxed also reasons that gamers are not blindly using ultra settings, so in real scenarios people are going to be lowering their settings to try to achieve a specific FPS target anyway. They also mention that lowering ingame settings doesn't usually affect the CPU FPS benchmark.
So in summary, the 1080p CPU benchmark is the ~highest possible target you can achieve, and then it's up to your GPU and ingame settings to decide how much of that target you can reach. It's a little more difficult to grasp and calculate mentally, but it prevents the 2k/4k benchmark data from showing what is effectively misleading "point in time" data that will not be useful if you have a different GPU or ingame settings. This is most clearly demonstrated by re-reviewing older CPUs in the future-proof section and showing that putting massive GPUs on old CPUs puts the FPS benchmarks of all resolutions to roughly the same value - i.e. the CPU doesn't truly have an effect w/r/t resolution, it's mainly just the GPU.
Yes, I understand the theory but at the end of the day I watch reviews so I don’t have to interpolate data or make guessing. If I see something like this I can clearly see that, given the same GPU I can expect that my 5900x should be around 6% slower than the 9800X 3D at 4k on average and here I can see that it is around 15% slower at 1440p which is useful for when using DLSS. This makes clear for me that an upgrade is not worth it. Those are the kind of details I want to view on reviews.