this post was submitted on 21 Jan 2024
324 points (97.1% liked)
Technology
59197 readers
3533 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
My Nvidia 1070 with 8gb vram is still playing all of my games. Not everything gets Ultra, nor my monitor isn't 4K. Forever I am the "value buyer". It's hard to put money into something that is marginally better though. I thought 16g would be a no-brainer.
Exactly, people get to caught up in the Digital Foundry-ification of ultra max settings running at a perfect ~120 unlocked frames. Relax my dudes and remember the best games of your life were perfect dark with your friends running at 9 FPS.
1080p is fine, medium settings are fine. If the game is good you won't sweat the details.
The frame rate was shat on at the time and with good reason, that was unplayable for me. Best times were Halo 4-16 local multiplayer.
As someone who really doesn't care much for game graphics I feel that a comment I wrote a few months ago also fits here:
FPS games tend to be better to run at lower settings to be more competitive anyway. You don't want all the visual noise.
I kind of feel the same way about TV resolution. I have a 1080p TV and a 720p TV and I'm honestly fine with them. Sure, there's better quality out there, but I can always go to the movies if I want that. And I have the advantages of TVs without any 'smart' bullshit. They can't even connect to the internet.
I'm not saying no one else should buy 8k TVs or whatever, if that's what you want, fine, but there are plenty of people I've talked to who feel the same way as me, so I'm glad they haven't done anything like make us all change to new TVs again like they did when they updated to HD.
I literally have a higher resolution computer monitor than I do TV. My computer monitor costs more than my TV did too!
30fps is fine too on most games.....
friend of mine makes do with a gtx960@720p and is perfectly fine with it, the fun games run. even new ones.
maybe an upgrade to digital foundry perfect 120fps would be worth it if it werent so damn expensive nowadays outside the us.
Not to shill for them but Alex makes it a point to run tests and to include optimized settings for non flagship hardware in every review he does. I'm not sure where your digital foundry nomenclatures are coming from.
And no, 30fps is not fine...
i was referring to the op i was responding to
You lost me at 1080p. It's a basic quality of life thing. Even 1440p is a HUGE upgrade even for regular computer use not even gaming.
I run 4k but I use/need it more work space at work than gaming.
I agree that this happens to an extent but Digital Foundry in particular makes a point to take into account performance of the cards most used by regular people and are one of the biggest forces in that space pushing people to not just hit "ultra" and move on as you can see with their optimized settings series and the like, as well as getting the best out of older games as in their retro series. They like games that look good and play smoothly, of course, but I don't think it's fair to associate them with that kind of ULTRA MAX OR DIE attitude.
I think there's sometimes an overcorrection from the "gameplay over graphics" crowd. I've been part of that group before and get it, it's frustrating when from your perspective the industry is ignoring the parts of games that you care about the most. But it's a strange thing to pick on because at the end of the day pretty things that feel smooth to play are wonderful! That can be done on a toaster with beautiful pixel art / low poly 3D models, but it can also be done in dramatically different ways by pushing high end hardware to its limits. There's room for both and I adore both. Games are art like anything else and it'd be strange to tell people who appreciate going to a beautiful movie shot on particularly nice film on-location in expensive places just because it's still a good movie if you watch it on an old laptop with awful web compression or because an underground mumblecore film from 2003 is also great.
Graphics aren't all that matter to me but if the primary joy someone gets from gaming is seeing ultra-detailed and perfectly rendered scenes the best way they possibly can, good for them. Personally, I like getting good visuals when I can but my primary concern is always framerate, as particularly in first person games even 60fps often triggers my motion sickness and forces me to stick to short sessions. Ultimately I see the this whole debate as a relic of the past that only made sense when the only games the average person had access to were AAA/AA releases. Low-spec gaming is better than it has ever been, with the indie scene continuing to go strong like it has for the past 15+ years and an ever-expanding backlog of classics which now run on just about anything every year.