this post was submitted on 02 May 2024
2 points (100.0% liked)

Games

16738 readers
695 users here now

Video game news oriented community. No NanoUFO is not a bot :)

Posts.

  1. News oriented content (general reviews, previews or retrospectives allowed).
  2. Broad discussion posts (preferably not only about a specific game).
  3. No humor/memes etc..
  4. No affiliate links
  5. No advertising.
  6. No clickbait, editorialized, sensational titles. State the game in question in the title. No all caps.
  7. No self promotion.
  8. No duplicate posts, newer post will be deleted unless there is more discussion in one of the posts.
  9. No politics.

Comments.

  1. No personal attacks.
  2. Obey instance rules.
  3. No low effort comments(one or two words, emoji etc..)
  4. Please use spoiler tags for spoilers.

My goal is just to have a community where people can go and see what new game news is out for the day and comment on it.

Other communities:

Beehaw.org gaming

Lemmy.ml gaming

lemmy.ca pcgaming

founded 1 year ago
MODERATORS
top 13 comments
sorted by: hot top controversial new old
[–] [email protected] 0 points 6 months ago (2 children)

I know it's a small set, but for gaming and is honestly king. Unless you want the absolute "I'm willing to pay double the cost for 5% more performance" top of the line, amd is just great.

For AI and compute.... They're far behind. CUDA just wins. I hope a joint standard will be coming up soon, but until then Nvidia wins

[–] [email protected] 0 points 6 months ago (1 children)

I bought a 7900xtx and have been VERY happy with it.

[–] [email protected] 0 points 6 months ago (1 children)

That's what I'm running, and it's honestly better than my partners 3090

[–] [email protected] 0 points 6 months ago (1 children)

The only thing it’s missing is dedicated video decode hardware (which is mostly a convenience) and an equivalent to shadow play. Otherwise it’s a great alternative to a 4080/S

[–] [email protected] 0 points 6 months ago* (last edited 6 months ago) (1 children)

Is amd relive not equivalent to shadow play? Can record gameplay in av1 without issue

[–] [email protected] 1 points 6 months ago

You can even skip the whole suite if you don't need the AMD per game driver tweaks. OBS now come with direct AMD av1 support and also can record HDR content.(which relive can't do.)

[–] [email protected] -1 points 6 months ago

Exactly. I just can't justify a high end GPU purchase if I can't also get some work out of it.

[–] [email protected] 0 points 6 months ago (1 children)

Everyone I know who's bought a GPU recently has gone used, including me

[–] [email protected] 1 points 6 months ago

That and there just hasn't been much gains in performance in recent years, so it makes sense to not upgrade for a while. And a lot of people upgraded all at once during the pandemic, so there are less people on the market for a new GPU.

[–] [email protected] -2 points 6 months ago (1 children)

What's NVidia seeing in the gaming space? Or do they conflate gaming and ML sales?

[–] [email protected] 0 points 6 months ago (2 children)

Who would buy consumer grade hardware for machine learning?

[–] [email protected] 2 points 6 months ago

Almost everyone?

[–] [email protected] 1 points 6 months ago

There are many different niches of ML. 99% of hobbyist would use consumer grade hardware. It's quite frankly more than good enough.

Even in commercial usage, consumer GPUs provide better value unless you need to do something that very specifically require a huge vram pool. Like connecting multiple A100 GPUs to have hundreds or tens of thousands of gigabyte vram. Those use cases only come up if you're making base models for general purpose.

If you're using it for single person use case, something like 4090 is actually the best hardware. Enough ram to run almost anything and it's higher clock speed than enterprise GPU means your results come back faster.

Even training doesn't require that much vram. Chat models are generally more vram heavy but if you're doing specific image training like stable diffusion for how to render your face, or some specific fetish porn, you only really need like 12GB of vram to do it. There are ways to even do it at lower like 8GB but 12 is sweet value spot where even 3060 or 4060ti can do. Consumer GPUs will get that trained in like 30min to 24hrs depending on settings and model.