this post was submitted on 17 Jun 2024
12 points (83.3% liked)
Videos
14270 readers
213 users here now
For sharing interesting videos from around the Web!
Rules
- Videos only
- Follow the global Mastodon.World rules and the Lemmy.World TOS while posting and commenting.
- Don't be a jerk
- No advertising
- No political videos, post those to [email protected] instead.
- Avoid clickbait titles. (Tip: Use dearrow)
- Link directly to the video source and not for example an embedded video in an article or tracked sharing link.
- Duplicate posts may be removed
Note: bans may apply to both [email protected] and [email protected]
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
The first half of this video is entirely dumb, which is shocking, because the second half actually accurately describes the issues the first half makes out to be "mysterious". It's not at all.
We can view model decisions AFTER they execute, but they are too fast to observe live. This is why constraints are put into place for reinforcement models to begin with. You want an expected outcome, just fast.
This video is confusing two different worlds that operate completely different from each other: computer vision models, and generative models.
We know exactly why vision models do what they do, because it's predetermined, and a result is expected. Training these models includes large sample sets which can be observed, and the resulting model has outputs describing what happened during training. There are a jillion tools out there that let you even run a step-by-step of such models to see what the before and after of the input is, and allow you to adjust to your liking if the result is not correct. We wouldn't be able to program them if not.
Generative models that are predictive operate differently. They attempt to guess a variation of input after a few filters, and then sort of run on their own. This is not reinforced learning, and is why it differs heavily from what this video describes.
There's a massive difference between the different operations of neural networks, and this video just confuses all of them in some spots, but accurately describes them in others. It's all over the place.
Base fact being that a model meant for vision is not having the same issues as one meant for languages or deep learning.
Yea, the video doesn't make it super clear, it's not a generative model at all. Those weird "ai" looking images are the result of taking the specific node and applying a filter to visualize what the node is looking for