this post was submitted on 01 Jun 2024
131 points (92.8% liked)

PC Gaming

8568 readers
501 users here now

For PC gaming news and discussion. PCGamingWiki

Rules:

  1. Be Respectful.
  2. No Spam or Porn.
  3. No Advertising.
  4. No Memes.
  5. No Tech Support.
  6. No questions about buying/building computers.
  7. No game suggestions, friend requests, surveys, or begging.
  8. No Let's Plays, streams, highlight reels/montages, random videos or shorts.
  9. No off-topic posts/comments, within reason.
  10. Use the original source, no clickbait titles, no duplicates. (Submissions should be from the original source if possible, unless from paywalled or non-english sources. If the title is clickbait or lacks context you may lightly edit the title.)

founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] -2 points 5 months ago (1 children)

There is literally a university study that proves over 50% failure in programming tasks. It's not a rational model, deal with it, get off the Kool aid, and move on.

[–] [email protected] 3 points 5 months ago (1 children)

If you didn't have an agenda/preconceived idea you wanted proven, you'd understand that a single study has never been used by any credible scientist to say anything is proven, ever.

Only people who don't understand how data works will say a single study from a single university proves anything, let alone anything about a model trained on billions of parameters across a field as broad as "programming".

I could feed GPT "programming" tasks that I know it would fail on 100% of the time. I also could feed it "programming" tasks I know it would succeed on 100% of the time. If you think LLMs have nothing to offer programmers, you have no idea how to use them. I've been successfully using GPT4T for months now, and it's been very good. It's better in static environments where it can be fed compiler errors to fix itself continually (if you ever look at more than a headline about GPT performance you'd know there's a substantial difference between zero-shot and 3-shot performance).

Bugs exist, but code heavily written by LLMs has not been proven to be any more or less buggy than code heavily written by junior devs. Our internal metrics have them within any reasonable margin of error (senior+GPT recently beating out senior+junior, but it's been flipping back and forth), and senior+GPT tickets get done much faster. The downside is GPT doesn't become a senior, where a junior does with years of training, though 2 years ago LLMs were at a 5th grade coding level on average, and going from 5th grade to surpassing college level and matching junior output is a massive feat, even if some luddites like yourself refuse to accept it.

[–] [email protected] -4 points 5 months ago (1 children)
[–] [email protected] 2 points 5 months ago

Holy shit you're right, I'm an idiot. Thanks for helping me shift my perspective.