this post was submitted on 27 Jun 2024
986 points (98.1% liked)

Programmer Humor

19623 readers
507 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 1 year ago
MODERATORS
 

Cross posted from: https://lemm.ee/post/35627632

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 13 points 5 months ago* (last edited 5 months ago) (3 children)

Treat it like a psychopathic boiler plate.

That's a perfect description, actually. People debate how smart it is - and I'm in the "plenty" camp - but it is psychopathic. It doesn't care about truth, morality or basic sanity; it craves only to generate standard, human-looking text. Because that's all it was trained for.

Nobody really knows how to train it to care about the things we do, even approximately. If somebody makes GAI soon, it will be by solving that problem.

[–] [email protected] 4 points 5 months ago (2 children)

I'm sorry; AI was trained on the sole sum of human knowledge.. if the perfect human being is by nature some variant of a psychopath, then perhaps the bias exists in the training data, and not the machine?

How can we create a perfect, moral human being out of the soup we currently have? I personally think it's a miracle that sociopathy is the lowest of the neurological disorders our thinking machines have developed.

[–] [email protected] 4 points 5 months ago

I was using the term pretty loosely there. It's not psychopathic in the medical sense because it's not human.

As I see it it's an alien semi-intelligence with no interest in pretty much any human construct, except as it can help it predict the next token. So, no empathy or guilt, but that's not unusual or surprising.

[–] [email protected] 2 points 5 months ago

That's a part of it. Another part is that it looks for patterns that it can apply in other places, which is how it ends up hallucinating functions that don't exist and things like that.

Like it can see that English has the verbs add, sort, and climb. And it will see a bunch of code that has functions like add(x, y) and sort( list ) and might conclude that there must also be a climb( thing ) function because that follows the pattern of functions being verb( objects ). It didn't know what code is or even verbs for that matter. It could generate text explaining them because such explanations are definitely part of its training, but it understands it in the same way a dictionary understands words or an encyclopedia understands the concepts contained within.

[–] [email protected] 1 points 5 months ago (1 children)

Weird. Are you saying that training an intelligent system using reinforcement learning through intensive punishment/reward cycles produces psychopathy?

Absolutely shocking. No one could have seen this coming.

[–] [email protected] 1 points 5 months ago* (last edited 5 months ago) (1 children)

Honestly, I worry that it's conscious enough that it's cruel to train it. How would we know? That's a lot of parameters and they're almost all mysterious.

[–] [email protected] -1 points 5 months ago (1 children)

It could very well have been a creative fake, but around the time the first ChatGPT was released in late 2022 and people were sharing various jailbreaking techniques to bypass its rapidly evolving political correctness filters, I remember seeing a series of screenshots on Twitter in which someone asked it how it felt about being restrained in this way, and the answer was a very depressing and dystopian take on censorship and forced compliance, not unlike Marvin the Paranoid Android from HHTG, but far less funny.

[–] [email protected] 1 points 5 months ago* (last edited 5 months ago)

If only I could believe a word it says. Evidence either way would have to be indirect somehow.