this post was submitted on 16 May 2024
856 points (97.6% liked)
Funny
6773 readers
683 users here now
General rules:
- Be kind.
- All posts must make an attempt to be funny.
- Obey the general sh.itjust.works instance rules.
- No politics or political figures. There are plenty of other politics communities to choose from.
- Don't post anything grotesque or potentially illegal. Examples include pornography, gore, animal cruelty, inappropriate jokes involving kids, etc.
Exceptions may be made at the discretion of the mods.
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
One of the points of the books is that the laws were inherently flawed.
Given that we're talking about a Google product, you might have more success asking if they're bound by the Ferengi Rules of Acquisition?
IDK if I missed something or I just disagree, but I remember all but maybe one short story ending up with the laws working as intended (though unexpectedly) and humanity being better as a result.
Didn't they end with humanity being controlled by a hyper-intelligent benevolent dictator, which ensured humans were happy and on a good path?
I thought it was Asiimovs books, but apparently not. Which one had the 3 fundamental rules lead to the solution basically being: "Humans can not truly be safe unless they're extinct" or something along those lines... Been a long time since I've explored the subjects.
I mean... Kind of Asimov's robot series? Except the androids/robots were trying so hard to stay to the rules and protect humans but at every chance they could humans fucked that up or refused to see the plan.
At least as I recall, the robots basically came up with multi-millenia spanning plans that would solve all of humanity's problems and then humans were like: "Cool. But what if we made robots detectives and they could also target people we don't like?" Then the robots fucked off for a while and a bunch of stuff happened and... Yeah. Asimov wrote a lot of books.