this post was submitted on 06 Oct 2023
2941 points (98.2% liked)
Piracy: ꜱᴀɪʟ ᴛʜᴇ ʜɪɢʜ ꜱᴇᴀꜱ
54636 readers
805 users here now
⚓ Dedicated to the discussion of digital piracy, including ethical problems and legal advancements.
Rules • Full Version
1. Posts must be related to the discussion of digital piracy
2. Don't request invites, trade, sell, or self-promote
3. Don't request or link to specific pirated titles, including DMs
4. Don't submit low-quality posts, be entitled, or harass others
Loot, Pillage, & Plunder
📜 c/Piracy Wiki (Community Edition):
💰 Please help cover server costs.
Ko-fi | Liberapay |
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I mean... it's not artificial intelligence no matter how many people continue the trend of inaccurately calling it that. It's a large language model. It has the ability to write things that look disturbingly close, even sometimes indistinguishable, to actual human writing. There's no good reason to mistake that for actual intelligence or rationality.
It seems to me that you misunderstand what artificial intelligence means. AI doesn't necessitate thought or sentience. If a computer can perform a complex task that is indistinguishable from the work of a human, it will be considered intelligent.
You may consider the classic turing test, which doesn't question why a computer program answers the way it does, only that it is indiscernable from a human response.
You may also consider this quote from John McCarthy on the topic:
There's more on this topic by IBM here.
You may also consider a few extra definitions:
Yep, all those definitions are correct and corroborate what the user above said. An LLM does not learn like an animal learns. They aren't intelligent. They only reproduce patterns similar to human speech. These aren't the same thing. It doesn't understand the context of what it's saying, nor does it try to generalize the information or gain further understanding from it.
It may pass the Turing test, but that's neither a necessary nor sufficient condition for intelligence. It is just a useful metric.
LLMs are expert systems, who's expertise is making believable and coherent sentences. They can "learn" to be better at their expert task, but they cannot generalise into other tasks.