this post was submitted on 22 Dec 2024
514 points (96.2% liked)
Technology
60076 readers
3561 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I suspect that if you took into consideration the millions of generations of evolution that "trained" the basic architecture of our brains, that advantage would shrink considerably.
I disagree. I'd argue evidence suggests we're just a more sophisticated version of a similar principle, refined over billions of years. We learn facts by rote, and learn similarities by rote until we develop enough statistical text (or audio) correlations to "understand" the world.
Conversations are a slightly meandering chain of statistically derived cliches. English adjective order is universally "understood" by native speakers based purely on what sounds right, without actually being able to explain why (unless you're a big grammar nerd). More complex conversations might seem novel, but they're just a regurgitation of rote memorized facts and phrases strung together in a way that seems appropriate to the conversation based on statistical experience with past conversations.
As with the evolution of our brains, which have operated on basically the same principles for hundreds of millions of years. The special sauce between human intelligence and a flatworm's is a refined model.
I'm not sure you can claim that absolutely. That kind of feature is an internal experience, you can't really confirm or deny if a GPT has something similar. Besides, humans have a pretty tenuous relationship with the concept of truth. There are certainly humans that consider objective falsehoods to be Truth.
Agree to disagree.
There is a lot that can be discussed in a philosophical debate. However, any 8 years old would be able to count how many letters are in a word. LLMs can't reliably do that by virtue of how they work. This suggests me that it's not just a model/training difference. Also evolution over million of years improved the "hardware" and the genetic material. Neither of this is compares to computing power or amount of data which is used to train LLMs.
I believe a lot of this conversation stems from the marketing (calling "intelligence") and the anthropomorphization of AI.
Anyway, time will tell. Personally I think it's possible to reach a general AI eventually, I simply don't think the LLMs approach is the one leading there.
Actually humans have more computing power than is required to run an LLM. You have this backwards. LLMs are comparably a lot more efficient given how little computing power they need to run by comparison. Human brains as a piece of hardware are insanely high performance and energy efficient. I mean they include their own internal combustion engines and maintenance and security crew for fuck's sake. Give me a human built computer that has that.
I agree here. I do think though that LLMs are closer than you think. They do in fact have both attention and working memory, which is a large step forward. The fact they can only process one medium (only text) is a serious limitation though. Presumably a general purpose AI would ideally have the ability to process visual input, auditory input, text, and some other stuff like various sensor types. There are other model types though, some of which take in multi-modal input to make decisions like a self-driving car.
I think a lot of people romanticize what humans are capable of while dismissing what machines can do. Especially with the processing power and efficiency limitations that come with the simple silicon based processors that current machines are made from.