this post was submitted on 04 Sep 2023
39 points (100.0% liked)
Technology
37691 readers
346 users here now
A nice place to discuss rumors, happenings, innovations, and challenges in the technology sphere. We also welcome discussions on the intersections of technology and society. If it’s technological news or discussion of technology, it probably belongs here.
Remember the overriding ethos on Beehaw: Be(e) Nice. Each user you encounter here is a person, and should be treated with kindness (even if they’re wrong, or use a Linux distro you don’t like). Personal attacks will not be tolerated.
Subcommunities on Beehaw:
This community's icon was made by Aaron Schneider, under the CC-BY-NC-SA 4.0 license.
founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Let me flip it around again - humans regularly "hallucinate", it's just not something we recognize as such. There's neuro-atypical hallucinations, yes, but there's also misperceptions, misunderstandings, brain farts, and "glitches" which regularly occur in healthy cognition, and we have an entire rest of the brain to prevent those. LLMs are most comparable to "broca's area", which neurological case studies suggest naturally produces a stream of nonsense (see: split brain patients explaining the actions of their mute half). It's the rest of our "cognitive architecture" which conditions that raw language model to remain self-consistent and form a coherent notion of self. Honestly this discussion on "conceptualization" is poorly conceived because it's unfalsifiable and says nothing about the practical applications. Why do I care if the LLM can conceptualize if it does whatever subset of conceptualization I need to complete a natural language task?
AI is being super overhyped right now, which is unfortunate because it really is borderline miraculous, yet somehow they've overdone it. Emergent properties are empirical observations of behaviors they're able to at least semi-consistently demonstrate - where it becomes "eye of the beholder" is when we dither on about psychology and philosophy about whether or not they're some kind of "conscious" - I would argue they aren't, and the architecture makes that impossible without external aid, but "conscious(ness)" is such a broad term that it barely has a definition at all. I guess to speedrun the overhype misinformation I see:
I'll add more if I see or think of any. And if you have any specific questions, I'd be happy to answer. Also I should note, I'm of course using a lot of anthropomorphizing language here but it's the closest we have to describing these concepts. They're not human, and while they may have comparable behaviors in isolation, you can't accurately generalize all human behaviors and their interactions onto the models. Even if they were AGI or artificial people, they would "think" in fundamentally different ways.
If you want a more approachable but knowledgeable discussion on LLMs and their capabilities, I would recommend a youtuber named Dave Shapiro. Very interesting ideas, he gets a bit far into hype and futurism but those are more or less contained within their own videos.`
Can you please tone down on the fallacies? Until now I've seen the following:
And now, the quoted excerpt shows two more:
Could you please show a bit more rationality? This sort of shit is at the very least disingenuous, if not worse (stupidity), it does not lead to productive discussion. Sorry to be blunt but you're just wasting the time of everyone here, this is already hitting Brandolini's Law.
I won't address the rest of your comment (there's guilt by association there BTW), or further comments showing the same lack of rationality. However I had to point this out, specially for the sake of other posters.