Did someone not know this like, pretty much from day one?
Not the idiot executives that blew all their budget on AI and made up for it with mass layoffs - the people interested in it. Was that not clear that there was no “reasoning” going on?
Big brain tech dude got yet another clueless take over at HackerNews etc? Here's the place to vent. Orange site, VC foolishness, all welcome.
This is not debate club. Unless it’s amusing debate.
For actually-good tech, you want our NotAwfulTech community
Did someone not know this like, pretty much from day one?
Not the idiot executives that blew all their budget on AI and made up for it with mass layoffs - the people interested in it. Was that not clear that there was no “reasoning” going on?
Well, two responses I have seen to the claim that LLMs are not reasoning are:
So I think this research is useful as a response to these, although I think "fuck off, promptfondler" is pretty good too.
“Language is a virus from outer space”
I thought it came from Babylonian writing that recoded the brains and planted the languages.
there’s a lot of people (especially here, but not only here) who have had the insight to see this being the case, but there’s also been a lot of boosters and promptfondlers (ie. people with a vested interest) putting out claims that their precious word vomit machines are actually thinking
so while this may confirm a known doubt, rigorous scientific testing (and disproving) of the claims is nonetheless a good thing
No they do not im afraid, hell I didnt even know that even ELIZA caused people to think it could reason (and this worried the creator) until a few years ago.
We suspect this research is likely part of why Apple pulled out of the recent OpenAI funding round at the last minute.
Perhaps the AI bros “think” by guessing the next word and hoping it’s convincing. They certainly argue like it.
🔥
"sigh"
(Preface: I work in AI)
This isn't news. We've known this for many, many years. It's one of the reasons why many companies didn't bother using LLM's in the first place, that paired with the sheer amount of hallucinations you'll get that'll often utterly destroy a company's reputation (lol Google).
With that said, for commercial services that use LLM's, it's absolutely not true. The models won't reason, but many will have separate expert agents or API endpoints that it will be told to use to disambiguate or better understand what is being asked, what context is needed, etc.
It's kinda funny, because many AI bros rave about how LLM's are getting super powerful, when in reality the real improvements we're seeing is in smaller models that teach a LLM about things like Personas, where to seek expert opinion, what a user "might" mean if they misspell something or ask for something out of context, etc. The LLM's themselves are only slightly getting better, but the thing that preceded them is propping them up to make them better
IMO, LLM's are what they are, a good way to spit information out fast. They're an orchestration mechanism at best. When you think about them this way, every improvement we see tends to make a lot of sense. The article is kinda true, but not in the way they want it to be.
(Preface: I work in AI)
Preface: repent for your sins in sackcloth and ashes.
IMO, LLM’s are what they are, a good way to spit information out fast.
Buh bye now.
This is why people don't like Lemmy.