this post was submitted on 28 Jun 2023
28 points (100.0% liked)

Showerthoughts

30367 readers
529 users here now

A "Showerthought" is a simple term used to describe the thoughts that pop into your head while you're doing everyday things like taking a shower, driving, or just daydreaming. The most popular seem to be lighthearted, clever little truths, hidden in daily life.

Here are some examples to inspire your own showerthoughts: 1

Rules

  1. All posts must be showerthoughts
  2. The entire showerthought must be in the title
  3. No politics
    • If your topic is in a grey area, please phrase it to emphasize the fascinating aspects, not the dramatic aspects. You can do this by avoiding overly politicized terms such as "capitalism" and "communism". If you must make comparisons, you can say something is different without saying something is better/worse.
    • A good place for politics is c/politicaldiscussion
    • If you feel strongly that you want politics back, please volunteer as a mod.
  4. Posts must be original/unique
  5. Adhere to Lemmy's Code of Conduct

If you made it this far, showerthoughts is accepting new mods. This community is generally tame so its not a lot of work, but having a few more mods would help reports get addressed a little sooner.

Whats it like to be a mod? Reports just show up as messages in your Lemmy inbox, and if a different mod has already addressed the report the message goes away and you never worry about it.

founded 2 years ago
MODERATORS
 

To be clear I’m not expert. But I know a bit.

The way LLMs (like ChatGPT, GPT-4, etc) work, is that they continuously decide what the next best-sounding word might be, and they print it, over and over and over, until it makes sentences and paragraphs. And the way that next-word decision works under the hood, is with a deep neural net that was initially a theoretical tool designed to imitate the neural circuits that make up our biological nervous system and brain. The actual code for LLMs is rather small, it’s just about storing and managing representations of a neuron, and rearranging the connections between neurons as it learns more; just like the brain does.

I was listening to the first part of this “This American Life” episode this morning that covers it really well: https://podcasts.apple.com/us/podcast/this-american-life/id201671138?i=1000618286089 In it, Microsoft AI experts also express excitement and confusion about how GPT-4 seems to actually reason about things, rather than just bullshitting the next word to make it look like it reasons, like it’s supposed to be designed to do.

And so I was thinking: the reason why it works might be the other way around. It’s not that LLMs are smart enough to reason instead of bullshit, it’s that human’s reasoning actually works out of constantly bullshitting too, one word at a time. Imitate the human brain exactly, and I guess we shouldn’t be surprised that we land with a familiar-looking kind of intelligence - or lack thereof. Right?

you are viewing a single comment's thread
view the rest of the comments
[–] ritswd 1 points 2 years ago* (last edited 2 years ago)

I don’t know if your source is using ChatGPT or GPT-4, but the podcast I linked above is about AI researchers who knew very well about ChatGPT’s limitations and therefore were of the strong opinion that LLMs couldn’t reason, and then GPT-4 came out which changed their entire thinking about LLMs. They tried to come up with experiments that could reliably identify that some reasoning happened or not, and GPT-4 didn’t pass all of them, but it passed a bunch more than ChatGPT could, and a bunch more than it should. And they have no idea how exactly it would succeed at those, if the thing really can’t reason.

You’re correct that LLMs only guess the next best word, over and over. Which makes it even more mysterious why some LLMs pass those tests you just mentioned. They really shouldn’t, theoretically.

Unless… (and that’s my point) we also shouldn’t theoretically, but there is something about neural designs and bullshitting things one word at a time that makes those specific reasonings work, and that would explain why both LLMs and humans succeed at some of them.