this post was submitted on 30 Dec 2024
653 points (98.5% liked)

Programmer Humor

19817 readers
138 users here now

Welcome to Programmer Humor!

This is a place where you can post jokes, memes, humor, etc. related to programming!

For sharing awful code theres also Programming Horror.

Rules

founded 2 years ago
MODERATORS
653
AI in reality (slrpnk.net)
submitted 3 days ago* (last edited 3 days ago) by [email protected] to c/[email protected]
 
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 9 points 3 days ago (5 children)

While I haven't read the paper, the comment's explanation seems to make sense. It supposedly contains a mathematical proof that making AGI from a finite dataset is a NP-hard problem. I have to read it and parse out the reasoning, if true, it would make for a great argument in cases like these.

https://lemmy.world/comment/14174326

[–] [email protected] 2 points 3 days ago (4 children)

If that is true, how does the brain work?

Call everything you have ever experienced the finite dataset.
Constructing your brain from dna works in a timely manner.
Then training it does too, you get visibly smarter with time, so on a linear scale.

[–] [email protected] 8 points 3 days ago* (last edited 3 days ago) (3 children)

I think part of the problem is that LLMs stop learning at the end of the training phase, while a human never stops taking in new information.

Part of why I think AGI is so far away is because to run the training in real-time like a human, it would take more compute than currently exists. They should be focusing on doing more with less compute to find new more efficient algorithms and architectures, not throwing more and more GPUs at the problem. Right now 10x the GPUs gets you like 5-10% better accuracy on whatever benchmarks, which is not a sustainable direction to go.

[–] veni_vedi_veni 2 points 3 days ago (1 children)

How does conversation context work though? Is that memory not a form of learning?

[–] [email protected] 6 points 3 days ago* (last edited 3 days ago) (1 children)

The context window is a fixed size. If the conversation gets too long, the start will get pushed out and the AI will not remember anything from the start of the conversation. It's more like having a notepad in front of a human, the AI can reference it, but not learn from it.

[–] [email protected] 3 points 3 days ago

Also, a key part of how GPT-based LLMs work today is they get the entire context window as their input all at once. Where as a human has to listen/read a word at a time and remember the start of the conversation on their own.

I have a theory that this is one of the reasons LLMs don't understand the progression of time.

load more comments (1 replies)
load more comments (1 replies)
load more comments (1 replies)