this post was submitted on 27 Dec 2024
375 points (95.2% liked)

Technology

60285 readers
4230 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 13 points 1 week ago (1 children)

It's a basic argument of generative complexity, I found the article some years ago while trying to find an earlier one (I don't think by the same author) that argued along the same complexity lines, essentially saying that if we worked like AI folks think we do we'd need to be so and so much trillion parameters and our brains would be the size of planets. That article talked about the need for context switching in generating (we don't have access to our cooking skills while playing sportsball), this article talks about the necessity to be able to learn how to learn. Not just at the "adjust learning rate" level, but mechanisms that change the resulting coding, thereby creating different such contexts, or at least that's where I see the connection between those two. In essence: To get to AGI we need AIs which can develop their own topology.

As to "rudeness": Make sure to never visit the Netherlands. Usually how this goes is that I link the article and the AI faithful I pointed it out to goes on a denial spree... because if they a) are actually into the topic, not just bystanders and b) did not have some psychological need to believe (including "my retirement savings are in AI stock") they c) would've come across the general argument themselves during their technological research. Or came up with it themselves, I've also seen examples of that: If you have a good intuition about complexity (and many programmers do) it's not unlikely a shower thought to have. Not as fleshed out as in the article, of course.