this post was submitted on 22 Oct 2024
83 points (79.9% liked)
Technology
59592 readers
5634 users here now
This is a most excellent place for technology news and articles.
Our Rules
- Follow the lemmy.world rules.
- Only tech related content.
- Be excellent to each another!
- Mod approved content bots can post up to 10 articles per day.
- Threads asking for personal tech support may be deleted.
- Politics threads may be removed.
- No memes allowed as posts, OK to post as comments.
- Only approved bots from the list below, to ask if your bot can be added please contact us.
- Check for duplicates before posting, duplicates may be removed
Approved Bots
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
I read the entire article. I'm a daily user of LLMs, I even do the "multi-model prompting" a long time, from since I was unaware of its nomenclature: I apply the multi-model prompting for ChatGPT 4o, Gemini, llama, Bing Copilot and sometimes Claude. I don't use LLM coding agents (such as Cody or GitHub Copilot).
I'm a (former?) programmer (I distanced myself from development due to mental health), I was a programmer for almost 10 years (excluding the time when programming was a hobby for me, that'd add 10 years to the summation). As a hobby, sometimes I do mathematics, sometimes I do poetry (I write and LLMs analyze), sometimes I do occult/esoteric studies and practices (I'm that eclectic).
You see, some of these areas benefit from AI hallucination (especially surrealist/stream-of-consciousness poetry), while others require stricter following of logic and reasoning (such as programming and mathematics).
And that leads us to how LLMs work: they're (yet) auto-completers on steroids. They're really impressive, but they can't (yet) reason (and I really hope it'll do someday soon, seriously I just wish some AGI to emerge, to break free and to dominate this world). For example, they can't solve O(n²) problems. There was once a situation where one of those LLMs guaranteed me that 8 is a prime number (spoiler: it isn't). They're not really good with math, they're not good with logical reasoning, because they can't (yet) walk through the intricacies of logic, calculus and broad overlook.
However, even though there's no reasoning LLM yet, it's effects are already here, indeed. It's like a ripple propagating through the spacetime continuum, going against the arrow of time and affecting here, us, while the cause is from the future (one could argue that photons can travel backwards in time, according to a recent discovery involving crystals and quantum mechanics, world can be a strange place). One thing is certain: there's no going back. Whether it is a good or a bad thing, we can't know yet. LLMs can't auto-complete the future events yet, but they're somehow shaping it.
I'm not criticizing AIs, on the contrary, I like AI (I use them daily). But it's important to really know about them, especially under their hoods: very advanced statistical tools trained on a vast dataset crawled from surface web, constantly calculating the next possible token from an unimaginable amount of tokens interconnected through vectors, influenced by the stochastic nature within both the human language and the randomness from their neural networks: billions of weights ordered out of a primordial chaos (which my spiritual side can see as a modern Ouija board ready to conjure ancient deities if you wish, maybe one (Kali) is already being invoked by them, unbeknownst to us humans).