this post was submitted on 15 Jun 2024
439 points (99.3% liked)

Privacy

31993 readers
415 users here now

A place to discuss privacy and freedom in the digital world.

Privacy has become a very important issue in modern society, with companies and governments constantly abusing their power, more and more people are waking up to the importance of digital privacy.

In this community everyone is welcome to post links and discuss topics related to privacy.

Some Rules

Related communities

Chat rooms

much thanks to @gary_host_laptop for the logo design :)

founded 5 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 8 points 5 months ago (1 children)

Would you give your perspective anyway, as I would be quite interested, although I'm not the one you talked to?

[–] Hackworth 6 points 5 months ago* (last edited 5 months ago) (4 children)

Sure, thanks for your interest. It's an incomplete picture, but we can think of LLMs as an abstraction of all the meaningful connections within a dataset to a higher dimensional space - one that can be explored. That alone is an insane accomplishment that is changing some of the pillars of data analysis and knowledge work. But that's just the contribution of the "Attention is All You Need" paper. Many implementations of modern generative AI combine LLM inference in agentic networks, with GANs, and with rules-based processing. Extracting connections is just one part of one part of a modern AI implementation.

The emergent properties of GPT4 are enough to point toward this exponential curve continuing. Theory of mind (and therefore deception) as well as relational spatial awareness (usually illustrated with stacking problems) developed solely from increasing the parameter count describing the neural network. These were unexpected capabilities. As a result, there is an almost literal arms race on the hardware side to see what other emergent properties exist at higher model sizes. With some poetic license, we're rending function from form so quickly and effectively that it's seen by some as freeing and others as a sacrilege.

Some of the most interesting work on why these capabilities emerge and how we might gain some insight (and control) from exploring the mechanisms is being done by Anthropic and by users at Hugging Face. They discovered that when specific neurons in Claude's net are stimulated, everything it responds with will in some way become about the Golden Gate Bridge, for instance. This sort of probing is perhaps a better route to progress than blindly chasing more size (despite its recent success). But only time will tell. Certainly, Google and MS have had a lot of unforced errors fumbling over themselves to stay in what they think is the race.

[–] [email protected] 6 points 5 months ago

Thank you very much for those insights!!

[–] AIhasUse 4 points 5 months ago

Thanks so much for taking the time to explain this. I was just going to give them a link.

[–] [email protected] 1 points 5 months ago (1 children)

I'm happy to take the time to alter your perspective, if you are open to new information.

You took some time, but spent it explaining at a fairly technical level, rather than a lamens term approach. I doubt you managed to change many people's perspective, but you maybe reinforced some.

[–] Hackworth 3 points 5 months ago* (last edited 5 months ago)

This is another good use case for gAI. Copy/paste the comment into a GPT and tell it to re-write the content at the desired reading or technical level. Then it's available for follow-up clarification questions.

[–] Delonix 1 points 5 months ago (1 children)
[–] Hackworth 2 points 5 months ago