this post was submitted on 30 May 2024
166 points (97.7% liked)

Videos

14419 readers
434 users here now

For sharing interesting videos from around the Web!

Rules

  1. Videos only
  2. Follow the global Mastodon.World rules and the Lemmy.World TOS while posting and commenting.
  3. Don't be a jerk
  4. No advertising
  5. No political videos, post those to [email protected] instead.
  6. Avoid clickbait titles. (Tip: Use dearrow)
  7. Link directly to the video source and not for example an embedded video in an article or tracked sharing link.
  8. Duplicate posts may be removed

Note: bans may apply to both [email protected] and [email protected]

founded 2 years ago
MODERATORS
all 24 comments
sorted by: hot top controversial new old
[–] [email protected] 48 points 6 months ago (5 children)

It's actually worse.

The video focuses on how you're leaking personal info all the time through the software that you use and the connections that you make, and ways to mitigate it.

However, have you guys heard about forensic linguistics? That's how the Unabomber was caught. The way that you use your language(s) is pretty unique to yourself, and can be used to uncover your identity. This was done manually by two guys, Fitzgerald and Shuy; they were basically identifying patterns in how Unabomber wrote to narrow down the suspects further and further, until they hit the right guy.

Now, let's talk about large "language" models, like Gemini or ChatGPT. Frankly, I believe that people who think that LLMs are "intelligent" or "comprehend language" themselves lack intelligence and language comprehension. But they were made to find and match patterns in written text, and rather good at it.

Are you getting the picture? What Fitzgerald and Shuy did manually 30 years ago can be automated now. And it gets worse, note how those LLMs "happen" to be developed by companies that you can't trust to die properly (Google, Amazon, Facebook, Apple, Microsoft and its vassal OpenAI).

So, while the video offers some solid advice regarding privacy, sadly it is not enough. If you're in some deep shit, and privacy is a life-or-death matter for you, I strongly advise you be always mindful of what and how you write.

And, for the rest of us: fighting individually for our right to privacy is not enough. We need to assemble and organise ourselves, to fight on legal grounds against those who are trying to kill it. You either fight for your rights or you lose them.

Just my two cents. I apologise as this is just side-related to the video, but I couldn't help it.

[–] [email protected] 25 points 6 months ago (2 children)

Conversely, you can now have your manifesto written by a locally run LLM.

[–] [email protected] 3 points 6 months ago

“Revise generically: [manifesto]” (certainly not the best prompt)

Folks seem to like Ollama per HackerNews threads: in a coding context here:

Not using Codestral (yet) but check out Continue.dev[1] with Ollama[2] running llama3:latest and starcoder2:3b. It gives you a locally running chat and edit via llama3 and autocomplete via starcoder2.

It's not perfect but it's getting better and better.

[1] https://www.continue.dev/ [2] https://ollama.com/

Please no unabombing though

Oh wow he wrote a 35k word manifesto… feel like that’s so rare you’d still stand a solid chance at being identified somehow.

[–] [email protected] 1 points 6 months ago

You could, but even then you need to put some thought on how to prompt and review/edit the output.

I've noticed from usage that LLMs are extremely prone to repeat verbatim words and expressions from the prompt. So if you ask something like "explain why civilisation is bad from the point of view of a cool-headed logician", you're likely outing yourself already.

A lot of the times the output will have "good enough" synonyms. That you could replace with more accurate words... and then you're outing yourself already. Or simply how you fix it so it sounds like a person instead of a chatbot, we all have writing quirks and you might end leaking them into the review.

And more importantly you need to aware that it is an issue, and that you can be tracked based on how and what you write.

[–] UmeU 13 points 6 months ago (1 children)

While forensic linguistics is pretty cool, the Unabomber was caught because they released his manifesto and his brother’s wife and brother recognized the unusual phrasing such as ‘Eat your cake and have it too’.

If an author has a large amount of known works then it’s not too difficult to identify other writings by that same author. But if the author does not have a large body of writing that is known to come from that individual, then the best we can do is determine an approximate age and geographic location where the Individual grew up, and that’s only when the unidentified writing is large enough, like in the case of the Unabomber where his manifesto was 30k words.

[–] [email protected] 1 points 6 months ago (1 children)

I did simplify the whole thing, as you noticed; but note that his SIL and brother identifying him is another example of the same process, David knew that expressions that Ted used like "cool-headed logicians" were highly unusual, not too unlike what the socio- and forensic linguists did there.

But if the author does not have a large body of writing that is known to come from that individual

Such as a Lemmy or Facebook account? Or any other online account associated with your writing, really; we produce far more text in the internet than ourselves realise.

And while a priori, your different accounts through different websites might look completely disconnected, as you connect two of them as coming from the same person, connecting a third one is easier. And a fourth. So goes on.

A small caveat is that while the corpus is bigger, so is the noise introduced by people from the other side of the world that happen to use the same patterns as the person whom you want to identify. Even then, I believe that the ability to bulk process text to find authorship grew considerably faster than the number of potential matches.

[–] [email protected] 0 points 6 months ago

Additional signal is not noise

[–] [email protected] 6 points 6 months ago (2 children)

Is there a best of on Lemmy

[–] StaticFalconar 4 points 6 months ago

Best of misinformation. The Unabomber was caught because the person that recognized the writing was his own family members. What the hell is this revisionist BS?

[–] [email protected] 3 points 6 months ago (1 children)
[–] [email protected] 4 points 6 months ago

Hi there! Looks like you linked to a Lemmy community using a URL instead of its name, which doesn't work well for people on different instances. Try fixing it like this: [email protected]

[–] AtariDump 3 points 6 months ago* (last edited 6 months ago) (2 children)

You can’t eat your cake and have it too.

[–] [email protected] 2 points 6 months ago* (last edited 6 months ago) (1 children)

Let's see how well this quote ages as subscription services become ubiquitous.

[–] AtariDump 2 points 6 months ago
[–] [email protected] 0 points 6 months ago

You can’t eat your cake and have it too.

And I can't "magically" know what you're referring to, either - given that you're replying to a rather long comment but providing exactly zero context on what specifically you're replying to.

Quotes, use them.

[–] [email protected] 1 points 6 months ago (1 children)

I don't believe they are intelligent now but I do believe we evolved as basically pattern matching machines.

[–] [email protected] 1 points 6 months ago (2 children)

I don't rule out the possibility of future intelligent text generators, but I don't think that large language models will do it.

I don't know enough about the evolution of cognition in human beings to believe that your second statement is wrong or right.

[–] [email protected] 1 points 6 months ago (1 children)

My second statement is certainly not comprehensive and rather offhand. Its just we have a ton of pattern matching behavior thats been optimized. Sometimes you may read things about it in why we see faces in everything and why we like certain sounds such that we have music and such.

[–] [email protected] 1 points 6 months ago (1 children)

Got it - that's a fair view then.

[–] [email protected] 2 points 6 months ago

yeah its just belief thought really. Its made me think on it more and its sorta a chicken and egg thing. like the whole basic things with cells. attracted to beneficial things like food and repulsed by dangerous things. so recognition becomes important but maybe just the like dislike is the thing and not the aparatus to identify it. shrug

[–] [email protected] 1 points 6 months ago (2 children)

How do you define intelligent here, if it’s not what LLMs can do? What’s something an “intelligent” text generator can do that an LLM cannot?

[–] [email protected] 1 points 6 months ago

If you genuinely believe that the current LLMs are intelligent, odds are that you have nothing meaningful to talk about LLMs, and thus I won't waste my time with you.

[–] SpookyCoffee 4 points 6 months ago

It’s one of the videos on his channel, that I didn’t watch on purpose. It’s just depressing thinking about it.