this post was submitted on 23 Jul 2023
309 points (95.0% liked)

Technology

59662 readers
3035 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

I keep seeing posts about this kind of thing getting people's hopes up, so let's address this myth.

What's an "AI detector"?

We're talking about these tools that advertise the ability to accurately detect things like deep-fake videos or text generated by LLMs (like ChatGPT), etc. We are NOT talking about voluntary watermarking that companies like OpenAI might choose to add in the future.

What does "effective" mean?

I mean something with high levels of accuracy, both highly sensitive (low false negatives) and highly specific (low false positives). High would probably be at least 95%, though this is ultimately subjective.

Why should the accuracy bar be so high? Isn't anything better than a coin flip good enough?

If you're going to definitively label something as "fake" or "real", you better be damn sure about it, because the consequences for being wrong with that label are even worse than having no label at all. You're either telling people that they should trust a fake that they might have been skeptical about otherwise, or you're slandering something real. In both cases you're spreading misinformation which is worse than if you had just said "I'm not sure".

Why can't a good AI detector be built?

To understand this part you need to understand a little bit about how these neural networks are created in the first place. Generative Adversarial Networks (GANs) are a strategy often employed to train models that generate content. These work by having two different neural networks, one that generates content similar to existing content, and one that detects the difference between generated content and the existing content. These networks learn in tandem, each time one network gets better the other one also gets better.

That this means is that building a content generator and a fake content detector are effectively two different sides of the same coin. Improvements to one can always be translated directly and in an automated way into improvements into the other one. This means that the generator will always improve until the detector is fooled about 50% of the time.

Note that not all of these models are always trained in exactly this way, but the point is that anything CAN be trained this way, so even if a GAN wasn't originally used, any kind of improved detection can always be directly translated into improved generation to beat that detection. This isn't just any ordinary "arms race", because the turn around time here is so fast there won't be any chance of being ahead of the curve... the generators will always win.

Why do these "AI detectors" keep getting advertised if they don't work?

  1. People are afraid of being saturated by fake content, and the media is taking advantage of that fear to sell snake oil
  2. Every generator network comes with its own free detector network that doesn't really work all that well (~50% accuracy) because it was used to create the generator originally, so these detectors are ubiquitous among AI labs. That means the people that own the detectors are the SAME PEOPLE that created the problem in the first place, and they want to make sure you come back to them for the solution as well.
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 54 points 1 year ago (4 children)

Good summary of the issues. I've been fairly disappointed with what a lot of people think the AI text generators are good for - replacement for search engines, magic oracle that can tell you any fact, something to write legal briefs. And the people who generate documents and then don't even proof read or fact checking them before using them for something important... Some uses are good, like basic code generation for programming tasks, but many are just silly.

The instances where some professor with zero clue about how AI text generation works or the issues you outline here has told a student "My AI detector said this was generated!" have been absurd, like one professor with obvious serious misunderstandings told a student "I asked ChatGPT if it wrote this and it said yes."

[–] Crackhappy 9 points 1 year ago (4 children)

Not to mention that this "AI" is in no way actually AI. It's just ML taken to a new level.

[–] [email protected] 19 points 1 year ago

It's not an AGI, but it's still AI

[–] deong 9 points 1 year ago

There’s no real distinction between the two. We don’t have a definition of AI or intelligence — never have. Inside the field, ML has some recognized connotations, but outside of specialist literature, they’re just marketing fluff.

[–] [email protected] 7 points 1 year ago (1 children)

It's interesting that it started a conversation about "if this thing can make output exactly like a human, does it matter?" but I agree... it's not conscious or 'thinking' about what it says. The output sure can be convincing, though.

[–] FlyingSquid 2 points 1 year ago (1 children)

I think a huge way that it matters is that it doesn't ask questions.

[–] [email protected] 2 points 1 year ago

That's a very good point. Even Eliza asked questions (and the last thing we need now is a ChatGPT therapist mode). It's also a matter of what it's programmed to do, but I don't believe that the system has awareness or curiosity.

[–] Crackhappy 3 points 1 year ago

There is a fundamental difference between recombinant regurgitation and creation.

[–] Aux 6 points 1 year ago (2 children)

The biggest issue with publicly available ML based text tools is that they're American centric. Detection of ChatGPT in the UK is simple - it creates texts using American spelling. And if you live outside of English speaking world, like most humans do, it's completely useless.

[–] [email protected] 7 points 1 year ago (2 children)

ChatGPT speaks other languages. It's actually a really good translator.

I just asked it to describe an organization using UK English and it indeed used 'organisation' instead (didn't check for other words).

[–] [email protected] 3 points 1 year ago (3 children)

Can it understand and create new compound words (in a language like German)? That's an issue I have with most spell checks and translators as well, it's forcing the language to be more like English

[–] [email protected] 6 points 1 year ago

Erfinde ein Rezept für Mohrrübensaftdressing [carrot juice dressing].

Rezept für Mohrrübensaftdressing: […] Beginne damit, frischen Mohrrübensaft herzustellen

I’d say it works.

[–] [email protected] 4 points 1 year ago (1 children)

Looks like it does: https://chat.openai.com/share/1b487711-c1be-468a-877b-98091449b55e

I asked it to translate 'meeting agreements' to Dutch and it came up with the word 'bijeenkomstafspraken', which is a valid but very uncommon Dutch word (I'm native Dutch and don't think I've seen it before). If I throw it into google with quotes around it, the first page is results with 'bijeenkomst afspraken', where 'afspraken' is used as the past tense of 'afspreken' (to agree) instead of as its noun (agreements).

It btw also suggested 'vergaderafspraken' as a translation, which is a way more common word.

[–] [email protected] 1 points 1 year ago (1 children)

That's nice, thanks for checking. I thought ChatGPT only worked at the level of whole words but it seems it chops them up internally.

[–] [email protected] 5 points 1 year ago

Correct, it's not just regurgitating words, it's predicting which token comes next. A token is sometimes a whole word, but for longer ones it's part of a word (and some other rules that define how tokenization works).

How it knows which token comes next is why the current generation of LLMs is so impressive. It seems to have learned the rules the underpin our languages, to the point that it seems to even understand the content. It doesn't just know the grammer rules (without anyone telling it, it just learned the patterns), it also knows which words belong to each other in which context.

It's your prompt + some preset other context (e.g. that it is an OpenAI LLM) that creates that context. So being able to predict a token correctly is one part, the other is having a good context. This is why prompt engineering quickly became a thing. This is also why supporting bigger contexts is another thing (but a larger context requires way more processing power, so there's a trade-off there).

It's btw not just the trained model + context that gives you the output of ChatGPT. I'm pretty sure there are layers before and after, possibly using other ML models, that filter content or make it more fit for processing. This is why you can't ask it how to make bombs, even though those recipes are in its training set and it very likely can create a recipe based on that.

[–] [email protected] 3 points 1 year ago (2 children)

I just ran a quick experiment, asking it to create compound German words. Easy. But maybe it just memorized them? So I asked it to create double-compound words (consisting of 2*2 = 4 words), and it delivered:

Natürlich! Hier sind einige zusammengesetzte Wörter aus zusammengesetzten Wörtern:

  1. Sonnenblumenfeldweg
  2. Wassermelonensaftglas
  3. Handschuhfachdeckel
  4. Zahnarztpraxisrezeption
  5. Flugzeugsitzplatzreservierung
  6. Straßenbahnhaltestellenschild
  7. Regenschirmgriffpolster
  8. Apfelbaumpflanzloch
  9. Fernseherfernbedienungsknopf
  10. Hausaufgabenerledigungszeit

Diese Wörter mögen etwas ungewöhnlich sein, aber sie folgen den Regeln der deutschen Sprache und sind zusammengesetzt aus jeweils vier Wörtern.

They all seem legit. Most are rare. I think this was the first time I read "Wassermelonensaftglas" as a native German reader. Sadly, the next iteration (compound words made from compound words made from compound words) failed, it just added one more word.

[–] [email protected] 3 points 1 year ago (1 children)

@[email protected] Was bedeuten diese Wörter?

  1. Sonnenblumenfeldweg
  2. Wassermelonensaftglas
  3. Handschuhfachdeckel
  4. Zahnarztpraxisrezeption
  5. Flugzeugsitzplatzreservierung
  6. Straßenbahnhaltestellenschild
  7. Regenschirmgriffpolster
  8. Apfelbaumpflanzloch
  9. Fernseherfernbedienungsknopf
  10. Hausaufgabenerledigungszeit
  11. Wassermelonensaftweg
[–] [email protected] 2 points 1 year ago (2 children)

Sicher! Hier sind die Bedeutungen:

  1. Sonnenblumenfeldweg: Ein Weg, der durch ein Feld von Sonnenblumen führt.
  2. Wassermelonensaftglas: Ein Glas, das mit Wassermelonensaft gefüllt ist.
  3. Handschuhfachdeckel: Der Deckel eines Handschuhfachs, normalerweise in einem Fahrzeug gefunden.
  4. Zahnarztpraxisrezeption: Der Empfangsbereich einer Zahnarztpraxis.
  5. Flugzeugsitzplatzreservierung: Der Vorgang oder die Bestätigung der Reservierung eines Sitzplatzes in einem Flugzeug.
  6. Straßenbahnhaltestellenschild: Ein Schild, das eine Straßenbahnhaltestelle kennzeichnet.
  7. Regenschirmgriffpolster: Ein Polster am Griff eines Regenschirms für zusätzlichen Komfort.
  8. Apfelbaumpflanzloch: Ein Loch, das zum Pflanzen eines Apfelbaums ausgegraben wurde.
  9. Fernseherfernbedienungsknopf: Ein Knopf auf einer Fernbedienung, um einen Fernseher zu steuern.
  10. Hausaufgabenerledigungszeit: Die Zeit, die für die Erledigung von Hausaufgaben vorgesehen ist.
  11. Wassermelonensaftweg: Könnte auf einen Weg hinweisen, in dessen Verlauf Wassermelonensaft verkauft oder verzehrt wird, aber dies ist eher ein Zusammenbau von Wörtern und hat keine verbreitete oder standardmäßige Bedeutung.
[–] [email protected] 2 points 1 year ago

Neat, that's a pretty reasonable definition for all of them, even for #11 (which I'm pretty sure isn't in its dataset, unlike the others which all have some matches on Google)

[–] hemko 1 points 1 year ago

@chatgpt translate above message to English

[–] FlyingSquid 1 points 1 year ago (1 children)

Of course when one of these grand mountain ranges goes stretching across the printed page, it adorns and ennobles that literary landscape--but at the same time it is a great distress to the new student, for it blocks up his way; he cannot crawl under it, or climb over it, or tunnel through it. So he resorts to the dictionary for help, but there is no help there. The dictionary must draw the line somewhere--so it leaves this sort of words out. And it is right, because these long things are hardly legitimate words, but are rather combinations of words, and the inventor of them ought to have been killed. They are compound words with the hyphens left out. The various words used in building them are in the dictionary, but in a very scattered condition; so you can hunt the materials out, one by one, and get at the meaning at last, but it is a tedious and harassing business. I have tried this process upon some of the above examples. "Freundshaftsbezeigungen" seems to be "Friendship demonstrations," which is only a foolish and clumsy way of saying "demonstrations of friendship." "Unabhaengigkeitserklaerungen" seems to be "Independencedeclarations," which is no improvement upon "Declarations of Independence," so far as I can see. "Generalstaatsverordnetenversammlungen" seems to be "General-statesrepresentativesmeetings," as nearly as I can get at it--a mere rhythmical, gushy euphuism for "meetings of the legislature," I judge. We used to have a good deal of this sort of crime in our literature, but it has gone out now. We used to speak of a things as a "never-to-be-forgotten" circumstance, instead of cramping it into the simple and sufficient word "memorable" and then going calmly about our business as if nothing had happened. In those days we were not content to embalm the thing and bury it decently, we wanted to build a monument over it.

-- Mark Twain, A Tramp Abroad

[–] [email protected] 1 points 1 year ago (1 children)

Okay, interesting. Of course it would be nice if languages were easy to understand and easy to learn. German seems to be on the hard end of this spectrum, but no language is free from unecessary complications like these. They all grew historically and organically, and were not constructed with accessibility in mind.

It is nearly impossible to get an objective view on languages, since each of us is inherently biased, and most of us don't speak another language so well that we could truly judge it. It's easy to spot silly things in other languages while we may be unaware of how difficult our mother language is to learn for foreigners.

The interpretation of the given examples feels wrong for me. While the technical part is correct, I think the conclusion is incorrect. For example, "Unabhaengigkeitserklärung" emphasizes the independence, while "Erklärung von Unabhängigkeit" emphasizes the demonstration. The two are not equivalent. Twain seemed to be ignorant about that and simply assumed a foreign language would follow the same rules as his own.

While I can understand Twains frustration in learning another language, his critique is based on a lack of understanding.

For some compound words, there is no straightforward equivalent. "Apfelbaum" (apple tree) could be "Baum, an dem Äpfel wachsen" (tree on which apples grow). But that leaves the question wether it's still an Apfelbaum when it does not grow apples in this moment, like in winter. "Baum des Apfels" (tree of the apple) can refer to a miniature tree on an apple. "Baum der Äpfel" (tree of the apples) might be okay.

Further, what he believes to be superior can sometimes be inferior. Consider cases like "The presentation on renewable energy technology investors." In this sentence, it's not clear whether "renewable energy technology" is a single entity modifying "investors," or if "renewable energy" and "technology investors" are separate entities, both modifying "presentation." The sentence could refer to a presentation for investors interested in renewable energy technology or to a presentation about investors who focus on renewable energy projects. Compound words prevent ambiguities like these.

Hyphens can help in these cases. They can also be used in German to make it easy to identify compound components, like it's required in https://en.wikipedia.org/wiki/Leichte_Sprache.

We used to speak of a things as a “never-to-be-forgotten” circumstance, instead of cramping it into the simple and sufficient word “memorable”

That's another interesting point to discuss. Which is easier for foreigners? Sure, a single, short word in itself is easy to learn. But it is a new word, which has to be learned. In this case, you have to learn which part of "memory" or "memorize" can be used, and which part must be replaced.

I also don't think "memorable" has the same meaning as "never-to-be-forgotten". Isn't "memorable" more fitting for positive things, while n-t-b-f is well suited for negative things? Was the Holocaust 'memorable'?

[–] FlyingSquid 2 points 1 year ago

Honestly, I posted it more because I thought it was funny than anything. I didn't expect such a deconstruction, but it's interesting!

[–] Aux 2 points 1 year ago

Well, I haven't used it for a while and according to comments below it changed a lot. So I stand corrected.

[–] [email protected] 1 points 1 year ago

So far, yes, only because they’ve been developed in the US and therefore trained on US English text. Eventually someone can make models for other languages and regions, but it is a lot of work and very expensive.

[–] [email protected] 3 points 1 year ago

I think Bing did a pretty good job at coming up with name suggestions for some Sims characters. Playing with a virtual doll house is in the more harmless end of the spectrum, but obviously people want to try LLMs with all sorts of tasks, where the stakes are much higher and consequences could be severe.

The more you use it, the more you’ll begin to understand how much you can or cannot trust an LLM. A sensible person would become more suspicious of the results, but people don’t always make sensible decisions.