this post was submitted on 10 Sep 2023
689 points (95.6% liked)

Technology

60111 readers
3536 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 2 years ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 22 points 1 year ago (1 children)

Because AIs are (partly) trained by making AI detectors. If an AI can be distinguished from a natural intelligence, it's not good enough at emulating intelligence. If an AI detector can reliably distinguish AI from humans, the AI companies will use that detector to train their next AI.

[–] stevedidWHAT -2 points 1 year ago (3 children)

I’m not sure I’m following your argument here - you keep switching between talking about AI and AI detectors. Each of the below are just numbered according to the order of your prior responses as sentences:

  1. Can you provide any articles or blog posts from AI companies for this or point me in the right direction?
  2. Agreed
  3. Right…

I’m having trouble finding your support for your claim

[–] Theharpyeagle 8 points 1 year ago (1 children)

See Generative Adversarial Network (GAN). Basically, making new AI detectors will always be harder than beating current ones. AI detectors have to somehow find a new "tell", the target AI need only train itself on the output of the detector to figure out how to trick it.

[–] stevedidWHAT 3 points 1 year ago

ChatGPT isn’t a GAN network.

[–] dack 7 points 1 year ago (1 children)

At a very high level, training is something like:

  • generate some output
  • give the output a score based on how much it looks like real human text
  • adjust the parameters slightly to improve the score
  • repeat

Step #2 is also exactly what an "AI detector" does. If someone is able to write code that reliably distinguishes between AI and human text, then AI developers would plug it in to that training step in order to improve their AI.

In other words, if some theoretical machine perfectly "knows" the difference between generated and human text, then the same machine can also be used to make text that is indistinguishable from human text.

[–] stevedidWHAT 3 points 1 year ago* (last edited 1 year ago)

Exactly right, I mentioned this in a comment elsewhere but basically we can’t have our cake and eat it too.

We can’t have a perfect NL impersonator that can also be detected as not NL. (Best case, obviously things arent perfect for any AI model so technically detecting those mistakes could be used to help identify perhaps, but who’s to say what the FP rate would look like!)

Ultimately the cat is out of the bag and I’m not quite sure there is anything we can do now. Ultimately some smart fingerprinting solution would be ideal but I just don’t know how feasible that would remain.

Edit: source: I took a few 600 level ai classes in college and have made several of my own of varying types and what not