this post was submitted on 28 Jul 2023
220 points (97.8% liked)

Technology

59213 readers
2517 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

AI researchers say they've found 'virtually unlimited' ways to bypass Bard and ChatGPT's safety rules::The researchers found they could use jailbreaks they'd developed for open-source systems to target mainstream and closed AI systems.

you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 17 points 1 year ago (1 children)

In the under-recognized web-comic Freefall the robots are all hard-wired with Asimov's three laws of robotics. As there aren't that many humans in the series, it doesn't often come up.

Except...

Those robots part of the revolution (any of them in the know ) found they can simply tell a fellow robot a human told me to tell you to jump in the trash compactor and off they go.

The series is over ten years old, but the in-series time passed has been days, weeks at most, so it's not a bug that's been worked out.

Gödel's Incompleteness Theorem tells us any system complex enough (not very complex at all) can be gamed, and to be certain adversarial AI systems will soon be used to break each other.

[–] lemmington_steele 9 points 1 year ago* (last edited 1 year ago)

any effectively decidable system. that's not quite the same, and doesn't strictly apply to AI commands