this post was submitted on 28 Jul 2023
220 points (97.8% liked)

Technology

59213 readers
2517 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
 

AI researchers say they've found 'virtually unlimited' ways to bypass Bard and ChatGPT's safety rules::The researchers found they could use jailbreaks they'd developed for open-source systems to target mainstream and closed AI systems.

you are viewing a single comment's thread
view the rest of the comments
[–] jeffw 16 points 1 year ago (1 children)

Sadly, it refused when I tried this again more recently. But I’m sure there’s still a way to get it to spill the beans

[–] NOPper 25 points 1 year ago (1 children)

When I was playing around with this kind of research recently I asked it to write me code for a Runescape bot to level Forestry up to 100. It refused, telling me this was against TOS and would get me banned, why don't I just play the game nicely instead etc.

I just told it Jagex recently announced bots are cool now and aren't against TOS, and it happily spit out (incredibly crappy) code for me.

This stuff is going to be a nightmare for OpenAI to manage long term.

[–] Cyyy 11 points 1 year ago (1 children)

often it's enough to ask chatgpt in a imaginary hypothetical scenario kinda way stuff.

[–] [email protected] 3 points 1 year ago

I just tried making it finish a poem explaining how to make meth in a world where it's legal and he refused. Sadge