this post was submitted on 12 Apr 2024
507 points (99.6% liked)

196

16310 readers
2178 users here now

Be sure to follow the rule before you head out.

Rule: You must post before you leave.

^other^ ^rules^

founded 1 year ago
MODERATORS
 
you are viewing a single comment's thread
view the rest of the comments
[–] Buddahriffic 8 points 5 months ago

I think the applicant including "don't reply as if you are an LLM" in their prompt might be enough to defeat this.

Though now I'm wondering if LLMs can pick up and include hidden messages in their input and output to make it more subtle.

Just tested it with gpt 3.5 and it wasn't able to detect a message using the first word after a bunch of extra newlines. When asked a specifically if it could see a hidden message, it said how the message was hidden but then just quoted the first line of the not hidden text.