this post was submitted on 06 Jul 2023
679 points (94.4% liked)
ChatGPT
8947 readers
1 users here now
Unofficial ChatGPT community to discuss anything ChatGPT
founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
view the rest of the comments
Bings version of chatgpt once said Vegito was the result of Goku and Vegeta performing the Fusion dance. That's when I knew it wasn't perfect. I tried to correct it and it said it didn't want to talk about it anymore. Talk about a diva.
Also one time, I asked it to generate a reddit AITA story where they were obviously the asshole. It started typing out "AITA for telling my sister to stop being a drama queen after her miscarriage..." before it stopped midway and, again, said it didn't want to continue this conversation any longer.
Very cool tech, but it's definitely not the end all, be all.
That’s actually fucking hilarious.
“Oh I’d probably use the meat grinder … uh I don’t walk to talk about this any more”
Bing chat seemingly has a hard filter on top that terminates the conversation if it gets too unsavory by their standards, to try and stop you from derailing it.
I was asking it (binggpt) to generate “short film scripts” for very weird situations (like a transformer that was sad because his transformed form was a 2007 Hyundai Tuscon) and it would write out the whole script, then delete it before i could read it and say that it couldn’t fulfil my request.
It knew it struck gold and actually sent the script to Michael Bay