this post was submitted on 20 Jul 2023
65 points (91.1% liked)

Technology

58100 readers
3184 users here now

This is a most excellent place for technology news and articles.


Our Rules


  1. Follow the lemmy.world rules.
  2. Only tech related content.
  3. Be excellent to each another!
  4. Mod approved content bots can post up to 10 articles per day.
  5. Threads asking for personal tech support may be deleted.
  6. Politics threads may be removed.
  7. No memes allowed as posts, OK to post as comments.
  8. Only approved bots from the list below, to ask if your bot can be added please contact us.
  9. Check for duplicates before posting, duplicates may be removed

Approved Bots


founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] [email protected] 4 points 1 year ago (1 children)

Yep, definitely. I have a plus subscription, and stuff that was easy for it just a few months ago now seems to take several back-and-forths to barely approach similar results.

Science content is where I noticed the most degradation. It just stares at me using blank “it’s not in my training data” answers to questions that used to have comprehensive responses a while ago.

I think they’re scaling down the models to make them cheaper to run?

[–] [email protected] 6 points 1 year ago

They’re definitely reducing model performance to speed up responses. ChatGPT was at its best when it took forever to write out a response. Lately I’ve noticed that ChatGPT will quickly forget information you just told it, ignore requests, hallucinate randomly, and has a myriad of other problems I didn’t have when the GPT-4 model was released.