this post was submitted on 03 Nov 2023
230 points (96.7% liked)

Programming

17008 readers
331 users here now

Welcome to the main community in programming.dev! Feel free to post anything relating to programming here!

Cross posting is strongly encouraged in the instance. If you feel your post or another person's post makes sense in another community cross post into it.

Hope you enjoy the instance!

Rules

Rules

  • Follow the programming.dev instance rules
  • Keep content related to programming in some way
  • If you're posting long videos try to add in some form of tldr for those who don't want to watch videos

Wormhole

Follow the wormhole through a path of communities [email protected]



founded 1 year ago
MODERATORS
you are viewing a single comment's thread
view the rest of the comments
[–] minorninth 9 points 10 months ago (4 children)

GPT-3.5 seems to have a problem of recency bias. With long enough input it can forget its prompt or be convinced by new arguments.

GPT-4 is not immune though better.

I’ve had some luck with a post-prompt. Put the user’s input, then follow up with a final sentence reminding the model of the prompt and desired output format.

[–] [email protected] 2 points 10 months ago (3 children)

Yes, that's by design, the networks work on transcripts per input, it does genuinely get cut off eventually, usually it purges an entire older line when the tokens exceed a limit.

[–] minorninth 2 points 10 months ago (1 children)

I’m talking about using the ChatGPT API to make a chat bot. Even when the user’s input is just one sentence, it can cause ChatGPT to forget its prompt.

[–] [email protected] 1 points 10 months ago* (last edited 10 months ago)

Ah, even then it could just be a consequence of training samples usually being chronological(most often the expected resolution for conflicting instructions is "whatever you heard last", with some exceptions when explicitly stated) so it learns to think that way. I did find the pattern also applies to GPT trained on long articles where you'd expect it not to, so wanted to just explain why that might be.

load more comments (1 replies)
load more comments (1 replies)