(no title)
tezza | 4 months ago
I worry that the garbage at the end will become part of the memory.
How many of your chats do you end… “that was rubbish/incorrect, i’m starting a new chat!”
tezza | 4 months ago
I worry that the garbage at the end will become part of the memory.
How many of your chats do you end… “that was rubbish/incorrect, i’m starting a new chat!”
rwhitman|4 months ago
GPT stores the incomplete chat and treats it as truth in memory. And it's very difficult to get it to un-learn something that's wrong. You have to layer new context on top of the bad information and it can sometimes run with the wrong knowledge even when corrected.
withinboredom|4 months ago
kromem|4 months ago
That injection (for various reasons) will essentially eat up a massive amount of the model's attention budget and most of the extended thinking trace if present.
I haven't really seen lower quality of responses with modern Claudes with long context for the models themselves, but in the web/app with the LCR injections the conversation goes to shit very quickly.
And yeah, LCRs becoming part of the memory is one (of several) things that's probably going to bite Anthropic in the ass with the implementation here.