When you correct it, it always seems to say "You are absolutely right! I apologize for the mistake," and then it gives another wrong answer with the same fully confident tone.
Lol. One probably has to work at OpenAI to try and fix their LLM thing. The one they let the public use probably doesnt learn from its users. Hopefully. I wouldn't let just any data in my LLM. When those things get garbage in they just put garbage out.
There's a "thumbs up" and "thumbs down" next to each generated response.
While this model may not be getting updated in real time, I would be surprised if that feedback isn't used when looking at updating the model with good feedback being used for retraining the model.
Once context gets long with the limited tokenization memory we currently have it seems to go insane rather fast. Would like to test on the 32k model to see how the same prompting differs.
I thought I was using GPT-4, but it appears that's only available to paying customers so far? The question I was getting repeated wrong answers to was "Can you make up a palindrome that starts with 'Dude'?" If you want to try that with GPT-4 I'd be interested to see if it can do it now or at least knows to say "I can't".
jacobwilliamroy|2 years ago
shagie|2 years ago
While this model may not be getting updated in real time, I would be surprised if that feedback isn't used when looking at updating the model with good feedback being used for retraining the model.
pixl97|2 years ago
Once context gets long with the limited tokenization memory we currently have it seems to go insane rather fast. Would like to test on the 32k model to see how the same prompting differs.
mckeed|2 years ago