Assuming that people only share conversations they think are good, would that be bad? Isn’t that the basis of RHLF?
There are a few times on Reddit that I want to explain something that I know well. But it will be a long post.
I’ll be lazy and ask ChatGPT the question, either verify it’s correct based on what I know, ask it to verify its answer on the web - the paid version has had web search for over year - or guide it to the correct answer if I notice something is incorrect.
Then I’ll share the conversation as the answer and tell the poster to read through the entire conversation and tell them that I didn’t just naively ask ChatGPT. It will be obvious from my chat session.
Its own answers, with feedback about whether the answers seem to have worked.
Learning to predict what word will lead to a successful solution (rather than just looking like existing speech) may prove to be a richer dataset than SO originally was.
scarface_74|1 year ago
There are a few times on Reddit that I want to explain something that I know well. But it will be a long post.
I’ll be lazy and ask ChatGPT the question, either verify it’s correct based on what I know, ask it to verify its answer on the web - the paid version has had web search for over year - or guide it to the correct answer if I notice something is incorrect.
Then I’ll share the conversation as the answer and tell the poster to read through the entire conversation and tell them that I didn’t just naively ask ChatGPT. It will be obvious from my chat session.
SketchySeaBeast|1 year ago
btilly|1 year ago
Learning to predict what word will lead to a successful solution (rather than just looking like existing speech) may prove to be a richer dataset than SO originally was.
sebazzz|1 year ago
Unless the feedback from the failing code review is piped back into the model it will still repeat the same garbage.