(no title)
v_CodeSentinal | 1 month ago
The only fix is tight verification loops. You can't trust the generative step without a deterministic compilation/execution step immediately following it. The model needs to be punished/corrected by the environment, not just by the prompter.
seanmcdirmid|1 month ago
embedding-shape|1 month ago
Personally I think it's too early for this. Either you need to strictly control the code, or you need to strictly control the tests, if you let AI do both, it'll take shortcuts and misunderstandings will much easier propagate and solidify.
Personally I chose to tightly control the tests, as most tests LLMs tend to create are utter shit, and it's very obvious. You can prompt against this, but eventually they find a hole in your reasoning and figure out a way of making the tests pass while not actually exercising the code it should exercise with the tests.
IshKebab|1 month ago
I find that this is usually a pretty strong indication that the method should exist in the library!
I think there was a story here a while ago about LLMs hallucinating a feature in a product so in the end they just implemented that feature.
SubiculumCode|1 month ago
vrighter|1 month ago
te7447|1 month ago
zoho_seni|1 month ago
exitb|1 month ago
CamperBob2|1 month ago
Often, if not usually, that means the method should exist.
HPsquared|1 month ago