(no title)
fibonachos | 4 months ago
From that I usually get a list of file+lines to manually review, along with some initial leads to chase.
Another use case is when fixing performance issues. I can feature flag my fix and ask the model to confirm the new code path will produce the same result for a given set of inputs. We also have test coverage for this kind of thing, but the LLM can do a once-over and point out some flaws before I ever run those tests.
I haven’t gotten to the point where it writes much code for me beyond the auto-complete, which has been a modest boost in efficiency.
noduerme|4 months ago
I guess I wouldn't trust it to confirm that new code would give the same result, but it can't hurt to ask, since if it told me the code wouldn't, that would make me look more closely at it.
I think as long as you look at it as part of a distillation process, and aim for false positives, and never actually trust it, it's good at helping to surface issues you may have missed.