top | item 46125076

(no title)

tenacious_tuna | 2 months ago

> Well now you get to read it.

Man, I wish this was true. I've given the same feedback on a colleague's clearly LLM-generated PRs. Initially I put effort into explaining why I was flagging the issues, now I just tag them with a sadface and my colleague replies "oh, cursor forgot." Clearly he isn't reading the PRs before they make it to me; so long as it's past lint and our test suite he just sends the PR.

I'd worry less if the LLMs weren't prone to modifying the preconditions of the test whenever they fail such that the tests get neutered, rather than correctly resolving the logic issues.

discuss

order

HaroldCindy|2 months ago

We need to develop new etiquette around submitting AI-generated code for review. Using AI for code generation is one thing, but asking other people review something that you neither wrote nor read is inconsiderate of their time.

daheza|2 months ago

I'm getting AI generated product requirements that they haven't read themselves. It is so frustrating. Random requirements like "this service must have a response time of 5s or less" - "A retry mechanism must be present". We have a specific SLA already for response time and the designs don't have a retry mechanism built.

The bad product managers have become 10x worse because they just generate AI garbage to spray at the engineering team. We are now writing AI review process for our user stories to counter the AI generation of the product team. I'd much rather spend my time building things than having AI wars between teams.

icedchai|2 months ago

At least they're running the test suite? I'm working with guys who don't even do that! I've also heard "I've fixed the tests" only to discover, yes, the tests pass now, but the behavior is no longer correct...