Show HN: I built a tool to assist AI agents to know when a PR is good to go
45 points| dsifry | 1 month ago |dsifry.github.io
It would poll CI in loops. Miss actionable comments buried among 15 CodeRabbit suggestions. Or declare victory while threads were still unresolved.
The core problem: no deterministic way for an agent to know a PR is ready to merge.
So I built gtg (Good To Go). One command, one answer:
$ gtg 123 OK PR #123: READY CI: success (5/5 passed) Threads: 3/3 resolved
It aggregates CI status, classifies review comments (actionable vs. noise), and tracks thread resolution. Returns JSON for agents or human-readable text.
The comment classification is the interesting part — it understands CodeRabbit severity markers, Greptile patterns, Claude's blocking/approval language. "Critical: SQL injection" gets flagged; "Nice refactor!" doesn't.
MIT licensed, pure Python. I use this daily in a larger agent orchestration system — would love feedback from others building similar workflows.
rootnod3|1 month ago
So the agent can now merge shit by itself?
Just the let damn thing push nto prod by itself at this point.
danenania|1 month ago
If there are CI failures or obvious issues that another AI can identify, why not have the agent keep going until those are resolved? This tool just makes that process more token efficient. Seems pretty useful to me.
ljm|1 month ago
blutoot|1 month ago
The reality is that probably 99.9999% of code bases on this earth (but this might drop soon, who knows) pre-date LLMs and organizing them in a way that coding agents can produce consistent results from sprint to sprint, will need a big plumbing work from all dev teams. And that will include refactoring, documentation improvements, building consensus on architectures and of course reshaping the testing landscape. So SWE's will have a lot of dirty work to do before we reach the aforementioned "scale".
However, a lot of platforms are being built from ground-up today in a post-CC (claude code) era . And they should be ready to hit that scale today.
glemion43|1 month ago
And yes there are plenty of use cases were ai code doesn't hurt anyone even if it gets merged automatically...
See it as an interesting new field of r&d...
literalAardvark|1 month ago
And you can do a more exhaustive test later, after the agents are done running amok to merge various things.
squeaky-clean|1 month ago
tayo42|1 month ago
The linked page explains how this fits into a development workflow
eg.
> A reviewer wrote “consider using X”… is that blocking or just a thought?
> AMBIGUOUS - Needs human judgment (suggestions, questions)
baxtr|1 month ago
dsifry|1 month ago
philipp-gayret|1 month ago
On a personal note; I hate LLM output to advertise a project. If you have something to share have the decency to type it out yourself or at least redact the nonsense from it.
dsifry|1 month ago
joshribakoff|1 month ago
furyofantares|1 month ago
forgotpwd16|1 month ago
aaronbrethorst|1 month ago
nyc1983|1 month ago
dsifry|1 month ago
Agent_Builder|1 month ago
[deleted]
mcolley|1 month ago
dsifry|1 month ago
So gtg builds all of that in and deterministically determines whether or not there are any actionable comments, and thus you can block the agent from moving forward until all actionable comments are thoroughly reviewed, acted upon or acknowledged, at which point it will change state and allow the PR to be merged.
joshuanapoli|1 month ago
dsifry|1 month ago