Does this produce actual lint rules, or are you templating out lint-like replies from a LLM using a response format?
If you're doing inference, just give me a cli that's userless and free. I'm happy to use left over codex plan tokens or gemini free tokens for this, and while the idea seems interesting and I might be upsellable to more features down the line, the price/offering is a non starter.
We combine determinism + LLMs to catch things a human would normally have to. If the LLM finds a violation, it generates a comment.
Big agree on the CLI being open and letting you bring your own inference provider. We’re holding off on it until we get more feedback from some of our hardcore users.
CuriouslyC|4 months ago
If you're doing inference, just give me a cli that's userless and free. I'm happy to use left over codex plan tokens or gemini free tokens for this, and while the idea seems interesting and I might be upsellable to more features down the line, the price/offering is a non starter.
dearilos|4 months ago
Big agree on the CLI being open and letting you bring your own inference provider. We’re holding off on it until we get more feedback from some of our hardcore users.