top | item 46593987

Show HN: API that falls back to humans when AI is unsure

5 points| Saurabh_Kumar_ | 2 months ago |sync-ai-11fj.vercel.app | reply

Hey HN, I’m the founder of SyncAI.

Previously, I was building internal tools for a fintech startup. We tried using GPT-4 Vision and various OCR APIs to automate our Accounts Payable. They worked great for ~90% of documents.

The problem was the other 10%: crumpled receipts, handwritten delivery notes, or invoices with weird layouts. In fintech, a 90% success rate isn’t automation; it’s a liability. We spent more time fixing the AI’s hallucinations than if we had just typed it manually.

I realized that for high-stakes operations, we didn’t need "better AI"—we needed a Safety Layer.

So I built SyncAI. It’s an extraction API that guarantees "Golden Records" (verified data) by routing uncertainty to humans.

How it works under the hood:

Ingestion: You send a PDF/Image via API.

Ensemble Extraction: We run it through a mix of OCR and LLMs to extract key fields.

Confidence Scoring: We calculate a confidence score for each specific field (not just the whole doc).

The Routing Logic:

If Confidence > 99% (and passes regex validation) → Auto-Approve (Returns JSON instantly).

If Confidence < 99% → Routes to HITL (Human-in-the-Loop) queue.

Verification: A human verifies only the low-confidence fields in a split-screen UI.

Callback: Your webhook receives the strictly typed, verified JSON.

We are positioning this for developers building Autonomous Finance Agents who need deterministic inputs.

I’ve set up a playground where you can try it without signing up. Upload your messiest invoice/receipt to see if we can parse it: https://sync-ai-11fj.vercel.app/

Pricing is usage-based (per doc). The documentation is here: https://sync-ai-11fj.vercel.app/docs.

I’d love to hear your feedback on the routing architecture or how you’re currently handling "AI hallucinations" in production pipelines.

2 comments

order
[+] Saurabh_Kumar_|2 months ago|reply
Hey HN, I’m Saurabh, founder of SyncAI.

While building fintech apps previously, I realized that GPT-4 is great, but getting it to read complex, messy invoices reliably (99.9%) is a nightmare. A 5% error rate is fine for a chatbot, but fatal for Accounts Payable.

I got tired of writing RegEx wrappers and retry logic, so I built SyncAI – a 'Safety Layer' for AI Agents.

How it works technically:

We ingest the PDF and run it through a mix of OCR + LLMs.

We calculate a 'Confidence Score' for every field extracted.

If confidence > 95%, it goes straight to your webhook.

If confidence < 95%, it routes to a Human-in-the-Loop (HITL) queue where a human verifies just that specific field.

Your Agent gets a strictly typed JSON 'Golden Record'.

Tech Stack: Python/FastAPI backend, React for the review dashboard, and we use a fine-tuned model for the routing logic.

The OCR Challenge: I know you guys are skeptical (as you should be). So I built a playground where you can upload your messiest, crumpled invoice to try it out without signing up: https://sync-ai-11fj.vercel.app/

Would love your feedback on the routing logic. I’ll be here answering questions all day!

[+] kundan_s__r|2 months ago|reply
This is a very pragmatic take. The “90% accuracy is a liability” line resonates — in high-stakes systems, partial automation often costs more than it saves.

What I like here is the field-level confidence gating instead of a single document score. That maps much better to real failure modes, where one bad value (amount, date, vendor) can invalidate the whole record.

One question I’m curious about: how stable are the confidence thresholds over time? In similar systems I’ve seen, models tend to get confidently wrong under distribution shift, which makes static thresholds tricky.

Have you considered combining confidence with explicit intent or scope constraints (e.g., what the system is allowed to infer vs. must escalate), rather than confidence alone?

Overall, this feels much closer to how production systems should treat AI — not as an oracle, but as a component that earns trust incrementally.