Launch HN: Parachute (YC S25) – Guardrails for Clinical AI
62 points| ariavikram | 6 months ago
Hospitals are racing to adopt AI. More than 2,000 clinical AI tools hit the U.S. market last year - from ambient scribes to imaging models. But new regulations (HTI-1, Colorado AI Act, California SB 3030, White House AI Action Plan) require auditable proof that these models are safe, fair, and continuously monitored.
The problem is, most hospital IT teams can’t keep up. They can’t vet every vendor, run stress tests, and monitor models 24/7. As a result, promising tools die in pilot hell while risk exposure grows.
We saw this firsthand while deploying AI at Columbia University Irving Medical Center, so we built Parachute. Columbia is now using it to track live AI models in production.
How it works: First, Parachute evaluates vendors against a hospital’s clinical needs and flags compliance and security risks before a pilot even begins. Next, we run automated benchmarking and red-teaming to stress test each model and uncover risks like hallucinations, bias, or safety gaps.
Once a model is deployed, Parachute continuously monitors its accuracy, drift, bias, and uptime, sending alerts the moment thresholds are breached. Finally, every approval, test, and runtime change is sealed into an immutable audit trail that hospitals can hand directly to regulators and auditors.
We’d love to hear from anyone with hospital experience who has an interest in deploying AI safely. We look forward to your comments!
jph|6 months ago
padolsey|6 months ago
We're looking for domain experts especially in high risk domains like healthcare, education, therapy. Then we'd work together co-authoring an eval in your specialism to expose and motivate AI labs to do better.
robertlagrant|6 months ago
ariavikram|6 months ago
pizzathyme|6 months ago
Next up is just great execution by you all!
That list of logos you all have - are those paying customers today?
Best of luck!
creata|6 months ago
Doesn't look like it. The first list of logos is standards bodies. The second list of logos is integrations.
potatoman22|6 months ago
ariavikram|6 months ago
jstummbillig|6 months ago
For example, consider what happens in this video: https://www.youtube.com/watch?v=AZhCYisIQB8&t=2s
Please don't make this mistake of thinking "aha, but you see, a human intervened!" This will never happen in the real world for the vast majority of humans in a similar scenario.
padolsey|6 months ago
Usually you can run human-in-the-loop spot checks to ensure that there's parity between your LLM evaluators and the equivalent specialist human evaluator.
siva7|6 months ago
znxnnxnx|6 months ago
seriusam|6 months ago
ariavikram|6 months ago
We use in-house evals (based on existing state-of-the-art benchmarks) to compare ambient scribes.
If you take a deeper look into the companies on our landing page, you will see that the first list refers to the compliance standards our workflows follow and the second refers to the existing tools we integrate with.
richwater|6 months ago
Impossible to deliver
potatoman22|6 months ago
Here's a good overview of fairness: https://learn.microsoft.com/en-us/azure/machine-learning/con... and there's plenty of papers discussing how to safely use predictive analytics and AI in healthcare.
I don't know if this product can give proof for safe and fair ML systems, but it's not impossible to use these things safely and fairly.
ariavikram|6 months ago
padolsey|6 months ago
sgt|6 months ago
zmmmmm|6 months ago
this is humans? I'm really not sure how this could be automated given the vast spectrum of applications and specific requirements complex organisations like hospitals have. It would have to boil down to "check box" compliance style analysis which in my experience usually leads to poor outcomes down the track (the worst product from every other point of view gets chosen because it checks the most arbitrary boxes on the security / compliance forms - then the integration bill dwarfs whatever it would have cost to address most of those things bespoke anyway).
fehudakjf|6 months ago
We've all seen how powerful language can be in legal defenses surrounding the for profit healthcare industry of the united states.
What new "pre-existing conditions" alike thought, and legal argument, terminating phrases will these large language models come up with for future generations?
nradov|6 months ago
iamgopal|6 months ago
j4coh|6 months ago
ariavikram|6 months ago
cactca|6 months ago
Here are a few questions that should be part of an evaluation of the Parachute platform to pressure test the claims made on the website and this post: 1) How many Parachute customers have passed regulatory audits by CMS, OCR, CLIA/CLAP, and the FDA? 2) What high quality peer-reviewed scientific evidence supports the claims of increased safety and detection of hallucinations and bias? 3) What liability does Parachute assume during production deployment? What are the SLAs? 4) How many years of regulatory experience does the team have with HIPPA, ISO, CFR, FDA, CMS, and state medical board compliance?
tony-yamin|6 months ago