(no title)
janchorowski | 1 year ago
The approach builds on the ability of LLMs to know when they don’t know how to answer. With proper LLM confidence calibration, the adaptive RAG is as accurate as a large context-based RAG, while being much cheaper to run.
What was really interesting for us here is that the basic idea is "geometric doubling" but it needs to be put into place with so much care, because of the counter-intuitive correlation effects of mistakes produced by LLM's for different prompts.
We provide runnable code examples, you will also find a reference implementation of the strategy in the Pathway LLM expansion pack:
https://github.com/pathwaycom/pathway/blob/main/python/pathw...
No comments yet.