nkaretnikov | 2 years ago | on: Unveiling Ragna: An Open Source RAG-Based AI Orchestration Framework
nkaretnikov's comments
nkaretnikov | 2 years ago | on: Unveiling Ragna: An Open Source RAG-Based AI Orchestration Framework
nkaretnikov | 2 years ago | on: Unveiling Ragna: An Open Source RAG-Based AI Orchestration Framework
I assume it's the former. It's explained in the post after the diagram, see "The diagram illustrates a series of steps [...]".
We get a user query. Based on that query, we pull relevant parts from the doc. Then, we submit both the query and the doc parts to an LLM. This limits the amount of data we need to send and allows the user to know which parts of the doc are relevant.
nkaretnikov | 2 years ago | on: Unveiling Ragna: An Open Source RAG-Based AI Orchestration Framework
In the blogpost, there's a demo: you upload a pdf, ask ML model a question, it answers and shows which pages it looked at in the document. If you don't trust the output, you can double-check the doc since you have the page numbers.
Ragna currently ships with pre-built extensions for OpenAI, MosaicML, Anthropic, and local LLMs as well as the Chroma and LanceDB vector databases. And it's open source, too!
nkaretnikov | 2 years ago | on: PyTorch: The Most Complex Software Project in the World Keeps Getting Better