(no title)
manibatra | 2 years ago
If you want a completely offline version you'd be running one of the open source LLMs locally. Otherwise put the embeddings in a VectorDB, query it for the context and send it to one of the completion APIs available (OpenAI, etc)
No comments yet.