(no title)
pstorm | 2 years ago
I'm particularly surprised by the speed considering all of the pre and post processing. I am doing some similar things and that is one bottlenecks. I'll dig in, but I'm curious what models you are using for each of these steps.
yuhongsun|2 years ago
We're using E5 base by default but there's an embedding model admin page to choose alternatives. There's also an API for it if you know what you're doing, you can even set one of the billion+ parameter LLM bi-encoders if you want (but you'd need a GPU for sure).
pdxandi|2 years ago