Show HN: Playground for comparing embedding models on Wikipedia+book retrieval
5 points| davidtsong | 2 years ago |embeds.ai
A few weeks ago, Shreyan and I were looking for an embedding model to use for RAG. We eventually came across the MTEB leaderboard, but we struggled to understand the benchmark scores.
We wanted a tool to test various embedding models with example queries on real-world datasets. After unsuccessfully looking for such a “playground”, we decided to just build one ourselves!
We embedded HuggingFace’s Simple Wikipedia dataset using @OpenAI, @Cohere, and 2 open-source models via @Baseten. We then stored the embeddings in @Supabase using pgvector. Finally, we built a web app using NextJS and deployed it on @Vercel.
Now we’re hosting the playground for anyone to use for free, as well as open-sourcing our work so people can try evaluating other models, datasets, or indexes.
Learn more here in our full blog post here: https://shreyanjain.substack.com/p/announcing-embedding-batt...
And the repo is here: https://github.com/EGCap/playground
If you have other suggestions / pain points from working with embedding models, vector DBs, or RAG, or if you would like to collaborate on any of the above or unrelated projects, please reach out! @shreyanj98 @davidtsong on Twitter
varunshenoy|2 years ago
Seems weird that every RAG app uses top-k especially since you might pull in information irrelevant to the context (e.g. if you were asking for the names of the authors of paper, you probably only want the top-1 embedding).
davidtsong|2 years ago
clueless_stats|2 years ago
davidtsong|2 years ago
sr33j|2 years ago
bigheadgpt|2 years ago
davidtsong|2 years ago
We'll definitely work on adding this model next. Seems promising! Thanks for sharing.
tigs_|2 years ago
shreyanj|2 years ago
ankitd33|2 years ago
davidtsong|2 years ago