top | item 47210534

Show HN: AI Sees Me – CLIP running in the browser

1 points| jayyvk | 1 hour ago |howaiseesme.com

I built a tool that runs OpenAI's CLIP model entirely in your browser using Transformers.js and ONNX Runtime Web. It encodes your webcam feed into vector embeddings and compares them against any text you type in real-time. No server, no API calls — all inference happens locally. The interesting technical challenge was getting CLIP to run at usable speeds in WASM while processing live video frames. Wanted to make the concept of embeddings and similarity scores tangible rather than abstract. Github: https://github.com/jayyvk/howaiseesme

discuss

order

No comments yet.