(no title)
bobnarizes | 2 months ago
I’m using SBERT-style embedding models for the semantic matching, which works very well in practice.
For non-text content, the app also analyzes images (OCR + object recognition) using Apple’s Vision framework. That part is surprisingly powerful, especially on Apple Silicon.
> I need to do something for images that are already classified/tagged via FastVLM
What’s the concrete use case you’re targeting with this?
cpursley|2 months ago
bobnarizes|2 months ago