(no title)
imperio59 | 1 month ago
So they're talking about this becoming an issue when chunking TBs of data (I assume), not your 1kb random string...
imperio59 | 1 month ago
So they're talking about this becoming an issue when chunking TBs of data (I assume), not your 1kb random string...
groby_b|1 month ago
memchunk has a throughput of 164 GB/s. A really fast embedder can deliver maybe 16k embeddings/sec, or ~1.6GB/s (if you assume 100 char sentences)
That's two orders of magnitude difference. Chunking is not the bottleneck.
It might be an architectural issue - you stuff chunks into a MQ, and you want to have full visibility in queue size ASAP - but otherwise it doesn't matter how much you chunk, your embedder will slow you down.
It's still a neat exercise on principle, though :)
viraptor|1 month ago