(no title)
owenpalmer | 9 days ago
Imagine a slot on your computer where you physically pop out and replace the chip with different models, sort of like a Nintendo DS.
owenpalmer | 9 days ago
Imagine a slot on your computer where you physically pop out and replace the chip with different models, sort of like a Nintendo DS.
roncesvalles|9 days ago
bagful|8 days ago
zupa-hu|8 days ago
(Still compelling!)
ekianjo|8 days ago
XorNot|9 days ago
amelius|8 days ago
With these speeds you can run it over USB2, though maybe power is limiting.
beAroundHere|9 days ago
Infact, I was thinking, if robots of future could have such slots, where they can use different models, depending on the task they're given. Like a Hardware MoE.
NitpickLawyer|8 days ago
Is this accurate? I don't know enough about hardware, but perhaps someone could clarify: how hard would it be to reverse engineer this to "leak" the model weights? Is it even possible?
There are some labs that sell access to their models (mistral, cohere, etc) without having their models open. I could see a world where more companies can do this if this turns out to be a viable way. Even to end customers, if reverse engineering is deemed impossible. You could have a device that does most of the inference locally and only "call home" when stumped (think alexa with local processing for intent detection and cloud processing for the rest, but better).
kilroy123|8 days ago
8cvor6j844qw_d6|9 days ago
sixtyj|8 days ago
Someone|8 days ago
I doubt it would scale linearly, but for home use 170 tokens/s at 2.5W would be cool; 17 tokens/s at 0,25W would be awesome.
On the other hand, this may be a step towards positronic brains (https://en.wikipedia.org/wiki/Positronic_brain)
Onavo|9 days ago