Show HN: Zero-power photonic language model–code
18 points| damir00 | 3 months ago |zenodo.org
Despite using only unitary operations and no attention mechanism, a 1024×32 model achieves coherent TinyStories generation after < 1.8 hours of training on a single consumer GPU.
This is Part 1 - the next step is physical implementation with $50 of optics from AliExpress.
tliltocatl|3 months ago
IronyMan100|3 months ago
cpldcpu|3 months ago
damir00|3 months ago
I apologize for not being clearer.
The goal isn't actually "zero power" - the goal is "so little heat dissipation in orbit is easy".
bastawhiz|3 months ago
If it does work, I think one of the biggest challenges will be adding enough complexity to it for it to do real, useful computation. Running the equivalent of GPT-2 is a cool tech demo, but if there's not an obvious path to scaling it up, it's a bit of a dead end.
damir00|3 months ago
I expect to have an answer this week...
ifuknowuknow|3 months ago