oh absolutely. burning a coal plant to decide if i should close discord is peak 2025 energy.
strictly speaking, using the local model (Ollama) is 'free' in terms of watts since my laptop is on anyway, but yeah, if the inefficiency is the art, I'm the artist.
fragmede|1 month ago
redfloatplane|1 month ago
evilduck|1 month ago
ImPrajyoth|1 month ago
I think we are moving toward a bilayered compute model: The Cloud: For massive reasoning.
The Local Edge: A small, resilient model that lives on-device and handles the OS loop, privacy, and immediate context.
BrainKernel is my attempt to prototype that Local Edge layer. Its messy right now, but I think the OS of 2030 will definitely have a local LLM baked into the kernel.
bdhcuidbebe|2 months ago
Now that’s a cursed take on power efficency
ImPrajyoth|2 months ago