top | item 38931314

(no title)

ryanklee | 2 years ago

It's an LLM technology that allows certain models to run on CPUs rather than big beefy GPUs. Makes running locally viable for consumers.

discuss

order

muricula|2 years ago

Is there a specific paper or something you can point me to? Or are you talking about like llama.cpp? Because I thought that referred to the fact that it was originally one c++ file named llama.cpp?

ryanklee|2 years ago

I assumed it was in reference to llama.cpp. It's a weak assumption, though.