top | item 38931314 (no title) ryanklee | 2 years ago It's an LLM technology that allows certain models to run on CPUs rather than big beefy GPUs. Makes running locally viable for consumers. discuss order hn newest muricula|2 years ago Is there a specific paper or something you can point me to? Or are you talking about like llama.cpp? Because I thought that referred to the fact that it was originally one c++ file named llama.cpp? ryanklee|2 years ago I assumed it was in reference to llama.cpp. It's a weak assumption, though. load replies (1)
muricula|2 years ago Is there a specific paper or something you can point me to? Or are you talking about like llama.cpp? Because I thought that referred to the fact that it was originally one c++ file named llama.cpp? ryanklee|2 years ago I assumed it was in reference to llama.cpp. It's a weak assumption, though. load replies (1)
ryanklee|2 years ago I assumed it was in reference to llama.cpp. It's a weak assumption, though. load replies (1)
muricula|2 years ago
ryanklee|2 years ago