top | item 46807641

(no title)

anhner | 1 month ago

It offers a GUI for easier configuration and management of models, and it allows you to store/load models as .gguf something ollama doesn't do (it stores the models across multiple files - and yes, I know you can load a .gguf in ollama but it still makes a copy in its weird format so now I need to either have a duplicate on my drive or delete my original .gguf)

discuss

order

hnlmorg|1 month ago

Thanks for the insights. I'm not familiar with .gguf. What's the advantage of that format?

atwrk|1 month ago

.gguf is the native format of llama.cpp and is widely used for quantized models (models with reduced float accuracy to reduce memory requirements).

llama.cpp is the actual engine running the llms, ollama is a wrapper around it.