top | item 43333313

(no title)

abroadwin | 11 months ago

Neat. It would be nice to provide an option to use an API endpoint without downloading an additional local model. I have several models downloaded via ollama and would prefer to use them without additional space being taken up by the default model.

discuss

order

Terretta|11 months ago

From the README:

Optionally, offload generation to speed up generation while extending the battery life of your MacBook.

Screenshot shows example, mentions OpenAI and gpt-4o.

abroadwin|11 months ago

But it still forces you to download a local model before you can use that feature.