top | item 43333313 (no title) abroadwin | 11 months ago Neat. It would be nice to provide an option to use an API endpoint without downloading an additional local model. I have several models downloaded via ollama and would prefer to use them without additional space being taken up by the default model. discuss order hn newest Terretta|11 months ago From the README:Optionally, offload generation to speed up generation while extending the battery life of your MacBook.Screenshot shows example, mentions OpenAI and gpt-4o. abroadwin|11 months ago But it still forces you to download a local model before you can use that feature.
Terretta|11 months ago From the README:Optionally, offload generation to speed up generation while extending the battery life of your MacBook.Screenshot shows example, mentions OpenAI and gpt-4o. abroadwin|11 months ago But it still forces you to download a local model before you can use that feature.
abroadwin|11 months ago But it still forces you to download a local model before you can use that feature.
Terretta|11 months ago
Optionally, offload generation to speed up generation while extending the battery life of your MacBook.
Screenshot shows example, mentions OpenAI and gpt-4o.
abroadwin|11 months ago