top | item 47195223 (no title) cpburns2009 | 1 day ago Does llama.cpp support Qwen3.5 yet? When I tried it before, it failed saying "qwen35moe" is an unsupported architecture. discuss order hn newest hnfong|1 day ago Yes, but make sure you grab the latest llama.cpp releaseNew model archs usually involve code changes. sowbug|1 day ago If you're running Ollama, you'll have to wait a little longer for its embedded version of llama.cpp to catch up. It can be a couple days or weeks behind. cpburns2009|1 day ago Awesome! It looks like the llama.cpp-hip AUR was updated today to b8179, and it works. reactordev|1 day ago You would need the Dynamic 2.0 GGUF as discussed in the article.But mmmmmm, Q8_K_XL looks mighty nice.
hnfong|1 day ago Yes, but make sure you grab the latest llama.cpp releaseNew model archs usually involve code changes. sowbug|1 day ago If you're running Ollama, you'll have to wait a little longer for its embedded version of llama.cpp to catch up. It can be a couple days or weeks behind. cpburns2009|1 day ago Awesome! It looks like the llama.cpp-hip AUR was updated today to b8179, and it works.
sowbug|1 day ago If you're running Ollama, you'll have to wait a little longer for its embedded version of llama.cpp to catch up. It can be a couple days or weeks behind.
cpburns2009|1 day ago Awesome! It looks like the llama.cpp-hip AUR was updated today to b8179, and it works.
reactordev|1 day ago You would need the Dynamic 2.0 GGUF as discussed in the article.But mmmmmm, Q8_K_XL looks mighty nice.
hnfong|1 day ago
New model archs usually involve code changes.
sowbug|1 day ago
cpburns2009|1 day ago
reactordev|1 day ago
But mmmmmm, Q8_K_XL looks mighty nice.