top | item 47195223

(no title)

cpburns2009 | 1 day ago

Does llama.cpp support Qwen3.5 yet? When I tried it before, it failed saying "qwen35moe" is an unsupported architecture.

discuss

order

hnfong|1 day ago

Yes, but make sure you grab the latest llama.cpp release

New model archs usually involve code changes.

sowbug|1 day ago

If you're running Ollama, you'll have to wait a little longer for its embedded version of llama.cpp to catch up. It can be a couple days or weeks behind.

cpburns2009|1 day ago

Awesome! It looks like the llama.cpp-hip AUR was updated today to b8179, and it works.

reactordev|1 day ago

You would need the Dynamic 2.0 GGUF as discussed in the article.

But mmmmmm, Q8_K_XL looks mighty nice.