top | item 44769113

(no title)

om8 | 7 months ago

To have a gpu inference, you need a gpu. I have a demo that runs 8B llama on any computer with 4 gigs of ram

https://galqiwi.github.io/aqlm-rs/about.html

discuss

order

adastra22|7 months ago

Any computer with a display has a GPU.

om8|7 months ago

Sure, but integrated graphics usually lacks vram for LLM inference.