(no title)
zyx321 | 6 months ago
I want to know if it's possible. 4GB for Linux, a bit of room for the calculations, and then you can load a 122GB model entirely into VRAM.
How would that perform in real life? Someone please benchmark it!
zyx321 | 6 months ago
I want to know if it's possible. 4GB for Linux, a bit of room for the calculations, and then you can load a 122GB model entirely into VRAM.
How would that perform in real life? Someone please benchmark it!
yencabulator|6 months ago
I have that split set at the minimum 2 GB and I'm giving the GPU a 20 GB model to process.