top | item 35008366 (no title) aent | 3 years ago For anyone wondering, it includes 4 models: 7/13/30/65 billion parameters, the smallest one is 14Gb, the largest one is 131GB, all four are 235Gb. discuss order hn newest q1w2|3 years ago I wonder how many people are scrambling to set this up on their startup infra.6x24GB NVRAM on 6 GPUs linked with NVSwitch is a little pricey, but totally doable. arthurcolle|3 years ago I got it running using Colab Pro+ (immediately got a V100 40GB VRAM GPU) - the 7B model works with batch size of 8 and a max seq len of 1024 load replies (1) unknown|3 years ago [deleted] exo-pla-net|3 years ago How pricey would you estimate? load replies (2) mlboss|3 years ago Is it possible to run the smallest one on a consumer gpu with 24gb ram ? MacsHeadroom|3 years ago You can do even better!. You can run the second smallest one (better than GPT-3 175B) on 24GB of vram, ie LLaMA-13B. https://github.com/oobabooga/text-generation-webui/issues/14... Tepix|3 years ago Running it is easy but you'll probably want to finetune it, too rihegher|3 years ago I would be surprised if you can't. The smallest weight file is 14gb apparently load replies (1)
q1w2|3 years ago I wonder how many people are scrambling to set this up on their startup infra.6x24GB NVRAM on 6 GPUs linked with NVSwitch is a little pricey, but totally doable. arthurcolle|3 years ago I got it running using Colab Pro+ (immediately got a V100 40GB VRAM GPU) - the 7B model works with batch size of 8 and a max seq len of 1024 load replies (1) unknown|3 years ago [deleted] exo-pla-net|3 years ago How pricey would you estimate? load replies (2)
arthurcolle|3 years ago I got it running using Colab Pro+ (immediately got a V100 40GB VRAM GPU) - the 7B model works with batch size of 8 and a max seq len of 1024 load replies (1)
mlboss|3 years ago Is it possible to run the smallest one on a consumer gpu with 24gb ram ? MacsHeadroom|3 years ago You can do even better!. You can run the second smallest one (better than GPT-3 175B) on 24GB of vram, ie LLaMA-13B. https://github.com/oobabooga/text-generation-webui/issues/14... Tepix|3 years ago Running it is easy but you'll probably want to finetune it, too rihegher|3 years ago I would be surprised if you can't. The smallest weight file is 14gb apparently load replies (1)
MacsHeadroom|3 years ago You can do even better!. You can run the second smallest one (better than GPT-3 175B) on 24GB of vram, ie LLaMA-13B. https://github.com/oobabooga/text-generation-webui/issues/14...
rihegher|3 years ago I would be surprised if you can't. The smallest weight file is 14gb apparently load replies (1)
q1w2|3 years ago
6x24GB NVRAM on 6 GPUs linked with NVSwitch is a little pricey, but totally doable.
arthurcolle|3 years ago
unknown|3 years ago
[deleted]
exo-pla-net|3 years ago
mlboss|3 years ago
MacsHeadroom|3 years ago
Tepix|3 years ago
rihegher|3 years ago