top | item 42901667

Deploy dedicated DeepSeek 32B on L40 GPUs ($8/hour)

19 points| wfalcon | 1 year ago |lightning.ai | reply

6 comments

order
[+] woodr77|1 year ago|reply
Everyone's saying I needed H100s for this. L40 is way easier for me to get my hands on. great news.
[+] ashenWon|1 year ago|reply
Is this running ollama, vllm or sglang under the hood? Curious about these performance numbers.
[+] lmilad|1 year ago|reply
How well does DeepSeek R1 handle generating long pieces of text with Qwen 32B?
[+] tchaton84|1 year ago|reply
Does it support largest Deepseek model ?
[+] yewnork|1 year ago|reply
curious the performance / price tradeoffs between deepseek-r1 671b, 70b, 32b
[+] neilbhatt|1 year ago|reply
nice, i can actually use my AWS start up creds