top | item 42901667 Deploy dedicated DeepSeek 32B on L40 GPUs ($8/hour) 19 points| wfalcon | 1 year ago |lightning.ai | reply 6 comments order hn newest [+] [-] woodr77|1 year ago|reply Everyone's saying I needed H100s for this. L40 is way easier for me to get my hands on. great news. [+] [-] ashenWon|1 year ago|reply Is this running ollama, vllm or sglang under the hood? Curious about these performance numbers. [+] [-] lmilad|1 year ago|reply How well does DeepSeek R1 handle generating long pieces of text with Qwen 32B? [+] [-] tchaton84|1 year ago|reply Does it support largest Deepseek model ? [+] [-] yewnork|1 year ago|reply curious the performance / price tradeoffs between deepseek-r1 671b, 70b, 32b [+] [-] neilbhatt|1 year ago|reply nice, i can actually use my AWS start up creds
[+] [-] woodr77|1 year ago|reply Everyone's saying I needed H100s for this. L40 is way easier for me to get my hands on. great news.
[+] [-] ashenWon|1 year ago|reply Is this running ollama, vllm or sglang under the hood? Curious about these performance numbers.
[+] [-] lmilad|1 year ago|reply How well does DeepSeek R1 handle generating long pieces of text with Qwen 32B?
[+] [-] yewnork|1 year ago|reply curious the performance / price tradeoffs between deepseek-r1 671b, 70b, 32b
[+] [-] woodr77|1 year ago|reply
[+] [-] ashenWon|1 year ago|reply
[+] [-] lmilad|1 year ago|reply
[+] [-] tchaton84|1 year ago|reply
[+] [-] yewnork|1 year ago|reply
[+] [-] neilbhatt|1 year ago|reply