top | item 33737039 (no title) jkbl | 3 years ago Yeah, but that's only true when you use one model for yourself. More VRAM is needed for running such a service. It currently loads 6 models per single GPU. And I think I have some VRAM left to add even more. discuss order hn newest No comments yet.
No comments yet.