(no title)
headlessvictim2 | 4 years ago
How would this work with GPU-bound machine learning models?
The model processing takes > 30 seconds and would still represent the bottleneck?
headlessvictim2 | 4 years ago
How would this work with GPU-bound machine learning models?
The model processing takes > 30 seconds and would still represent the bottleneck?
pjgalbraith|4 years ago
Here is an example of what it could look like: https://docs.microsoft.com/en-us/azure/architecture/patterns...
headlessvictim2|4 years ago
Right now, we use ELB (Elastic Load Balancer) to sit in front of multiple GPU instances.
Is this sufficient or do you suggest adding Celery into this architecture?