Classical ML Still Relevant?
17 points| Sanej | 3 years ago
With all the proliferation of DL and LLM along with near unlimited compute, energy and bandwidth do we still need classical ML approach for solving the problems? Is DL / NN going to take over everything?
apohn|3 years ago
If Domain Expertise + Feature Engineering + ML model can get you 90% of the way there and it runs on a tiny cloud instance that takes 30 minutes to train, is a DL based approach that pushes you to 91% worth it from an ROI instance if takes a 4xGPU cluster 2 days to train it, not to mention inference costs? Especially if you need to explain what the model is doing?"
This above is exactly the situation I'm in now with my job. I'm on the "Get useful stuff to production so we can save money" side of things, and we have R&D teams who try to approach the same problems using DL and all the latest methods. At least for the use cases our team focuses on, they haven't been able to do more than set $$$ on fire via GPUs. For us, Domain Knowledge + Good Data Engineering is the secret.
I think ML is going to be around for a long time because it works, even though DL is dominating the news right now. Just because a neurologist can also diagnose and treat common medical conditions (e.g a pneumonia), that doesn't mean we need every doctor to be neurologist.
Sanej|3 years ago
Salgat|3 years ago
Sanej|3 years ago
softwaredoug|3 years ago
ML isn't deep learning or not deep learning. It's fundamentally to me about a statistical formulation of a business problem.
It's how you would evaluate ML, formulate business tasks into an objective function, understand and develop training data, and what the features actually measure what’s important in the domain.
Sanej|3 years ago
PaulHoule|3 years ago
There are problems where classical ML works fine and if it works, why change it?
In text classification it depends on the problem but often the old methods work very well and there is not a lot of room for neural methods to do better.
For images or audio however I think a deep network would almost always be in the picture.
Often people use a pretrained neural network to make an embedding and then use classical ML methods to make a classifier that works on that embedding.
The data prep and evaluation process is very much the same no matter what kind of model you are using.
fdgsdfogijq|3 years ago
This couldnt be further from the truth. NLP/text algorithms have seen model improvements from NNs more than any other field.
alpineidyll3|3 years ago
That said, it's pretty saturated as a field of study. People work on uncertainty quantification etc. But it's unclear what numbers people would want to improve.
jononor|3 years ago
Sanej|3 years ago
bjourne|3 years ago
PartiallyTyped|3 years ago
It will only take over the cases where you have vast swaths of data, don't have reasonable preprocessing approaches that simplify the task, and don't need statistical guarantees.
psyklic|3 years ago
NNs also often perform similarly or worse than simpler models when you have "medium-sized" (and/or tabular) data. In fact, I nearly always start with simpler models when consulting -- why immediately make it complicated if a smaller, more interpretable model works well?
michaericalribo|3 years ago
stevofolife|3 years ago