When I worked at CERN around 2010, Boosted Decision Trees were the most popular classifier, exactly due to the (potential for) explainability along with its power of expression.
We had a cultural aversion for neural networks back then, especially if the model was used in physics analysis directly.
Times have changed…
ekjhgkejhgk|16 hours ago
I've always thought that the idea that decision trees are "explainable" is very overstated. The moment that you go past a couple of levels in depth, it becomes an un-interpretable jungle. I've actually done the exercise of inspecting how a 15-depth decision trees makes decision, and I found it impossible to interpret anything.
In a neural network you can also follow the successive matrix multiplications and relu etc through the layers, but you end up not knowing how the decision is made.
Thoughts?
lokimedes|15 hours ago
My second job after physics was AI for defense, and boy is the dream of explainable AI alive there.
Honesty anyone who “needs” AI to be understandable by dissection, suffers from control issues :)
srean|19 hours ago
This makes me a little concerned -- the use of parameters rich opaque models in Physics.
Ptolemaic system achieved a far better fit of planetary motion (over the Copernican system) because his was a universal approximator. Epicyclic system is a form of Fourier analysis and hence can fit any smooth periodic motion. But the epicycles were not the right thing to use to work out the causal mechanics, in spite of being a better fit empirically.
In Physics we would want to do more than accurate curve fitting.
lokimedes|19 hours ago
unknown|16 hours ago
[deleted]
wodenokoto|21 hours ago
boccaff|20 hours ago
longer answer: Random forests use the average of multiple trees that are trained in a way to reduce the correlation between trees (bagging with modified trees). Boosting trains sequentially, with each classifier working on the resulting residuals so far.
I am assuming that you meant boosted decision trees, sometimes gradient boosted decisions trees, as usually one have boosted decision trees. I think xgboost added boosted RF, and you can boost any supervised model, but it is not usual.
hansvm|20 hours ago
For better or for worse (usually for better), boosted decision trees work harder to optimize the tree structure for a given problem. Random forests rely on enough trees being good enough.
Ignoring tree split selection, one technique people sometimes do makes the two techniques more related -- in gradient boosting, once the splits are chosen it's a sparse linear algebra problem to optimize the weights/leaves (iterative if your error is not MSE). That step would unify some part of the training between the two model types.