Submitted by GraciousReformer t3_118pof6 in MachineLearning
[deleted] t1_j9jgblt wrote
Reply to comment by randomoneusername in [D] "Deep learning is the only thing that currently works at scale" by GraciousReformer
[deleted]
randomoneusername t1_j9jkzs7 wrote
The statement stand-alone that you have there is very vague. Can I assume is taking about NLP or CV projects ?
On tabular data even with non linear relationship normal boosting, ensemble algorithms can scale and be top of the game.
bloodmummy t1_j9jzvnr wrote
It strikes me that people who tout DL as a hammer-for-all-nails never touched tabular data in their lives. Go try to do a couple of Kaggle Tabular competitions and you'll soon realise that DL can be very dumb, cumbersome, and data-hungry. Ensemble models,Decision Tree models, and even feature-engineered Linear Regression models still rule there and curb-stomp DL all day long ( For most cases ).
Tabular data is also still the type of data most-used with ML. I'm not a "DL-hater" if there is such a thing, in fact my own research is using DL only. But it isn't a magical wrench, and it won't be.
Mefaso t1_j9jgvoz wrote
Anything that scales sub-quadraticaly?
Anything "big-data"
GraciousReformer OP t1_j9jib2n wrote
Then why DL?
suflaj t1_j9jjetb wrote
Because it requires the least amount of human intervention
Also because it subjectively sounds like magic to people who don't really understand it, so it both sells to management and to consumers.
At least it's easier for humans to cope it is like magic than to accept that a lot of what AI can do is just stuff that is trivial and doesn't require humanity to solve.
Viewing a single comment thread. View all comments