Ask HN: Is deep learning obsession in college ill founded?
43 points| muazzam | 5 years ago
I'm a CS junior (about to become a senior) and in our last year, people choose their capstone project that they work on for the entire year. For some years (say since 2017), deep learning projects completely dominate the other projects in term of number and the awards that they go on to win. I understand the principles behind it, even find it cool, but the whole inscrutable nature of it is problematic to me.
[+] [-] ageitgey|5 years ago|reply
But that isn't really a problem. In most cases, the projects you do as an undergrad don't affect your professional life in any way after you get your first job. Very, very few undergrad projects turn into real projects that anyone uses after the student graduates.
So don't worry too much about it. Ten years ago, every senior project was an app. Twenty years ago, every senior project was a website. It's just a sign of the times and doesn't matter in the long run.
[+] [-] rvz|5 years ago|reply
Spot on.
On top of that we have 'AI' models getting fooled over adversarial attacks which just involve a single pixel change. As long as these issues are not tackled or not researched well enough, then we'll be pretty much be heading into another AI Winter and the hype cycle will go through its trough of disillusionment phase. Being unable to inspect the black-box nature of such deep-learning systems is why highly regulated industries involving danger to life such as healthcare or other safety critical industries label deep-learning solutions as unsafe for them.
Sure, all you see right now are other students and startups 'applying' deep learning everywhere, but they are hardly advancing the field unlike DeepMind and OpenAI are. In terms of learning, it's something good to learn as a student at college, but creating a AI startup now requires using Google, Amazon or Microsofts data center's for training which is clearly not sustainable anyway.
Security related projects and research are always where it's at.
[+] [-] visarga|5 years ago|reply
It's not all or nothing as you present it. ML models can be useful even if they are imperfect - and we should not forget humans aren't perfect either. For example, a model could reduce 50% of the time necessary to enter an invoice into the database. It's imperfect, yet useful.
A model need not run alone without any safety. It can have plain old programming rules to validate its outputs, or use human in the loop.
> Sure, all you see right now are other students and startups 'applying' deep learning everywhere, but they are hardly advancing the field unlike DeepMind and OpenAI are.
On the contrary, I would say that what DeepMind and OpenAI are doing is largely irrelevant for industry. There is a huge number of domains where no ML model has been created, and that is because there are so few people who can make them. The low hanging fruit hasn't been picked yet. It's like electricity at the beginning of the 20th century. The work these students and startups are doing is the good, useful work. You don't need DeepMind grade models to solve most real problems.
> creating a AI startup now requires using Google, Amazon or Microsofts data center's for training
You can train most useful models on a single machine today. Some, like Logistic Regression, train in seconds or minutes. Others take an hour, or a day. Some heavy ones take a week to train. If you don't do hyper-parameter search or cutting edge research you only need a few runs to get a working model. It's data tagging that usually takes months or years.
[+] [-] _bxg1|5 years ago|reply
[+] [-] sin7|5 years ago|reply
The vast majority of us are users. We massage the data to be in a certain shape, then feed it through a machine that someone else created. We can change the parameters. We can change the data. But few of us are going to look in to the code of a random forest function.
I've switched tracks and started doing web development. Playing with the hyper parameters in machine learning is no different than changing the feel of a drop down by changing the colors, fonts and other things to fit a certain aesthetic.
I could be wrong, but I have yet to meet anyone that has done anything besides use packages created by others to call themselves data scientists. I think that opens it up to becoming just another tool no different than Excel.
[+] [-] visarga|5 years ago|reply
[+] [-] tuatoru|5 years ago|reply
It's a problem whereever reliable operation is required, or analytic tractability (explanation) is required, or where resources available for data labeling are limited.
Its niche appears quite small, unless and until solid mathematical foundations are developed for it.
[+] [-] uoaei|5 years ago|reply
[+] [-] uoaei|5 years ago|reply
To understand why neural networks work, you will have to understand how a whole host of smaller, simpler ML models work in excruciating detail. Multiple linear regression, logistic regression, etc. What they mean, how they work, what's really going on "inside", what the underlying probabilistic model represents, etc.
Neural networks are great because it takes basically all of those smaller ideas and concatenates them into a super flexible statistical machine. It's really cool to see the "in->out" but it's even cooler once you have a good grasp on what's going on in the intermediate steps.
In my experience, almost everyone working with neural networks don't have those details down. This goes 100-fold for non-research roles. They learned the Keras API and are happy stacking layers, and as long as the output looks nice they push to production. For most cases empirical validation is probably enough, because NNs usually can achieve some incremental improvement just by virtue of the fact they have so many damn degrees of freedom. But to get a well-performing, well-founded model, you need to know the ins and outs.
[+] [-] deuslovult|5 years ago|reply
Imo deep learning is so popular because it "works". For a classification problem, if you try a linear baseline and a deep learning model, and you do a reasonable job of hyperparameter tuning and experimental design, it's likely you will outperform a simpler model. This holds true across many problem spaces.
I think the issue is that modern DL frameworks make it a little too easy to get pretty good performance on new problems. Other techniques generally require more background knowledge to make reasonable modeling assumptions, and still frequently perform worse than a naively applied DL approach.
I think DL will remain, in practice and education, a very popular tool. But it is essential to learn traditional statistical inference and other background to appropriately contextualize DL models so it isn't just some form of black magic.
[+] [-] mattkrause|5 years ago|reply
It's easy to beat a naive logistic regression model with a good neural network, but the gap often closes once you start trying to tune the logistic model too. (And it's not like the neural networks aren't tuned either--architecture search, data augmentation, etc).
Recent review on medical data: https://www.sciencedirect.com/science/article/abs/pii/S08954...
[+] [-] poulsbohemian|5 years ago|reply
So, pick something in "AI" that is the hotness of the moment, learn what you can, do your best, and then get on with life and career.
[+] [-] md2020|5 years ago|reply
[+] [-] visarga|5 years ago|reply
Neural networks performance on a problem is a benchmark of its real difficulty. It gives us insight, a new perspective.
In millennia of deliberations what have philosophers have discovered about the nature of intelligence? And then .. a neural net beats us at all board games, another can solve differential equations, another can translate, another can see, and so on. Have we really not learned anything by these inventions?
Another advantage of DL is that it frames the problem of intelligence in mathematical concepts and rigorous evaluation.
I, for one, have reconsidered all my spiritual beliefs after learning about the agent-environment-reward model of reinforcement learning. A new way of framing the agent and life, so parsimonious and powerful. And it does not require a soul, or a god, or anything outside out real environment, and yet can explain so much.
The whole machine learning paradigm is another powerful concept through which we can understand how we might function. Previously you might wonder how emotion, thought, sensation, imagination and will relate to each other. Now we can understand how they might be implemented and wired together, and what principles support their function.
[+] [-] Irishsteve|5 years ago|reply
It's not ideal - but if it wasn't DL it would be another topic / application.
[+] [-] debbiedowner|5 years ago|reply
It's not completely inscrutable in why it works, follow the thread on deep learning theory research by starting from the names here: http://www.vision.jhu.edu/tutorials/CVPR17-Tutorial-Math-Dee...
[+] [-] en4bz|5 years ago|reply
[+] [-] makapuf|5 years ago|reply
[+] [-] AnimalMuppet|5 years ago|reply
Given all that you've said, a capstone that tries to dent the "inscrutable nature" of deep learning might be an interesting choice.
[+] [-] omarhaneef|5 years ago|reply
I will point out that the real win is with new data sources, and simple linear regressions may still work there.
[+] [-] burfog|5 years ago|reply
Make a Rust front-end for the GNU Compiler Collection.
Emulate something.
Write a hypervisor.
[+] [-] muazzam|5 years ago|reply
[+] [-] sys_64738|5 years ago|reply
[+] [-] Kenji|5 years ago|reply