top | item 17622045

Jeff Dean’s ML System Architecture Blueprint

231 points| trcytony | 7 years ago |medium.com | reply

29 comments

order
[+] mark_l_watson|7 years ago|reply
I work in this field, but nothing state of the art (relatively simple LSTM models and GAN models). While I found the article informative, it was also a little depressing to see how far research goes beyond what I am working on. I spend about 8 hours a week off-work-hours studying and reading papers and I find it difficult to keep up.
[+] antpls|7 years ago|reply
I think it makes more sense to focus on the benchmarks. Benchmarks change less often than the underlying algorithms/models, and they are easier to follow.

Once a model performs consistently well on a given benchmarks over several years, then it makes sense to get more into the details.

For example in the NLP field, in 2018, there is a focus on multi-tasks models. Some studies (don't have the refs at hand, sorry) suggest that different models generalize differently (and some time better) when trained on several tasks at once.

Anyway, those papers and models are the result of team of researchers working on the problem full time, with tons of data at their hands. If any sane individuals were able to keep up with the state of the art, it wouldn't be a research field, I guess :-)

[+] jahjaylee|7 years ago|reply
Although a neat thought, are the number of papers on ArXiv really something worth comparing to Moore's law? Like at that point, what can't you compare to Moore's law...
[+] spullara|7 years ago|reply
I once asked Gordon Moore what the software equivalent of Moore's law was, he responded without pause: "the number of bugs doubles every 18 months".
[+] zeusk|7 years ago|reply
Number of things being compared to moore law?
[+] randcraw|7 years ago|reply
IMHO, the two are incomparable. The exponential growth rate of Moore's Law was driven largely by a linear rate of shrinkage in 2D, which drove up clock rates geometrically as microarchitecture component distances shrank two-fold (until CMOS' heat finally fought back). ML has no similar geometric driven basis that will continue to drive its rate of growth superlinear.

I suspect this plot is Dean's way of paying homage to Patterson, since he and Hennessy were famous for similar plots describing CPU performance in their two architecture textbooks.

[+] joe_the_user|7 years ago|reply
Even more, an exponent increase in the number of papers on a subject might mean something but it could easily mean something other than the optimistic scenario.
[+] kingvash|7 years ago|reply
Google really seems to be leading the pack with investments in silicon (e.g. TPU and recently announced edge tpu[1]). Other traditional silicon companies (Nvidia, Intel) seem to get it but I have yet to see investments from other tech companies (Amazon, FB, Netflix).

[1] https://techcrunch.com/2018/07/25/google-is-making-a-fast-sp...)

[+] zeusk|7 years ago|reply
Microsoft Research is pushing along as well, they just don't publicize it as much.

Azure has already deployed FPGAs (they believe, being able to deploy and make changes including changes to workload on the fly is more beneficial than the efficiency compared to using an ASIC) for networking and accelerated ML (Project Catapult and Project Brainwave).

tbh, I do agree with using FPGAs over ASICs given the speed at which the tech is moving. Google has already cycled through 3 versions of the TPU.

[+] seanmcdirmid|7 years ago|reply
I heard that Google was actually late to the game on this, not realizing the value of GPUs over (just) distributed computing when the DNN trend started. Now they even have custom silicon...
[+] typeformer|7 years ago|reply
The blueprint calls for the AI to train on a data set of everything Jeff Dean does or thinks for the entirety of his life.
[+] typeformer|7 years ago|reply
No love for bad Jeff Dean jokes here I guess.
[+] drewmassey|7 years ago|reply
Interesting. There is a kind of obvious conflict between cloud resources and hardware intensive applications like ML. The zeitgeist is obviously swinging.
[+] ratsimihah|7 years ago|reply
Am I the only one thrown off by Jeff Dean wearing a suit?
[+] thelastidiot|7 years ago|reply
It's called respecting your audience. Everyone probably looked like from a 1970s movie in the room.
[+] sytelus|7 years ago|reply
First, -1 to authors for publishing paper in IEEE Macro which is behind paywall. We need to start mass boycott of all IEEE journals considering they are as bas as Elsevier but have successfully painted themselves as good guys. Number of papers in ML that I came across and behind paywall are mostly from IEEE. In any case, we expect better from authors in Google Brain to agree to publish behind any paywall!

Second, I fail to see any real takeaway or key new insight. Number of papers grows exponentially in many fields in initial periods.