I am an average solo user / applied researcher using Windows locally with a GTX 1070 8GB and looking for speed and documentation first, so Theano is way ahead in these departments. That said, we are also told TensorFlow is the next big thing because of scalability (?). TensorFlow works under Windows with Anaconda and Python 3.5 through Keras as well, so I have it available indeed and can try the benchmarks. Where do we stand, really? Thanks.
[+] [-] asah|8 years ago|reply
TensorFlow is a platform "winner" and approx 100% of all innovations will quickly be ported to TensorFlow - TBD which of the others will "keep up" with innovations as they continue to come out.
other recommendations:
- by default, TensorFlow allocates 100% of GPU RAM for each process. You'll want to control this: https://stackoverflow.com/questions/34199233/how-to-prevent-...
- Keras. yes, this. Dramatically reduces code by 2-10x, without loss of control AFAICT.
- cloud hardware. Pretty quickly, you'll want to scale and run multiple tests at once, and e.g. quickly backup & copy data, replicate system images, etc. I use Google Cloud Hosting and it's much easier (and cheaper) than AWS. Haven't tried Azure but heard good things. At least once, Google's internet bandwidth has saved hours waiting for data transfers.
[+] [-] laingc|8 years ago|reply
Far better, in my view, is to work with the newer Estimators API. It is almost as fool-proof as Keras, but instead of trying to be a framework as such, the Estimators/learn API essentially just wraps up some of the boilerplate that you need with raw tensorflow, and internally looks fairly similar to the code you might write yourself. Consequently, it preserves the composability of TF far better than Keras.
[+] [-] deepnotderp|8 years ago|reply
[+] [-] jasonb778|8 years ago|reply
[deleted]
[+] [-] rryan|8 years ago|reply
Instead of coding imperatively, you write code to build a computation graph. The graph is a data structure that fully describes the computation you want to perform (e.g. training or inference of a machine learning model).
* That graph can be executed immediately, or stored for later.
* Since it's a serializable data structure, you can version it quite easily.
* You can deploy it to production without production having to depend on ANY of the code that built the graph, only the runtime necessary to execute it.
* You can run a compiler on it (such as XLA or TensorFlow's built in graph rewriter) to produce a more efficient version of the graph.
* In some circumstances, you can even compile the runtime away, producing a single .h/.o that you can link directly into e.g. a mobile app.
It's a beautiful and highly useful abstraction that allows TensorFlow to have both a great development and production story in one framework. Most frameworks only have a good story for either development or production.
If you are a machine learning researcher who doesn't need or care about deploying your work (i.e. mostly publishing papers), you may not want the overhead of having to deal with building a graph, and may prefer something that computes imperatively like PyTorch. If you are building products / services that use ML and developing/training your own models (as opposed to taking pre-trained models and using them), there is really no credible competitor to TensorFlow.
Disclaimer: I work at Google. I spend all day writing TensorFlow models. I'm not on the TensorFlow team nor do I speak for them or Google.
[+] [-] fnl|8 years ago|reply
MXNet has amalgamation http://mxnet.io/how_to/smart_device.html#amalgamation-making...
CNTK provides a managed ("evaluation") library solution to deploy your models and embed them in C, C++, C#, Python, and even an experimental Java version. https://docs.microsoft.com/en-us/cognitive-toolkit/CNTK-Eval...
How's that not competitive to TF? MXNet's approach is a bit unwieldy, yes, but seems easily streamlined. And CNTK's deployment method looks perfectly fine. Note I haven't checked other DL libs, but it seems unreasonable that Microsoft and Amazon have no "competitive" solution for deployment.
[+] [-] unknown|8 years ago|reply
[deleted]
[+] [-] unknown|8 years ago|reply
[deleted]
[+] [-] sirfz|8 years ago|reply
1) Transition was fairly straightforward, both APIs' interfaces are more-or-less similar and share some design characteristics.
2) Having said that, TF's API is easier to use and without a doubt a lot easier to read.
3) Consistency: Deploying Theano in different environments surprised me on several occasions with different output compared to the training environment. TF is more consistent on this front (never had such issues).
4) Running multiprocessing with Theano + GPU is a disaster (due to forking) so I end up having to create process pools before initializing Theano. No such issues with TF.
5) TF provides many helpful operators (such as queues and batching ops) as well as monitoring tools (Tensorboard) and debugging tools.
6) Its development is extremely rapid, new releases every couple of months with a lot of improvements and new features every time.
In short, TF is what Theano should have been. A lot of new papers are being developed in TF as well so it helps to understand it.
[+] [-] digitalzombie|8 years ago|reply
How stable is the api then?
I think google is a bit notorious for this (e.g. Angular vs Angular 2).
[+] [-] paulsutter|8 years ago|reply
We use several frameworks because sample code from different papers uses different frameworks. It's not that big of a deal.
[+] [-] cs702|8 years ago|reply
* https://twitter.com/fchollet/status/765212287531495424
* https://news.ycombinator.com/item?id=12391744
* https://github.com/aymericdamien/TopDeepLearning
[+] [-] aisofteng|8 years ago|reply
[+] [-] julsimon|8 years ago|reply
[+] [-] visarga|8 years ago|reply
[+] [-] DLEnthusiast|8 years ago|reply
According to https://medium.com/@karpathy/icml-accepted-papers-institutio... , 3 of the top research labs in the world are DeepMind, Google Brain (and the rest of Google), and Microsoft Research. Let's see:
* DeepMind: TensorFlow
* Google Brain: TensorFlow
* Microsoft Research: CNTK
Ok, so what about academia? The top deep learning groups in academia are:
* Montreal: Theano
* Toronto: TensorFlow
* IDSIA: TensorFlow
So, what about the greater academic research community? Maybe we could get some data about who uses what by looking at the frameworks cited by researchers in their papers. Andrej did that: it's mainly TensorFlow and Caffe. https://medium.com/@karpathy/a-peek-at-trends-in-machine-lea...
[+] [-] kasbah|8 years ago|reply
[+] [-] curiousgal|8 years ago|reply
and yet a lot of researchers are using Caffe.
[+] [-] ssivark|8 years ago|reply
Why is the choice of framework such a big deal? Is it unreasonable to expect someone well-versed in one framework to be able to pick up another reasonably fast if/when collaborating with someone proficient in the latter ?
[+] [-] laingc|8 years ago|reply
It may become important if you end up having a ton of models running in production that need to be maintained and further developed, but in general for new applications I would say that substantially less than 5% of your time would (should!) be spent actually writing any code.
[+] [-] anxman|8 years ago|reply
Disclosure: Work for Google Cloud
[+] [-] asah|8 years ago|reply
[+] [-] nafizh|8 years ago|reply
[+] [-] k__|8 years ago|reply
[+] [-] mmv|8 years ago|reply
Creating a small neural network and training it over the MNIST dataset is like the 'todo list' starter project of this kind of frameworks.
[+] [-] bajaj|8 years ago|reply
[+] [-] massaman_yams|8 years ago|reply
[+] [-] ma2rten|8 years ago|reply
* more widely used, more example code
* developed by a bigger team, likely to improve faster
* easier to deploy
* training with Cloud ML
* better support for distributed training
* no compile time (this can be long especially for RNNs)
[+] [-] yuanchuan|8 years ago|reply
As such you won't need to implement/convert your model in another format for usage.
[+] [-] torbjorn|8 years ago|reply
[+] [-] johnsmith21006|8 years ago|reply
[+] [-] ci5er|8 years ago|reply
[+] [-] manis404|8 years ago|reply
[+] [-] chronic7ui|8 years ago|reply
[+] [-] he0001|8 years ago|reply
[+] [-] he0001|8 years ago|reply