Been using Tensorflow embedded in a mobile app for a few months and honestly, I’m constantly surprised at how well thought-out the tooling is, and how quickly you can get results. Conversely, I think a few things are still unnecessarily dense (installing dependencies, optimizing hyper-parameters, and some of the embedded/XLA stuff is very raw). Kudos to the team though. It sounds like they’re on the right track with TF overall, and focusing on performance (including the XLA stuff) + ease of use (high-level, Keras API) is absolutely what I want as a user right now. Keep up the great work, y’all.
Would you happen to know if it requires additional code to support the Hexagon digital signal processor (DS)from Qualcomm Or is it automatic (kinda like switching between Tensorflow-CPU and Tensorflow-GPU)? I mainly work with Tensorflow on a PC so I'm not too familiar with the embedded variants of Tensorflow. Thanks!
It's worth pointing out Tensorflow is basically Google's clone of Theano, including a lot of the same design decisions. They've improved some things but it's not like Google handed us the secret to fire here. It's just a good implementation of the same things a lot of people have been working on for years.
I agree 100%. I'm not sure what AMD is thinking, but without support from major ML tools there is no chance of competing against NVidia in this space - and this space will grow larger and larger.
I totally agree with you as well. I was looking for a new graphics card and was debating between the GTX1050 or the RX480. I ended up getting the 1050 since it has CUDA and CUDANN support even though the RX480 has better specs.
> Plus, soon Google will open-source code that will multiply the speed of TensorFlow — specifically version three of Google’s Inception neural network model — by 58.
Uh, nope, that was speedup on 64 GPUs (or CPU cores, can't remember). i.e. it scales linearly, something that TF hasn't always been good at v other frameworks. I'm amazed a journalist with (I assume) basic technical competence could make this mistake.
I have a couple of applications in mind, mostly time series predictions. But the machine learning field seems to be vast and I don't know where to start.
The ML/DNN rabbit-hole goes deep. If the video above leaves you wanting more, http://www.deeplearningbook.org/ does a good job on drilling into more specifics for the various techniques used. The examples on the tensorflow webpage are also very good.
http://cs231n.github.io/ is a great site for beginners. I've been following the site along the Udacity Self Driving Car nanodegree. The CS231 material has helped me understand the concepts significantly.
Edit: I should mention that the class mainly focuses on neural networks and image recognition. However, once you have the foundation, you can apply your skillset to a vast range of applications.
Don't worry that just because it isn't using deep nets that it isn't state of the art or won't get the job done well. That would be like thinking python's built-in sort function isn't sufficient because it doesn't use Spark.
Deep learning is primarily a study of multi-layered neural networks, spanning over a great range of model architectures. This course is taught in the MSc program in Artificial Intelligence of the University of Amsterdam. In this course we study the theory of deep learning, namely of modern, multi-layered neural networks trained on big data. The course focuses particularly on computer vision and language modelling, which are perhaps two of the most recognizable and impressive applications of the deep learning theory.
I would recommend starting with a spreadsheet-sized dataset (no more than a few thousand records) where you want to predict one of the columns, use binary decision trees to try to predict it's value. Use either Azure ML Studio, or Jupyter with the Sci-Kit Learn library, depending on your comfort level with programming.
if you are based in the SF Bay Area you can come to Data Weekends (www.dataweekends.com). They are 2-day workshops to get started with Machine Learning and Deep Learning (full disclosure: I run them)
I am looking at the Martin Wicke talk. The Estimator API is very reminiscent of SparkML. Nice to see that the tensorflow crew are flexible enough to take good ideas from projects such as SparkML and Keras (now included natively in the TF stack).
Other highlights include the hotspot compiler (I was not that impressed so far, but it's early days for them), and embedded visualizations (looked quite cool) for visually inspecting learnt manifolds.
I stumbled across a three-chapter preview of the upcoming book Learning TensorFlow on Safari Books Online and went through them in a sitting. It was so accessible - both the book and TensorFlow itself - and inspired me to start learning math so that when the rest of the book comes out I will be better prepared to go deeper. I love learning in general, but haven't been this excited about learning something totally new (for me) in a long time.
I'll discuss this a bit during my talk at the dev summit.
The short answer is no.
The long answer is yes, but only if you create the model in Python, export it, and then feed training data in other languages. There are some people doing exactly that.
Long term, I'd like to give all languages equal footing, but there's quite a bit of work left.
TensorFlow comes with an easy-to-use Python interface and no-nonsense interfaces in other languages to build and execute computational graphs. Write stand-alone TensorFlow Python, C++, Java, or Go programs, or try things out in an interactive TensorFlow iPython notebook where you can keep notes, code, and visualizations logically grouped. This is just the start though — we're hoping to entice you to contribute interfaces to your favorite language — be it Lua, JavaScript, or R.
Could MacBook Pros (with Intel HD Graphics 3000 384 MB, to be more specific) train with GPU? I've always wanted to train algorithms but without using the GPU it is really slow.
I doubt the integrated Intel Card would be supported, even if it is, using the CPU would be just as good if not better. A lot of the high performance you see on GPUs are because of very highly optimized libraries available for Nvidia cards (like CuDNN) and so on.
Great news. I have several TensorFlow examples in a new book I am writing. I need to read up on the new higher level APIs, and can hopefully shorten the book example prob-grams.
We changed the URL from https://www.tensorflow.org/, which doesn't say anything about 1.0, to an article which gives a bit of background. If someone suggests a better URL we can change it again.
The URL when I first clicked this was the GitHub release notes, which is far more informative and apropos to the HN audience than either the TF landing page or vague VentureBeat pseudonews.
[+] [-] timanglade|9 years ago|reply
[+] [-] syntaxing|9 years ago|reply
[+] [-] redsummer|9 years ago|reply
[+] [-] codesternews|9 years ago|reply
[+] [-] andy_ppp|9 years ago|reply
https://www.youtube.com/watch?v=vq2nnJ4g6N0
I wish AMD graphics cards were supported fully. I really think AMD should find a way to work with the Tensor Flow team on this...
[+] [-] cityhall|9 years ago|reply
[+] [-] amenod|9 years ago|reply
[+] [-] syntaxing|9 years ago|reply
[+] [-] plexicle|9 years ago|reply
There are quite a few breaking changes but there is a very helpful conversion script here: https://github.com/tensorflow/tensorflow/tree/r1.0/tensorflo....
You can find the breaking changes in the 1.0 release here: https://github.com/tensorflow/tensorflow/releases/tag/v1.0.0
[+] [-] cityhall|9 years ago|reply
[+] [-] taliesinb|9 years ago|reply
Uh, nope, that was speedup on 64 GPUs (or CPU cores, can't remember). i.e. it scales linearly, something that TF hasn't always been good at v other frameworks. I'm amazed a journalist with (I assume) basic technical competence could make this mistake.
[+] [-] plexicle|9 years ago|reply
https://github.com/tensorflow/tensorflow/releases/tag/v1.0.0
You can follow the Summit live here: https://www.youtube.com/watch?v=LqLyrl-agOw
[+] [-] cstuder|9 years ago|reply
I have a couple of applications in mind, mostly time series predictions. But the machine learning field seems to be vast and I don't know where to start.
[+] [-] gregatragenet3|9 years ago|reply
The ML/DNN rabbit-hole goes deep. If the video above leaves you wanting more, http://www.deeplearningbook.org/ does a good job on drilling into more specifics for the various techniques used. The examples on the tensorflow webpage are also very good.
[+] [-] syntaxing|9 years ago|reply
Edit: I should mention that the class mainly focuses on neural networks and image recognition. However, once you have the foundation, you can apply your skillset to a vast range of applications.
[+] [-] imh|9 years ago|reply
Don't worry that just because it isn't using deep nets that it isn't state of the art or won't get the job done well. That would be like thinking python's built-in sort function isn't sufficient because it doesn't use Spark.
[+] [-] jray|9 years ago|reply
Deep learning is primarily a study of multi-layered neural networks, spanning over a great range of model architectures. This course is taught in the MSc program in Artificial Intelligence of the University of Amsterdam. In this course we study the theory of deep learning, namely of modern, multi-layered neural networks trained on big data. The course focuses particularly on computer vision and language modelling, which are perhaps two of the most recognizable and impressive applications of the deep learning theory.
http://uvadlc.github.io/
[+] [-] sputknick|9 years ago|reply
[+] [-] anhari|9 years ago|reply
https://www.reddit.com/r/MachineLearning/wiki/index
[+] [-] ghego|9 years ago|reply
[+] [-] jamesblonde|9 years ago|reply
[+] [-] drakonka|9 years ago|reply
[+] [-] Omnipresent|9 years ago|reply
[+] [-] fest|9 years ago|reply
[+] [-] jjuel|9 years ago|reply
[+] [-] whorleater|9 years ago|reply
[+] [-] phodo|9 years ago|reply
[+] [-] sixbrx|9 years ago|reply
[+] [-] vomjom|9 years ago|reply
The short answer is no.
The long answer is yes, but only if you create the model in Python, export it, and then feed training data in other languages. There are some people doing exactly that.
Long term, I'd like to give all languages equal footing, but there's quite a bit of work left.
[+] [-] sremani|9 years ago|reply
-----------------------------------------------------
Language options
TensorFlow comes with an easy-to-use Python interface and no-nonsense interfaces in other languages to build and execute computational graphs. Write stand-alone TensorFlow Python, C++, Java, or Go programs, or try things out in an interactive TensorFlow iPython notebook where you can keep notes, code, and visualizations logically grouped. This is just the start though — we're hoping to entice you to contribute interfaces to your favorite language — be it Lua, JavaScript, or R.
--------------------------------------------------------
[+] [-] desku|9 years ago|reply
[+] [-] thasaleni|9 years ago|reply
[+] [-] mrcabada|9 years ago|reply
[+] [-] govg|9 years ago|reply
[+] [-] jamesblonde|9 years ago|reply
[+] [-] mark_l_watson|9 years ago|reply
[+] [-] alvivar|9 years ago|reply
I really like Python, but F# <3
[+] [-] ndesaulniers|9 years ago|reply
[+] [-] shmageggy|9 years ago|reply
[+] [-] 120bits|9 years ago|reply
[+] [-] chrisra|9 years ago|reply
[+] [-] amelius|9 years ago|reply
[+] [-] anonfunction|9 years ago|reply
[+] [-] dang|9 years ago|reply
[+] [-] runesoerensen|9 years ago|reply
[+] [-] plexicle|9 years ago|reply
[+] [-] unknown|9 years ago|reply
[deleted]
[+] [-] aschampion|9 years ago|reply