Very cool to see this! It turns out my wife and I bought Andy Barto’s (and his wife’s) house.
During the process, there was a bidding war. They said “make your prime offer” so, knowing he was a mathematician, we made an offer that was a prime number :-)
Nice! Well deserved. They make both editions of their RL textbook available as a free to read PDF. I have been a paid AI practitioner since 1982, and I must admit that RL is one subject I personally struggle mastering, and the Sutton/Barto book, the Cousera series on RL taught by Professors White and White, etc. personally helped me: recommended!
Indeed a bitter lesson. I once enjoyed encoding human knowledge into a computer because it gives me understanding of what's going on. Now everything is becoming a big black box that is hard to reason about. /sigh/
Also, Moore's law has become a self-fulfilling prophecy. Now more than ever, AI is putting a lot of demand on computational power, to the point which drives chip makers to create specialized hardware for it. It's becoming a flywheel.
This depends a little bit on what the goal of AI research is. If it is (and it might well be) to build machines that excel at tasks previously thought to be exclusively reserved to, or needing to involve, the human mind, then these bitter lessons are indeed worthwhile.
But if you do AI research with the idea that by teaching machines how to do X, we might also be able to gain insight in how people do X, then ever more complex statistical setups will be of limited information.
Note that I'm not taking either point of view here. I just want to point out that perhaps a more nuanced approach might be called for here.
> In computer vision, there has been a similar pattern. Early methods conceived of vision as searching for edges, or generalized cylinders, or in terms of SIFT features. But today all this is discarded.Modern deep-learning neural networks use only the notions of convolution and certain kinds of invariances, and perform much better.
I was there, at that moment where pattern matching for vision started to die.
That was not completely lost though, learning from that time is still useful on other places today.
This describes Go AIs as a brute force strategy with no heuristics, which is false as far as I know. Go AIs don't search the entire sample space, they search based on their training data of previous human games.
I remember the article, and remember how badly it missed the point... The goal of writing a chess program that could beat a world champion wasn't to beat the world champion... the goal was to gain understanding into how anyone can play chess well. The victory in that match would've been equivalent to eg. drugging Kasparov prior to the match, or putting a gun to his head and telling him to lose: even cheaper and more effective.
Good time to remind everyone that Sutton is a human successionist and doesn't care if humans all die. He is not to be trusted nor celebrated: https://www.youtube.com/watch?v=NgHFMolXs3U
The ACM award is for their professional academic achievements - this fetishism to dig into another person’s personal life and find the most weird thing they said as the thing that paints over all of their life’s achievements as evil must stop.
It’s silly and dangerous. Because you don’t like thing A and they said/did thing A all of their lofty accomplishments get nullified by anyone. And worst of all internet gives your opinion the same weight as someone else (or the rest of us) who knows a lot about thing B that could change the world. From a strictly professional capacity.
This works me up because this is what’s dividing up people right now at a much larger scale.
Have you ever met Sutton? He is the most heart-warming, caring and passionate hippy I have ever met. He does not want all humans to die. The talk you link also doesn't support your claim. Perhaps I missed it, in that case, do leave a timestamp.
In the talk, he says it will lead to an era of prosperity for humanity, however without humanity being in sole control of their destiny. His conclusion slide (at 12:33) literally has the bullet point "the best hope for a long-term future for humanity". That is opposite to you saying he "doesn't care if humans all die".
If I plan for my succession, I don't hope nor expect my daughter will murder me. I'm hoping for a long retirement in good health after which I will quietly pass in my sleep, knowing I left her as well as I could in a symbiotic relationship with the universe.
That seems to be a harsh and misleading framing of his position. My own reading is that he believes it is inevitable that humans will be replaced by transhumans. That seems more like wild sci-fi utopianism than ill-will. It doesn't seem like a reason to avoid celebrating his academic achievements.
It is interesting that you bring this to the attention, but I don't see why we should not trust or celebrate someone if they have views that you don't agree with.
Edit: especially since I think your implied claim that Sutton would actively want everyone to die seems very much unfounded.
Very disappointing. I do not understand how people earnestly defend the successionist view as a good future, but I thought he might at least give some interesting arguments.
This talk isn't that. There are no substantive arguments for why we should embrace this future and his representation of the opposite side isn't in good faith either, instead he chose to present straw-man versions of them.
He concludes with "A successful succession offers [...] the best hope for a long-term future for humanity.
How this can possibly be true when ai succession necessarily includes replacement eludes me. He does mention transhumanism on a slide, but it seems extremely unlikely that he's actually talking about that and the whole succession spiel is just unfortunate wording.
I used their RL book for a course I taught. It's beautifully written and freely available (http://incompleteideas.net/book/the-book-2nd.html)! I kept getting distracted by the beautiful writing that I would miss the actual content.
Huge congratulations to Andrew Barto and Richard Sutton on the well-deserved Turing Award! as a student, their textbook Reinforcement Learning: An Introduction was my gateway into the field. I still remembered that how Chapter 6 on ‘Temporal Difference Learning’ fundamentally reshaped the way I thought about sequential decision-making.
a timeless classic that I still highly recommend reading today!
This is a long time coming. To see through an idea from start to finish and make this span an entire field instead of a sub chapter in a dynamic programming book.
I wish a lot more games actually ended up using RL - the place where all of this started in the first place - would be really
cool!
Well deserved, RL will only gain more importance as time goes on thanks to its (and neural nets) flexibility. The bitter lesson won't feel so bitter as we scale.
He gave up his US citizenship years ago but he explains some of the reasons why he left. I'll also say that the AI research coming out of Canada is pretty great as well so I think it makes sense to do research there.
As someone who grew up in Edmonton, attended the U of A, and had the good fortune of receiving an incredible CS education at a discount price, I'm incredibly grateful for his (and the other amazing professors there) immense sacrifice.
Great people and cheap cost of living, but man do I not miss the city turning into brown sludge every winter.
He's been there since he left Bell Labs, in the mid 2000's, I think. The U of A is, or was, rich with Alberta oil sands money and willing to use it to fund "curiosity-driven research", which is pretty nice if you're willing to live where the temperatures go down to -40 in the winter.
These guys are great but unfortunately the ai sutton and barto book is really bad.
You would do better with Grokking Machine Learning by trask, and then a couple months of implementing ml papers.
I second this suggestion. Read Grokking Deep Reinforcement Learning before reading Sutton. Well, the Sutton book is free, so take a peak, but if the formulas scare you then read Grokking Deep Reinforcement Learning.
[+] [-] zackkatz|1 year ago|reply
During the process, there was a bidding war. They said “make your prime offer” so, knowing he was a mathematician, we made an offer that was a prime number :-)
So neat to see him be recognized for his work.
[+] [-] mark_l_watson|1 year ago|reply
EDIT: the example programs for their book are available in Common Lisp and Python. http://incompleteideas.net/book/the-book-2nd.html
[+] [-] ofirpress|1 year ago|reply
[+] [-] cxr|1 year ago|reply
[+] [-] khaledh|1 year ago|reply
Also, Moore's law has become a self-fulfilling prophecy. Now more than ever, AI is putting a lot of demand on computational power, to the point which drives chip makers to create specialized hardware for it. It's becoming a flywheel.
[+] [-] kleiba|1 year ago|reply
But if you do AI research with the idea that by teaching machines how to do X, we might also be able to gain insight in how people do X, then ever more complex statistical setups will be of limited information.
Note that I'm not taking either point of view here. I just want to point out that perhaps a more nuanced approach might be called for here.
[+] [-] jdright|1 year ago|reply
I was there, at that moment where pattern matching for vision started to die. That was not completely lost though, learning from that time is still useful on other places today.
[+] [-] Buttons840|1 year ago|reply
[+] [-] DavidPiper|1 year ago|reply
[+] [-] crabbone|1 year ago|reply
[+] [-] perks_12|1 year ago|reply
[+] [-] blufish|1 year ago|reply
[+] [-] porridgeraisin|1 year ago|reply
[+] [-] barrenko|1 year ago|reply
[+] [-] incognito124|1 year ago|reply
[+] [-] jxjnskkzxxhx|1 year ago|reply
[+] [-] zelphirkalt|1 year ago|reply
[+] [-] darkoob12|1 year ago|reply
[+] [-] vonneumannstan|1 year ago|reply
[+] [-] textlapse|1 year ago|reply
It’s silly and dangerous. Because you don’t like thing A and they said/did thing A all of their lofty accomplishments get nullified by anyone. And worst of all internet gives your opinion the same weight as someone else (or the rest of us) who knows a lot about thing B that could change the world. From a strictly professional capacity.
This works me up because this is what’s dividing up people right now at a much larger scale.
I wish you well.
[+] [-] 317070|1 year ago|reply
In the talk, he says it will lead to an era of prosperity for humanity, however without humanity being in sole control of their destiny. His conclusion slide (at 12:33) literally has the bullet point "the best hope for a long-term future for humanity". That is opposite to you saying he "doesn't care if humans all die".
If I plan for my succession, I don't hope nor expect my daughter will murder me. I'm hoping for a long retirement in good health after which I will quietly pass in my sleep, knowing I left her as well as I could in a symbiotic relationship with the universe.
[+] [-] zoogeny|1 year ago|reply
That seems to be a harsh and misleading framing of his position. My own reading is that he believes it is inevitable that humans will be replaced by transhumans. That seems more like wild sci-fi utopianism than ill-will. It doesn't seem like a reason to avoid celebrating his academic achievements.
[+] [-] smokel|1 year ago|reply
Edit: especially since I think your implied claim that Sutton would actively want everyone to die seems very much unfounded.
[+] [-] cowsandmilk|1 year ago|reply
His last slide literally says “best hope for a long-term future for humanity”. That’s literally the opposite of what you’re claiming.
[+] [-] visarga|1 year ago|reply
[+] [-] nycticorax|1 year ago|reply
[+] [-] Version467|1 year ago|reply
This talk isn't that. There are no substantive arguments for why we should embrace this future and his representation of the opposite side isn't in good faith either, instead he chose to present straw-man versions of them.
He concludes with "A successful succession offers [...] the best hope for a long-term future for humanity. How this can possibly be true when ai succession necessarily includes replacement eludes me. He does mention transhumanism on a slide, but it seems extremely unlikely that he's actually talking about that and the whole succession spiel is just unfortunate wording.
[+] [-] ks2048|1 year ago|reply
[+] [-] rhema|1 year ago|reply
[+] [-] cxie|1 year ago|reply
a timeless classic that I still highly recommend reading today!
[+] [-] textlapse|1 year ago|reply
I wish a lot more games actually ended up using RL - the place where all of this started in the first place - would be really cool!
[+] [-] jimbohn|1 year ago|reply
[+] [-] j7ake|1 year ago|reply
Shows he has integrity and is not a careerist focused on prestige and money above all else.
[+] [-] armSixtyFour|1 year ago|reply
He gave up his US citizenship years ago but he explains some of the reasons why he left. I'll also say that the AI research coming out of Canada is pretty great as well so I think it makes sense to do research there.
[+] [-] tbrockman|1 year ago|reply
Great people and cheap cost of living, but man do I not miss the city turning into brown sludge every winter.
[+] [-] jp57|1 year ago|reply
[+] [-] Philpax|1 year ago|reply
[+] [-] optimalsolver|1 year ago|reply
[+] [-] jamesblonde|1 year ago|reply
[+] [-] rvz|1 year ago|reply
[+] [-] wegfawefgawefg|1 year ago|reply
[+] [-] Buttons840|1 year ago|reply
[+] [-] 317070|1 year ago|reply
[+] [-] carabiner|1 year ago|reply
[+] [-] pklee|1 year ago|reply
[+] [-] PartiallyTyped|1 year ago|reply
[+] [-] nextworddev|1 year ago|reply
[+] [-] byyoung3|1 year ago|reply