RichardSSutton Profile Banner
Richard Sutton Profile
Richard Sutton

@RichardSSutton

Followers
57K
Following
120
Media
33
Statuses
366

Student of mind and nature, libertarian, chess player, cancer survivor. @ Keen, UAlberta, Amii, https://t.co/u8za2Kod54, The Royal Society, Turing Award

Edmonton, Alberta, Canada
Joined October 2010
Don't wanna be here? Send us removal request.
@RichardSSutton
Richard Sutton
2 years
AI researchers seek to understand intelligence well enough to create beings of greater intelligence than current humans. Reaching this profound intellectual milestone will enrich our economies and challenge our societal institutions. It will be unprecedented and
69
139
927
@professor_ajay
Ajay Agrawal
27 days
Investors are making a multi-trillion dollar bet on the value of scaling LLMS. Against this backdrop @korymath role plays a character delivering the opening keynote at MatrixCon. Someone asks what @karpathy means: "Probably we are still not sufficiently bitter lesson pilled and
1
3
16
@RichardSSutton
Richard Sutton
27 days
And the new Superintelligence Research Lab will be centered in... Edmonton!
@lazyuniverse
Giri ATG
28 days
Launching our Research Lab : Advancing experience powered, decentralized superintelligence - built for continual learning, generalization & model-based planning. Press Release : https://t.co/iPYXb1nzYr We’re solving the hardest challenges in real-world industries, robotics,
19
23
414
@RayDalio
Ray Dalio
1 month
History and logic have made clear that sanctions reduce the demand for fiat currencies and debts denominated in them and support gold. Throughout history, before and during shooting wars, there have been financial and economic wars that we now call sanctions (which means cutting
138
508
3K
@sfiscience
Santa Fe Institute
1 month
Congratulations to SFI Professor Melanie Mitchell (@MelMitchell1), a winner of the 2025 National Academies Eric and Wendy Schmidt Awards for Excellence in Science Communications (@SciCommAwards). Mitchell is recognized for her writing and podcasting on topics related to AI and
8
19
170
@kjaved_
Khurram Javed
1 month
As long as robotic systems are stuck with sim2real, there is little hope of having general-purpose robots. Sim2real only makes economic sense for a handful of situations where the cost of failure is astronomically high. It's easy to spot these situations because even humans
@M33pinator
Kris D.
1 month
Reinforcement learning 🧠 on robots 🤖 can’t stay in simulation forever. My new post explores why direct, on-hardware learning matters and how we also need smarter mechanical design to enable it. https://t.co/0GKl8naFJ8
14
21
190
@RichardSSutton
Richard Sutton
1 month
Learning is the derivative of knowledge.
94
138
2K
@RichardSSutton
Richard Sutton
1 month
To learn more about temporal difference learning, you could read the original paper ( https://t.co/0cGg3YD4Ws) or watch this video ( https://t.co/dOa3rfOPhn).
Tweet card summary image
videolectures.net
@kjaved_
Khurram Javed
2 months
The Dwarkesh/Andrej interview is worth watching. Like many others in the field, my introduction to deep learning was Andrej’s CS231n. In this era when many are involved in wishful thinking driven by simple pattern matching (e.g., extrapolating scaling laws without nuance), it’s
18
123
1K
@CsabaSzepesvari
Csaba Szepesvari
1 month
@karpathy It seems to me that not only you, but too many people talk about RL as if these two things were the same, which prevents a more nuanced discussion. 2/2
4
5
104
@RichardSSutton
Richard Sutton
1 month
Well said
@CsabaSzepesvari
Csaba Szepesvari
1 month
@karpathy @karpathy I think it would be good to distinguish RL as a problem from the algorithms that people use to address RL problems. This would allow us to discuss if the problem is with the algorithms, or if the problem is with posing a problem as an RL problem. 1/x
6
15
428
@RichardSSutton
Richard Sutton
2 months
More on LLMs, RL, and the bitter lesson, on the Derby Mill podcast.
@professor_ajay
Ajay Agrawal
2 months
5
18
241
@beforeasi
before
2 months
@RichardSSutton @dwarkesh_sp no you didn't misspoke there, Richard. I miss quoted, the video and caption says " training " itself. Apologies, 🙏
0
1
27
@kjaved_
Khurram Javed
2 months
This is a reasonable take on the podcast. One thing I would add is that people underestimate just how much babies learn as opposed to what they are born with. One of the big differences between us and other animals might just be that we rely much more on learning because we have
@karpathy
Andrej Karpathy
2 months
Finally had a chance to listen through this pod with Sutton, which was interesting and amusing. As background, Sutton's "The Bitter Lesson" has become a bit of biblical text in frontier LLM circles. Researchers routinely talk about and ask whether this or that approach or idea
22
31
293
@RichardSSutton
Richard Sutton
2 months
Still timely
@RichardSSutton
Richard Sutton
3 years
Lots of exaggeration about AI lately. The hype is that LLMs have anything to do with intelligence. The FUD is that AIs will enslave us. I like this cartoon in the New Yorker because it suggests the ridiculousness of both memes.
27
46
543
@kjaved_
Khurram Javed
2 months
This will age poorly. I largely have an optimistic view of LLMs; I use multiple LLM tools daily, and I don't think the LLM tech stack is a bubble—it will create a lot of value. I disagree that the length of tasks that LLMs can do has been doubling every 7 months. There are tasks
@Mononofu
Julian Schrittwieser
2 months
As a researcher at a frontier lab I’m often surprised by how unaware of current AI progress public discussions are. I wrote a post to summarize studies of recent progress, and what we should expect in the next 1-2 years: https://t.co/B7438Z9lOF
54
71
828
@JosephModayil
Joseph Modayil
2 months
This is a thoughtful writeup, as I expect from Rod Brooks. I think he is right on the importance of input representation, touch, and physical safety in deployment. I also think he underestimates the potential for representation and subgoal discovery with reinforcement learning.
@rodneyabrooks
Rodney Brooks
2 months
I have just finished and just published some weekend reading for you. 9,600 words of not easy reading, on why today's humanoid robots won't learn to be dexterous.
3
4
23
@erikbryn
Erik Brynjolfsson
2 months
Two things can be true at the same time: 1. Without additional advances, LLMs won't get us to general intelligence. 2. Even without additional advances, LLMs will radically transform the economy.
60
134
759
@RichardSSutton
Richard Sutton
2 months
💯
@chris_hayduk1
Chris Hayduk
2 months
Everyone posting about the Dwarkesh interview (including Dwarkesh himself!) is missing this subtle point. When LLMs imitate, they imitate the ACTION (ie the token prediction to produce the sequence). When humans imitate, they imitate the OUTPUT but must discover the action
28
46
674
@RichardSSutton
Richard Sutton
2 months
Dwarkesh and I had a frank exchange of views. I hope we moved the conversation forward. Dwarkesh is a true gentleman.
@dwarkesh_sp
Dwarkesh Patel
2 months
.@RichardSSutton, father of reinforcement learning, doesn’t think LLMs are bitter-lesson-pilled. My steel man of Richard’s position: we need some new architecture to enable continual (on-the-job) learning. And if we have continual learning, we don't need a special training
82
223
4K
@RichardSSutton
Richard Sutton
2 months
Mike is a powerful thinker and researcher. Very well deserved.
@AmiiThinks
Amii
2 months
Amii Fellow and Canada @CIFAR_News AI Chair Michael Bowling was appointed to Canada's AI Strategy Task Force. We're incredibly proud to see Michael's expertise recognized at this level. Congratulations on a well-deserved appointment! Read: https://t.co/GZwYksNX2O
0
2
103
@KyleCranmer
Kyle Cranmer
2 months
Is scale all you need? Or is there still a role for incorporating domain knowledge and inductive bias? While I was in Heidelberg, I took some time to write a short essay on this question called "The Bittersweet Lesson". https://t.co/DQEItqXomF #HLF25
2
9
113