Luke Rowe
@Luke22R
Followers
146
Following
966
Media
9
Statuses
81
PhD student at @Mila_Quebec, focusing on autonomous driving. Previously @Waymo and @torc_robotics.
Montréal, Québec
Joined February 2022
Waymo is coming to London next year.
waymo.com
Waymo is expanding to London, with plans to offer rides starting in 2026
0
4
44
Waymos are 80% less likely to get into a serious crash than human drivers. An 80% reduction in car crash deaths in the US would mean more lives saved than if you eliminated all homicides. Great piece by @KelseyTuoc
theargumentmag.com
Self-driving cars are way safer than human drivers
24
105
573
“GRPO” is just rebranded REINFORCE. Everything “unique” about GRPO like the advantage normalization and (biased) KL regularization are pretty much useless. Kill GRPO. It’s always been REINFORCE.
7
11
185
NO verifiers. NO Tools. Qwen3-4B-Instruct can match DeepSeek-R1 and o3-mini (high) with ONLY test-time scaling. Presenting Recursive Self-Aggregation (RSA) — the strongest test-time scaling method I know of! Then we use aggregation-aware RL to push further!! 📈📈 🧵below!
22
102
786
🚨Announcing the World Modeling Workshop 2026 🚨 📅 When: Feb 4–6, 2026 📍Where: Mila (Montréal) + Online (free) 💡 What: Keynotes, Methods Deep Dive, and Tutorials 🌐 https://t.co/WukFtNON3o ✉️ worldmodel.mila@gmail.com 🧵 Details below:
6
59
239
@jxmnop Origin of most of these innovations is Canada 🇨🇦 though 😜
2
1
37
🚨 Excited to share our new work: "Stable Gradients for Stable Learning at Scale in Deep Reinforcement Learning"! 📈 We propose gradient interventions that enable stable, scalable learning, achieving significant performance gains across agents and environments! Details below 👇
2
36
171
How can we make behavioural cloning (BC) achieve better combinatorial generalization on out-of-distribution goals? We propose BYOL-γ: an auxiliary self-predictive loss to improve generalization for goal-conditioned BC. 🧵1/6
1
15
74
Poutine: Vision-Language-Trajectory Pre-Training and Reinforcement Learning Post-Training Enable Robust End-to-End Autonomous Driving.
arxiv.org
Maintaining good driving behavior in out-of-distribution scenarios remains a critical challenge in autonomous driving. A promising direction is to leverage the generalist knowledge and reasoning...
0
2
14
Why do we keep sampling from the same distribution the model was trained on? We rethink this old paradigm by introducing Feynman-Kac Correctors (FKCs) – a flexible framework for controlling the distribution of samples at inference time in diffusion models! Without re-training
arxiv.org
While score-based generative models are the model of choice across diverse domains, there are limited tools available for controlling inference-time behavior in a principled manner, e.g. for...
🧵(1/6) Delighted to share our @icmlconf 2025 spotlight paper: the Feynman-Kac Correctors (FKCs) in Diffusion Picture this: it’s inference time and we want to generate new samples from our diffusion model. But we don’t want to just copy the training data – we may want to sample
1
28
136
One of the most striking, non-text AI plots I've seen since ChatGPT launched. Scaling keeps working, this time for Waymo's tooling.
4
22
189
This was joint work with my amazing colleagues at @Mila_Quebec: Rodrigue de Schaetzen, @rogg1111 , @chrisjpal , @duckietown_coo Check out our report here:
0
0
0
Why did Poutine work? • Plug-and-play VLM – Built on Qwen 2.5 VL 3B. No custom perception backbone or action headers needed. • Simple and effective training recipe – Self-supervised vision-language-trajectory pre-training followed by lightweight RL preference fine-tuning.
1
0
0
This challenge pushed the limits of vision-based end-to-end planning in rare, long-tail scenarios. We show that VLMs can be repurposed into effective planners in the long-tail.
1
0
0
Excited that our paper "Addressing Concept Mislabeling in Concept Bottleneck Models Through Preference Optimization" was accepted to ICML 2025! We show how Preference Optimization can reduce the impact of noisy concept labels in CBMs. 🧵/9
1
23
36
(1/n)🚨You can train a model solving DFT for any geometry almost without training data!🚨 Introducing Self-Refining Training for Amortized Density Functional Theory — a variational framework for learning a DFT solver that predicts the ground-state solutions for different
3
40
157
New preprint! 🧠🤖 How do we build neural decoders that are: ⚡️ fast enough for real-time use 🎯 accurate across diverse tasks 🌍 generalizable to new sessions, subjects, and species? We present POSSM, a hybrid SSM architecture that optimizes for all three of these axes! 🧵1/7
4
25
60
🚗💥Introducing Ctrl-Crash: controllable video generation for autonomous driving! SOTA models struggle to generate physically realistic car crashes. We propose an image2video diffusion model with bounding box and crash type control. Website: https://t.co/vNBYhbx3c4 🧵->
2
13
23