
Wenlong Huang
@wenlong_huang
Followers
4K
Following
3K
Media
33
Statuses
555
PhD Student @StanfordSVL @StanfordAILab. Previously @Berkeley_AI @GoogleDeepMind @NVIDIARobotics. Robotics, Foundation Models.
Stanford, CA
Joined May 2019
What structural task representation enables multi-stage, in-the-wild, bimanual, reactive manipulation?. Introducing ReKep: LVM to label keypoints & VLM to write keypoint-based constraints, solve w/ optimization for diverse tasks, w/o task-specific training or env models. 🧵👇
18
105
513
RT @lucas_manuelli: Today I’m proud to share what I’ve been working on recently with my team at @BostonDynamics along with our collaborator….
bostondynamics.com
Useful humanoid robots will require a long list of competencies. They will need the ability to manipulate a diverse range of objects (e.g. hard/soft, heavy/delicate, rigid/articulated, large/small),...
0
24
0
RT @YunfanJiang: Our @StanfordSVL's BEHAVIOR Bot got a package 📦 from Jensen today. What’s inside? It’s ready to unbox faster than you c….
0
4
0
RT @BaldassarreFe: Say hello to DINOv3 🦖🦖🦖. A major release that raises the bar of self-supervised vision foundation models. With stunning….
0
278
0
RT @DJiafei: Reasoning is central to purposeful action. Today we introduce MolmoAct — a fully open Action Reasoning Model (ARM) for robotic….
0
66
0
RT @BDuisterhof: Missed our #RSS workshop on structured world models for robot manipulation?🦾 Or want to rewatch 📷 your favorite talks? We….
youtube.com
https://swomo-rss.github.io/ Physics-based models have been crucial for manipulation, enabling sim-to-real learning, model-predictive control, manipulation p...
0
1
0
RT @ManlingLi_: 🏆Thrilled to receive @aclmeeting Inaugural Dissertation Award Honorable Mention. “Multimodality” has moved incredibly fas….
0
15
0
RT @neil_nie_: Thank you for sharing our work @ycombinator! Please checkout to schedule a free pilot to see how our….
vernerobotics.com
Verne Robotics provides AI software for robotic arms with no cap-ex. Physical AI within reach. Train in hours and automate today.
0
12
0
RT @mihirp98: 🚨 The era of infinite internet data is ending, So we ask:. 👉 What’s the right generative modelling objective when data—not co….
0
183
0
RT @ManlingLi_: Excited that @RuohanZhang76 is joining NU @northwesterncs ! If you are thinking about pursuing a PhD, definitely reach out….
0
5
0
RT @binghao_huang: Tactile interaction in the wild can unlock fine-grained manipulation! 🌿🤖✋. We built a portable handheld tactile gripper….
0
52
0
RT @RussTedrake: TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: .
0
107
0
RT @shivanshpatel35: 🚀 Introducing RIGVid: Robots Imitating Generated Videos!.Robots can now perform complex tasks—pouring, wiping, mixing—….
0
33
0
Exciting to see more works leveraging VLM-inferred keypoints as a bridge between semantic knowledge and low-level behaviors, especially those dexterous skills 🤩.
We find keypoint trajectories to be a powerful interface between VLM planning & RL control.VLM: Generates object + hand motion plan from a task prompt & RGB-D image (perception + commonsense).RL policy: Conditioned on the plan, learns low-level dexterous control (0-shot sim2real).
0
0
11
RT @ManlingLi_: Can VLMs build Spatial Mental Models like humans?. Reasoning from limited views?.Reasoning from partial observations?.Reaso….
0
58
0
RT @YuXiang_IRVL: “As a PHD student, your job is not publishing a paper every quarter. Focus on a problem in deep understanding and solve i….
0
80
0
RT @charles_rqi: Tesla Robotaxi: A New Era Begins. I’ve (very fortunately) been part of multiple robotaxi launches. But this one is differe….
0
449
0
RT @gan_chuang: Attending RSS for the first time and giving a talk tomorrow at the Learning Structured World Models for Robotic Manipulatio….
0
9
0
Join us tomorrow in SGM 124 for the SWOMO workshop at #RSS2025! We will have 6 amazing talks and a panel in the end to discuss structured world modeling for robotics!. Latest schedule and information at
Excited to announce the “Structured World Models for Robotic Manipulation” workshop at #RSS2025 in LA!. Website: Call for Papers (Deadline: May 23): Come join us with a stellar lineup of speakers to discuss the various important &
4
11
50
RT @kaiwynd: Can we learn a 3D world model that predicts object dynamics directly from videos? . Introducing Particle-Grid Neural Dynamics….
0
34
0
RT @YXWangBot: 🤖 Does VLA models really listen to language instructions? Maybe not 👀.🚀 Introducing our RSS paper: CodeDiffuser -- using VLM….
0
27
0