philippswu Profile Banner
Philipp Wu Profile
Philipp Wu

@philippswu

Followers
2K
Following
2K
Media
50
Statuses
414

PhD @Berkeley_AI advised by @pabbeel. Previously @MetaAI @covariantai.

Berkeley, CA
Joined November 2018
Don't wanna be here? Send us removal request.
@philippswu
Philipp Wu
2 years
๐ŸŽ‰Excited to share a fun little hardware project weโ€™ve been working on. GELLO is an intuitive and low cost teleoperation device for robot arms that costs less than $300. We've seen the importance of data quality in imitation learning. Our goal is to make this more accessible 1/n
26
107
690
@smithlaura1028
Laura Smith
19 hours
Excited to share what we've been brewing at PI! Weโ€™re working on making robots more helpful by making them faster and more reliable through real-world practice, even on delicate behaviors like carrying this very full latte cup
3
18
174
@kevin_zakka
Kevin Zakka
19 days
I spend so much time in MuJoCo I made a Halloween edition ๐ŸŽƒ๐Ÿ‘ป
10
8
198
@kevin_zakka
Kevin Zakka
19 days
Didnโ€™t go max speed for safety and space reasons but pretty happy with the result!
@kevin_zakka
Kevin Zakka
19 days
Coming to mjlab today! This is vanilla RL, no motion imitation/AMP. Natural gaits emerge from minimal rewards: velocity tracking, upright torso, speed-adaptive joint regularization, and contact quality (foot clearance, slip, soft landings). No reference trajectories or gait
4
18
188
@kevin_zakka
Kevin Zakka
25 days
Super happy and honored to be a 2025 Google PhD Fellow! Thank you @Googleorg for believing in my research. I'm looking forward to making humanoid robots more capable and trustworthy partners ๐Ÿค—
@Googleorg
Google.org
26 days
๐ŸŽ‰ We're excited to announce the 2025 Google PhD Fellows! @GoogleOrg is providing over $10 million to support 255 PhD students across 35 countries, fostering the next generation of research talent to strengthen the global scientific landscape. Read more: https://t.co/0Pvuv6hsgP
35
4
196
@alescontrela
Alejandro Escontrela
28 days
Simulation drives robotics progress, but how do we close the reality gap? Introducing GaussGym: an open-source framework for learning locomotion from pixels with ultra-fast parallelized photorealistic rendering across >4,000 iPhone, GrandTour, ARKit, and Veo scenes! Thread ๐Ÿงต
11
64
333
@Jesse_Y_Zhang
Jesse Zhang
1 month
Great to see people already using video rewinding (proposed in https://t.co/oE9qpYUeDV) to improve their reward models! This paper adds stage prediction to better reward long-horizon tasks. Excited to see how much better reward models can get!
@QianzhongChen
Qianzhong Chen
2 months
๐Ÿš€ Introducing SARM: Stage-Aware Reward Modeling for Long-Horizon Robot Manipulation Robots struggle with tasks like folding a crumpled T-shirtโ€”long, contact-rich, and hard to label. We propose a scalable reward modeling framework to fix that. 1/n
1
13
131
@philippswu
Philipp Wu
2 months
We've seen reward models make a massive impact in improving language models with RLHF. Excited to see how reward models can play a similar role in robotics! Follow David's thread for the paper! https://t.co/TMNw7Cpfzr
@QianzhongChen
Qianzhong Chen
2 months
0
0
4
@philippswu
Philipp Wu
2 months
Very excited about this overall research direction. Having such a precise model that can measure task progress enables countless possibilities for improved policy performance. As the saying goes "You can't improve what you don't measure".
1
0
5
@philippswu
Philipp Wu
2 months
One of my favorite results: despite the reward model and policy being trained on the same data, the policy misidentifies the crumpled shirt as folded and pushes the unfinished fold to the corner, where the reward model correctly identifies the failure. https://t.co/R1BVQyZy17
@QianzhongChen
Qianzhong Chen
2 months
๐Ÿ” Robust to OOD + noisy policy rollouts SARM isnโ€™t just for clean human demos. It accurately estimates progress even on noisy, OOD policy rollouts. ๐Ÿ‘‡ a demo where SARM detects recession and not cheated by โ€œfake finishโ€. 3/n
1
0
4
@philippswu
Philipp Wu
2 months
๐ŸŽ†New paper! We explore how to learn precise reward models that can accurately measure task progress for robot manipulation. See @QianzhongChen's thread for more details.
@QianzhongChen
Qianzhong Chen
2 months
๐Ÿš€ Introducing SARM: Stage-Aware Reward Modeling for Long-Horizon Robot Manipulation Robots struggle with tasks like folding a crumpled T-shirtโ€”long, contact-rich, and hard to label. We propose a scalable reward modeling framework to fix that. 1/n
2
6
78
@pabbeel
Pieter Abbeel
2 months
Very excited to start sharing some of the work we have been doing at Amazon FAR. In this work we present OmniRetarget, which can generate high-quality interaction-preserving data from human motions for learning complex humanoid skills. High-quality re-targeting really helps
@zhenkirito123
Zhen Wu
2 months
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐—ข๐—บ๐—ป๐—ถ๐—ฅ๐—ฒ๐˜๐—ฎ๐—ฟ๐—ด๐—ฒ๐˜๐ŸŽฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐—บ๐—ถ๐—ป๐—ถ๐—บ๐—ฎ๐—น RL: - 5 rewards, - 4 DR
8
39
292
@qiayuanliao
Qiayuan Liao
2 months
SOTA data generation from OmniRetarget + SOTA formulation from BeyondMimic = mind-blowing performance
@zhenkirito123
Zhen Wu
2 months
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐—ข๐—บ๐—ป๐—ถ๐—ฅ๐—ฒ๐˜๐—ฎ๐—ฟ๐—ด๐—ฒ๐˜๐ŸŽฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐—บ๐—ถ๐—ป๐—ถ๐—บ๐—ฎ๐—น RL: - 5 rewards, - 4 DR
0
5
40
@brenthyi
Brent Yi
2 months
New project from @kevin_zakka I've been using + helping with! Ridiculously easy setup, typed codebase, headless vis => happy ๐Ÿ˜Š
@kevin_zakka
Kevin Zakka
2 months
I'm super excited to announce mjlab today! mjlab = Isaac Lab's APIs + best-in-class MuJoCo physics + massively parallel GPU acceleration Built directly on MuJoCo Warp with the abstractions you love.
1
3
37
@danijarh
Danijar Hafner
2 months
Excited to introduce Dreamer 4, an agent that learns to solve complex control tasks entirely inside of its scalable world model! ๐ŸŒŽ๐Ÿค– Dreamer 4 pushes the frontier of world model accuracy, speed, and learning complex tasks from offline datasets. co-led with @wilson1yan
82
358
3K
@philippswu
Philipp Wu
2 months
๐Ÿคฏ effortless sim2real, excited to use this!!
@kevin_zakka
Kevin Zakka
2 months
I'm super excited to announce mjlab today! mjlab = Isaac Lab's APIs + best-in-class MuJoCo physics + massively parallel GPU acceleration Built directly on MuJoCo Warp with the abstractions you love.
0
4
25
@kevin_zakka
Kevin Zakka
2 months
I'm super excited to announce mjlab today! mjlab = Isaac Lab's APIs + best-in-class MuJoCo physics + massively parallel GPU acceleration Built directly on MuJoCo Warp with the abstractions you love.
32
140
858
@Ken_Goldberg
Ken Goldberg
2 months
โ€œโ€ฆturns out the โ€˜renderโ€™ subset of simulation is well-matched to the โ€˜quasi-staticโ€™ subset of manipulation.โ€
@mzubairirshad
Zubair Irshad
2 months
Great talk by @uynitsuj presenting the Real2Render2Real paper at CoRL in Seoul! A great collaboration between @UCBerkeley and @ToyotaResearch. The code for this paper is now available, check it out here: https://t.co/ayTC0nPtf4 @ToyotaResearch @AUTOLab_Cal @Ken_Goldberg
0
8
26
@philippswu
Philipp Wu
2 months
Excited! Kevin always delivers๐Ÿš€๐Ÿš€๐Ÿš€
@kevin_zakka
Kevin Zakka
2 months
Meet mjlab. Powered by MuJoCo Warp. Drops Monday.
0
0
27
@philippswu
Philipp Wu
2 months
BLUE!
@oprydai
Mustafa
2 months
3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms 3D print robotic arms
0
0
12