Zhen Wu
@zhenkirito123
Followers
2K
Following
363
Media
8
Statuses
32
Research Intern @ Amazon FAR (Frontier AI & Robotics). CS @Stanford. Humanoid Robots & Character Animation ๐ค
California, USA
Joined April 2022
I've long wondered if we can make a humanoid robot do a ๐๐ฎ๐น๐น๐ณ๐น๐ถ๐ฝ - and we just made it happen by leveraging ๐ข๐บ๐ป๐ถ๐ฅ๐ฒ๐๐ฎ๐ฟ๐ด๐ฒ๐ with BeyondMimic tracking! This came after our original OmniRetarget experiments, with only minor tweaks to RL training: relaxing a
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐ข๐บ๐ป๐ถ๐ฅ๐ฒ๐๐ฎ๐ฟ๐ด๐ฒ๐๐ฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐บ๐ถ๐ป๐ถ๐บ๐ฎ๐น RL: - 5 rewards, - 4 DR
177
554
4K
Open-source: complete codebase covering multiple simulation backends, training, retargeting, and real-world inference. Infra built for humanoid, but also readily modified for quadruped (also included). Lots of infra gems/conveniences we rely on consistently. Hopefully equally
Sim-to-real learning for humanoid robots is a full-stack problem. Today, Amazon FAR is releasing a full-stack solution: Holosoma. To accelerate research, we are open-sourcing a complete codebase covering multiple simulation backends, training, retargeting, and real-world
10
55
468
Sim-to-real learning for humanoid robots is a full-stack problem. Today, Amazon FAR is releasing a full-stack solution: Holosoma. To accelerate research, we are open-sourcing a complete codebase covering multiple simulation backends, training, retargeting, and real-world
19
132
579
Meet BFM-Zero: A Promptable Humanoid Behavioral Foundation Model w/ Unsupervised RL๐ https://t.co/3VdyRWgOqb ๐งฉONE latent space for ALL tasks โกZero-shot goal reaching, tracking, and reward optimization (any reward at test time), from ONE policy ๐คNatural recovery & transition
5
72
253
one of the best moments at BAIR lab actually never imagined to spot prof @redstone_hong on a random day
one of the best moments at BAIR lab actually never imagined to spot prof sergey levine from physical intelligence on a random day
4
5
68
Excited to introduce TWIST2, our next-generation humanoid data collection system. TWIST2 is portable (use anywhere, no MoCap), scalable (100+ demos in 15 mins), and holistic (unlock major whole-body human skills). Fully open-sourced: https://t.co/fAlyD77DEt
22
111
486
How can we standardize conditioning signals in image/video models to achieve the iterative editing & portability that Universal Scene Descriptors provide in graphics? Introducing Neural USD: An object-centric framework for iterative editing & control ๐งต
1
15
80
Simulation drives robotics progress, but how do we close the reality gap? Introducing GaussGym: an open-source framework for learning locomotion from pixels with ultra-fast parallelized photorealistic rendering across >4,000 iPhone, GrandTour, ARKit, and Veo scenes! Thread ๐งต
11
65
338
ResMimic: a two-stage residual framework that unleashes the power of pre-trained general motion tracking policy. Enable expressive whole-body loco-manipulation with payloads up to 5.5kg without task-specific design, generalize across poses, and exhibit reactive behavior.
10
72
296
We are open-sourcing over 4 hours of high-quality, retargeted trajectories! Website: https://t.co/EeW71PaXVI ArXiv: https://t.co/8jxK1svcmT Datasets: https://t.co/xUq8seOxtM Huge shout out to the amazing team: @lujieyang98, @x_h_ucb, @akanazawa, @pabbeel, @carlo_sferrazza,
huggingface.co
3
6
64
Standing on the shoulders of giants! Our work builds on amazing research in the community๐ก. We use the "interaction mesh" ๐ธ๏ธ [1], [2] to preserve spatial relationships and leverage the minimal RL formulation from works like BeyondMimic [3]. Our long-horizon sequence is a nod to
1
1
38
Our grand finale: A complex, long-horizon dynamic sequence, all driven by a proprioceptive-only policy (no vision/LIDAR)! In this task, the robot carries a chair to a platform, uses it as a step to climb up, then leaps off and performs a parkour-style roll to absorb the landing.
5
28
155
But how much better is our data? ๐ค Compared to widely-used baselines, our motions show far fewer physical artifactsโvirtually zero foot-skating and penetrationโwhile better preserving contact. This allows us to use an open-sourced RL framework (BeyondMimic) without
2
0
33
And it's not just for a specific robot! Our framework is highly general and adapts to different robot embodiments, including the @UnitreeRobotics H1 and the @boosterobotics T1. We can retarget complex object-carrying and platform-climbing skills across these different robots with
1
0
34
What about scalability? OmniRetarget transforms a SINGLE human demo into diverse motion clips. We can systematically vary terrain height, object size, and initial poses. Best of all, these augmented skills transfer directly from sim to our real-world hardware! ๐คโก๏ธ๐ฆพ 4/9
1
1
36
The result of this high-quality data? We can train diverse skills like box carrying ๐ฆ, slope crawling ๐พ, and platform climbing ๐ง with a radically simplified RL process! All policies use just 5 reward terms, achieving successful zero-shot sim-to-real transfer! ๐ฏโก๏ธ๐ฆพ 3/9
1
0
55
Existing retargeting often produces artifacts like foot-skating and penetration โ. To compensate, RL policies rely on complex ad-hoc reward terms, forcing a trade-off between accurate motion tracking and correcting errors like slipping or bad contacts. OmniRetarget fixes this
3
4
73
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐ข๐บ๐ป๐ถ๐ฅ๐ฒ๐๐ฎ๐ฟ๐ด๐ฒ๐๐ฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐บ๐ถ๐ป๐ถ๐บ๐ฎ๐น RL: - 5 rewards, - 4 DR
31
156
673
Excited to share that I'll be joining @UTAustin in Fall 2026 as an Assistant Professor with @utmechengr @texas_robotics! I'm looking for PhD students interested in humanoids, dexterous manipulation, tactile sensing, and robot learning in general -- consider applying this cycle!
50
38
464
We're hiring interns (and full-times) all year long! Please email me if interested.
41
85
2K