Zhen Wu Profile
Zhen Wu

@zhenkirito123

Followers
2K
Following
363
Media
8
Statuses
32

Research Intern @ Amazon FAR (Frontier AI & Robotics). CS @Stanford. Humanoid Robots & Character Animation ๐Ÿค–

California, USA
Joined April 2022
Don't wanna be here? Send us removal request.
@zhenkirito123
Zhen Wu
2 months
I've long wondered if we can make a humanoid robot do a ๐˜„๐—ฎ๐—น๐—น๐—ณ๐—น๐—ถ๐—ฝ - and we just made it happen by leveraging ๐—ข๐—บ๐—ป๐—ถ๐—ฅ๐—ฒ๐˜๐—ฎ๐—ฟ๐—ด๐—ฒ๐˜ with BeyondMimic tracking! This came after our original OmniRetarget experiments, with only minor tweaks to RL training: relaxing a
@zhenkirito123
Zhen Wu
3 months
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐—ข๐—บ๐—ป๐—ถ๐—ฅ๐—ฒ๐˜๐—ฎ๐—ฟ๐—ด๐—ฒ๐˜๐ŸŽฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐—บ๐—ถ๐—ป๐—ถ๐—บ๐—ฎ๐—น RL: - 5 rewards, - 4 DR
177
554
4K
@pabbeel
Pieter Abbeel
18 days
Open-source: complete codebase covering multiple simulation backends, training, retargeting, and real-world inference. Infra built for humanoid, but also readily modified for quadruped (also included). Lots of infra gems/conveniences we rely on consistently. Hopefully equally
@carlo_sferrazza
Carlo Sferrazza
18 days
Sim-to-real learning for humanoid robots is a full-stack problem. Today, Amazon FAR is releasing a full-stack solution: Holosoma. To accelerate research, we are open-sourcing a complete codebase covering multiple simulation backends, training, retargeting, and real-world
10
55
468
@carlo_sferrazza
Carlo Sferrazza
18 days
Sim-to-real learning for humanoid robots is a full-stack problem. Today, Amazon FAR is releasing a full-stack solution: Holosoma. To accelerate research, we are open-sourcing a complete codebase covering multiple simulation backends, training, retargeting, and real-world
19
132
579
@li_yitang
Yitang Li
1 month
Meet BFM-Zero: A Promptable Humanoid Behavioral Foundation Model w/ Unsupervised RL๐Ÿ‘‰ https://t.co/3VdyRWgOqb ๐ŸงฉONE latent space for ALL tasks โšกZero-shot goal reaching, tracking, and reward optimization (any reward at test time), from ONE policy ๐Ÿค–Natural recovery & transition
5
72
253
@arthurallshire
Arthur Allshire
1 month
one of the best moments at BAIR lab actually never imagined to spot prof @redstone_hong on a random day
@himanshustwts
himanshu
1 month
one of the best moments at BAIR lab actually never imagined to spot prof sergey levine from physical intelligence on a random day
4
5
68
@ZeYanjie
Yanjie Ze
1 month
Excited to introduce TWIST2, our next-generation humanoid data collection system. TWIST2 is portable (use anywhere, no MoCap), scalable (100+ demos in 15 mins), and holistic (unlock major whole-body human skills). Fully open-sourced: https://t.co/fAlyD77DEt
22
111
486
@alescontrela
Alejandro Escontrela
2 months
How can we standardize conditioning signals in image/video models to achieve the iterative editing & portability that Universal Scene Descriptors provide in graphics? Introducing Neural USD: An object-centric framework for iterative editing & control ๐Ÿงต
1
15
80
@alescontrela
Alejandro Escontrela
2 months
Simulation drives robotics progress, but how do we close the reality gap? Introducing GaussGym: an open-source framework for learning locomotion from pixels with ultra-fast parallelized photorealistic rendering across >4,000 iPhone, GrandTour, ARKit, and Veo scenes! Thread ๐Ÿงต
11
65
338
@SihengZhao
Siheng Zhao
2 months
ResMimic: a two-stage residual framework that unleashes the power of pre-trained general motion tracking policy. Enable expressive whole-body loco-manipulation with payloads up to 5.5kg without task-specific design, generalize across poses, and exhibit reactive behavior.
10
72
296
@zhenkirito123
Zhen Wu
3 months
We are open-sourcing over 4 hours of high-quality, retargeted trajectories! Website: https://t.co/EeW71PaXVI ArXiv: https://t.co/8jxK1svcmT Datasets: https://t.co/xUq8seOxtM Huge shout out to the amazing team: @lujieyang98, @x_h_ucb, @akanazawa, @pabbeel, @carlo_sferrazza,
Tweet card summary image
huggingface.co
3
6
64
@zhenkirito123
Zhen Wu
3 months
Standing on the shoulders of giants! Our work builds on amazing research in the community๐Ÿ’ก. We use the "interaction mesh" ๐Ÿ•ธ๏ธ [1], [2] to preserve spatial relationships and leverage the minimal RL formulation from works like BeyondMimic [3]. Our long-horizon sequence is a nod to
1
1
38
@zhenkirito123
Zhen Wu
3 months
Our grand finale: A complex, long-horizon dynamic sequence, all driven by a proprioceptive-only policy (no vision/LIDAR)! In this task, the robot carries a chair to a platform, uses it as a step to climb up, then leaps off and performs a parkour-style roll to absorb the landing.
5
28
155
@zhenkirito123
Zhen Wu
3 months
But how much better is our data? ๐Ÿค” Compared to widely-used baselines, our motions show far fewer physical artifactsโ€”virtually zero foot-skating and penetrationโ€”while better preserving contact. This allows us to use an open-sourced RL framework (BeyondMimic) without
2
0
33
@zhenkirito123
Zhen Wu
3 months
And it's not just for a specific robot! Our framework is highly general and adapts to different robot embodiments, including the @UnitreeRobotics H1 and the @boosterobotics T1. We can retarget complex object-carrying and platform-climbing skills across these different robots with
1
0
34
@zhenkirito123
Zhen Wu
3 months
What about scalability? OmniRetarget transforms a SINGLE human demo into diverse motion clips. We can systematically vary terrain height, object size, and initial poses. Best of all, these augmented skills transfer directly from sim to our real-world hardware! ๐Ÿค–โžก๏ธ๐Ÿฆพ 4/9
1
1
36
@zhenkirito123
Zhen Wu
3 months
The result of this high-quality data? We can train diverse skills like box carrying ๐Ÿ“ฆ, slope crawling ๐Ÿพ, and platform climbing ๐Ÿง— with a radically simplified RL process! All policies use just 5 reward terms, achieving successful zero-shot sim-to-real transfer! ๐ŸŽฏโžก๏ธ๐Ÿฆพ 3/9
1
0
55
@zhenkirito123
Zhen Wu
3 months
Existing retargeting often produces artifacts like foot-skating and penetration โŒ. To compensate, RL policies rely on complex ad-hoc reward terms, forcing a trade-off between accurate motion tracking and correcting errors like slipping or bad contacts. OmniRetarget fixes this
3
4
73
@zhenkirito123
Zhen Wu
3 months
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing ๐—ข๐—บ๐—ป๐—ถ๐—ฅ๐—ฒ๐˜๐—ฎ๐—ฟ๐—ด๐—ฒ๐˜๐ŸŽฏ, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with ๐—บ๐—ถ๐—ป๐—ถ๐—บ๐—ฎ๐—น RL: - 5 rewards, - 4 DR
31
156
673
@carlo_sferrazza
Carlo Sferrazza
3 months
Excited to share that I'll be joining @UTAustin in Fall 2026 as an Assistant Professor with @utmechengr @texas_robotics! I'm looking for PhD students interested in humanoids, dexterous manipulation, tactile sensing, and robot learning in general -- consider applying this cycle!
50
38
464
@rocky_duan
Rocky Duan
4 months
We're hiring interns (and full-times) all year long! Please email me if interested.
41
85
2K