Ruihan Yang Profile
Ruihan Yang

@RchalYang

Followers
2K
Following
688
Media
19
Statuses
189

Applied Scientist @ Amazon Frontier AI & Robotics (FAR) PhD from @UCSanDiego Robot Learning / Embodied AI

San Diego, CA
Joined July 2017
Don't wanna be here? Send us removal request.
@RchalYang
Ruihan Yang
1 year
At IROS 2024 now. I’ll present our work HarmonicMM tomorrow 10AM at WeAT2 Also open to all kinds of discussion! Let me know if you’d like to chat!
@RchalYang
Ruihan Yang
2 years
How to tackle complex household tasks like opening doors, and cleaning tables in real? Introducing HarmonicMM: Our latest model seamlessly combines navigation and manipulation, enabling robots to tackle household tasks using only RGB visual observation and robot proprioception.
1
6
45
@cheryyun_l
Yongyuan Liang
5 days
Unified multimodal models can generate text and images, but can they truly reason across modalities? 🎨 Introducing ROVER, the first benchmark that evaluates reciprocal cross-modal reasoning in unified models, the next frontier of omnimodal intelligence. 🌐 Project:
5
27
285
@RchalYang
Ruihan Yang
13 days
I recently learned that my friend has been trying to scare Waymo by jumping in front of it deliberately out of nowhere. (He's safe and sound for now) I'm really curious whether this type of data is included in training. 😅
@WaymoCommunity
Waymo community
25 days
每個人都應該安全咁分享道路。了解下Waymo點��努力令我哋嘅道路更安全。
0
0
0
@RchalYang
Ruihan Yang
1 month
Lucky enough to see live demos….🤯🤯🤯
@zhenkirito123
Zhen Wu
1 month
Humanoid motion tracking performance is greatly determined by retargeting quality! Introducing 𝗢𝗺𝗻𝗶𝗥𝗲𝘁𝗮𝗿𝗴𝗲𝘁🎯, generating high-quality interaction-preserving data from human motions for learning complex humanoid skills with 𝗺𝗶𝗻𝗶𝗺𝗮𝗹 RL: - 5 rewards, - 4 DR
0
1
9
@RchalYang
Ruihan Yang
1 month
Residual RL for pretrained policies at ease in real world by amazing @larsankile . Check it out!
@larsankile
Lars Ankile
1 month
How can we enable finetuning of humanoid manipulation policies, directly in the real world? In our new paper, Residual Off-Policy RL for Finetuning BC Policies, we demonstrate real-world RL on a bimanual humanoid with 5-fingered hands (29 DoF) and improve pre-trained policies
0
1
29
@RchalYang
Ruihan Yang
2 months
Over the past few years, a lot of progress (not just robot learning) has come from working on somewhat similar hardware — making it much easier to share knowledge. Of course, if NVIDIA is shipping actually humanoid, i would like to see how it works.
@yukez
Yuke Zhu
2 months
People who are really serious about robot learning should make their own robot hardware.
1
0
18
@rocky_duan
Rocky Duan
3 months
We're hiring interns (and full-times) all year long! Please email me if interested.
41
86
2K
@RchalYang
Ruihan Yang
3 months
My personal opinion: mobile aloha (Bimanual + Wheel) / Vega (Bimanual + dexhand + wheel) / Digit (Bimanual + Wheel) / Optimus (obviously) are humanoids.
1
0
7
@RchalYang
Ruihan Yang
3 months
Turns out, when we discuss “humanoid robot” everyone’s picturing something totally different. So I made this figure, and next time, i'll show this, before discussion.
15
19
195
@RchalYang
Ruihan Yang
4 months
We evaluate EgoVLA on our Ego Humanoid Manipulation Benchmark * Human pretraining improves performance across both short- & long-horizon tasks * Fine-tuned EgoVLA outperforms baselines, especially on challenging, multi-step behaviors * Pretraining boosts generalization to
1
1
9
@RchalYang
Ruihan Yang
4 months
To enable reproducible, scalable evaluation, we introduce Ego Humanoid Manipulation Benchmark — a diverse humanoid manipulation benchmark using Isaac Lab and a testbed for manipulation policy generalization. • 12 tasks: from atomic to multi-stage skills • 25 visual background
1
1
7
@RchalYang
Ruihan Yang
4 months
At its core, EgoVLA leverages a unified human-robot action space built on the MANO hand model. We retarget robot hand motions into MANO space, allowing human and robot actions to be represented identically. During deployment, EgoVLA predicts MANO wrist + hand motion from video.
2
0
7
@RchalYang
Ruihan Yang
4 months
EgoVLA learns manipulation by predicting future wrist & hand motion from diverse egocentric human videos across different backgrounds and tasks. It uses a vision-language backbone (NVILA-2B) and an action head to model both perception and control: * Inputs: RGB history, language
1
2
14
@RchalYang
Ruihan Yang
4 months
How can we leverage diverse human videos to improve robot manipulation? Excited to introduce EgoVLA — a Vision-Language-Action model trained on egocentric human videos by explicitly modeling wrist & hand motion. We build a shared action space between humans and robots, enabling
6
72
491
@RchalYang
Ruihan Yang
5 months
that means you are so lucky to deeply understand multiple problems during your phd.
@_wenlixiao
Wenli Xiao
5 months
ummm… As a robotics PhD student, I’m genuinely worried that the problem I find important now will be solved in the next 2 years—by MORE DATA, without any need to understand the underlying structure. And this happens in many areas😂
0
0
5
@RchalYang
Ruihan Yang
5 months
When it comes to scaling data, it’s not just about scale—it’s also about distribution. Leveraging generative models, even simple ones, can help improve both. Great work led by @jianglong_ye & @kaylee_keyi!
@jianglong_ye
Jianglong Ye
5 months
How to generate billion-scale manipulation demonstrations easily? Let us leverage generative models! 🤖✨ We introduce Dex1B, a framework that generates 1 BILLION diverse dexterous hand demonstrations for both grasping 🖐️and articulation 💻 tasks using a simple C-VAE model.
1
2
7
@RchalYang
Ruihan Yang
5 months
Thank you @xiaolonw for all the support and guidance over the past six years! It’s been a truly transformative experience, and I’m so grateful for everything I’ve learned along the way. Hard to believe this chapter is coming to a close.
@xiaolonw
Xiaolong Wang
5 months
Congratulations to the graduation of @Jerry_XU_Jiarui @JitengMu @RchalYang @YinboChen ! I am excited for their future journeys in industry: Jiarui -> OpenAI Jiteng -> Adobe Ruihan -> Amazon Yinbo -> OpenAI
7
1
57
@yswhynot
yisha
6 months
For years, I’ve been tuning parameters for robot designs and controllers on specific tasks. Now we can automate this on dataset-scale. Introducing Co-Design of Soft Gripper with Neural Physics - a soft gripper trained in simulation to deform while handling load.
7
38
131
@RchalYang
Ruihan Yang
6 months
Great progress by Optimus
@Tesla_Optimus
Tesla Optimus
6 months
I’m not just dancing all day, ok
0
0
2
@rocky_duan
Rocky Duan
6 months
Our robotics team will be at ICRA next week in Atlanta! Having started a new research team at Amazon building robot foundation models, we're hiring across all levels, full-time or intern, and across both SW and Research roles. Ping me at drockyd@amazon.com and let's have a chat!
2
21
201