Joel Jang Profile
Joel Jang

@jang_yoel

Followers
2K
Following
2K
Media
50
Statuses
369

Senior Research Scientist @nvidiaai GEAR Lab, world modeling lead. On leave from PhD at @uwcse

Seattle, US
Joined March 2021
Don't wanna be here? Send us removal request.
@jang_yoel
Joel Jang
5 months
Introducing πƒπ«πžπšπ¦π†πžπ§! We got humanoid robots to perform totally new π‘£π‘’π‘Ÿπ‘π‘  in new environments through video world models. We believe video world models will solve the data problem in robotics. Bringing the paradigm of scaling human hours to GPU hours. Quick 🧡
9
78
382
@NVIDIARobotics
NVIDIA Robotics
1 month
The rise of humanoid platforms presents new opportunities and unique challenges. πŸ€– Join @yukez at #CoRL2025 as he shares the latest research on robot foundation models and presents new updates with the #NVIDIAIsaac GR00T platform. Learn more πŸ‘‰ https://t.co/LrzONs1Gzc
49
29
157
@RoboPapers
RoboPapers
2 months
Full episode dropping soon! Geeking out with @jang_yoel on DreamGen - Unlocking Generalization in Robot Learning through Video World Models https://t.co/4GkmxHMqSW Co-hosted by @chris_j_paxton @micoolcho
2
7
31
@elonmusk
Elon Musk
2 months
@DrJimFan Tesla has this too for Optimus. As you say, it is essential for humanoid robot training.
90
168
2K
@DrJimFan
Jim Fan
2 months
World modeling for robotics is incredibly hard because (1) control of humanoid robots & 5-finger hands is wayyy harder than β¬†οΈβ¬…οΈβ¬‡οΈβž‘οΈ in games (Genie 3); and (2) object interaction is much more diverse than FSD, which needs to *avoid* coming into contact. Our GR00T Dreams work was
@DrJimFan
Jim Fan
5 months
What if robots could dream inside a video generative model? Introducing DreamGen, a new engine that scales up robot learning not with fleets of human operators, but with digital dreams in pixels. DreamGen produces massive volumes of neural trajectories - photorealistic robot
36
171
1K
@TheHumanoidHub
The Humanoid Hub
3 months
A humanoid robot policy trained solely on synthetic data generated by a world model. Research Scientist Joel Jang presents NVIDIA's DreamGen pipeline: β¦Ώ Post-train the world model Cosmos-Predict2 with a small set of real teleoperation demos. β¦Ώ Prompt the world model to
10
43
221
@DrJimFan
Jim Fan
3 months
I've been a bit quiet on X recently. The past year has been a transformational experience. Grok-4 and Kimi K2 are awesome, but the world of robotics is a wondrous wild west. It feels like NLP in 2018 when GPT-1 was published, along with BERT and a thousand other flowers that
183
328
4K
@jang_yoel
Joel Jang
4 months
Check out Cosmos-Predict2, a new SOTA video world model trained specifically for Physical AI (powering GR00T Dreams & DreamGen)!
@hanna_mao
Hanzi Mao
4 months
We build Cosmos-Predict2 as a world foundation model for Physical AI builders β€” fully open and adaptable. Post-train it for specialized tasks or different output types. Available in multiple sizes, resolutions, and frame rates. πŸ“· Watch the repo walkthrough
0
6
44
@zhengyiluo
Zhengyi β€œZen” Luo
4 months
Nvidia GEAR RSS 2025 Squad Rolling Out
11
6
151
@jang_yoel
Joel Jang
4 months
πŸš€ GR00T Dreams code is live! NVIDIA GEAR Lab's open-source solution for robotics data via video world models. Fine-tune on any robot, generate 'dreams', extract actions with IDM, and train visuomotor policies with LeRobot datasets (GR00T N1.5, SmolVLA). https://t.co/7Fndn7zDJB
Tweet card summary image
github.com
Nvidia GEAR Lab's initiative to solve the robotics data problem using world models - NVIDIA/GR00T-Dreams
@jang_yoel
Joel Jang
5 months
Introducing πƒπ«πžπšπ¦π†πžπ§! We got humanoid robots to perform totally new π‘£π‘’π‘Ÿπ‘π‘  in new environments through video world models. We believe video world models will solve the data problem in robotics. Bringing the paradigm of scaling human hours to GPU hours. Quick 🧡
6
44
151
@youliangtan
youliang
4 months
How we improve VLA generalization? πŸ€” Last week we upgraded #NVIDIA GR00T N1.5 with minor VLM tweaks, FLARE, and richer data mixtures (DreamGen, etc.) ✨. N1.5 yields better language following β€” post-trained on unseen Unitree G1 with 1K trajectories, it follows commands on
2
23
187
@qsh_zh
Qinsheng Zhang
4 months
πŸš€ Introducing Cosmos-Predict2! Our most powerful open video foundation model for Physical AI. Cosmos-Predict2 significantly improves upon Predict1 in visual quality, prompt alignment, and motion dynamicsβ€”outperforming popular open-source video foundation models. It’s openly
7
62
204
@chris_j_paxton
Chris Paxton
4 months
Assuming that we need ~2 trillion tokens to get to a robot GPT, how can we get there? I went through a few scenarios looking at how we can combine simulation data, human video data, and looking at the size of existing robot fleets. Some assumptions: - We probably need some real
12
35
214
@AiYiyangZ
Yiyang Zhou
4 months
πŸ”₯ ReAgent-V Released! πŸ”₯ A unified video framework with reflection and reward-driven optimization. ✨ Real-time self-correction. ✨ Triple-view reflection. ✨ Auto-selects high-reward samples for training.
1
18
44
@jang_yoel
Joel Jang
4 months
Giving a talk about GR00T N1, GR00T N1.5, and GR00T Dreams in NVIDIA GTC Paris 06.11 2PM - 2:45PM CEST. If you are at Vivatech in Paris, please stop by the "An Introduction to Humanoid Robotics" Session!
@NVIDIARobotics
NVIDIA Robotics
4 months
Are you curious about #humanoidrobotics? Join our experts at #GTCParis for a deep dive into the #NVIDIAIsaac GR00T platform and its four pillars: 🧠 Robot foundation models for cognition and control 🌐 Simulation frameworks built on @nvidiaomniverse and #NVIDIACosmos πŸ“Š Data
1
6
63
@ruijie_zheng12
Ruijie Zheng
4 months
Representation also matters for VLA models! Introducing FLARE: Robot Learning with Implicit World Modeling. With future latent alignment objective, FLARE significantly improves policy performance on multitask imitation learning & unlocks learning from egocentric human videos.
6
24
118
@adcock_brett
Brett Adcock
5 months
Nvidia also announced DreamGen, a new engine that scales robot learning with digital dreams It produces large volumes of photorealistic robot videos (using video models) paired with motor action labels and unlocks generalization to new environments https://t.co/rWTboFmM7z
3
7
94
@TheHumanoidHub
The Humanoid Hub
5 months
NVIDIA has published a paper on DREAMGEN – a powerful 4-step pipeline for generating synthetic data for humanoids that enables task and environment generalization. - Step 1: Fine-tune a video generation model using a small number of human teleoperation videos - Step 2: Prompt
2
32
158
@snasiriany
Soroush Nasiriany
5 months
It’s not a matter of if, it’s a matter of when, video models and world models are going to be a central tool for building robot foundation models.
@jang_yoel
Joel Jang
5 months
Introducing πƒπ«πžπšπ¦π†πžπ§! We got humanoid robots to perform totally new π‘£π‘’π‘Ÿπ‘π‘  in new environments through video world models. We believe video world models will solve the data problem in robotics. Bringing the paradigm of scaling human hours to GPU hours. Quick 🧡
0
1
11