
Ankit Goyal
@imankitgoyal
Followers
2K
Following
230
Media
33
Statuses
167
Research Scientist @Nvidia | Developing Foundation Models for Robotics | Ph.D. @PrincetonCS | Previously: @UMich, @IITKanpur
Seattle, WA
Joined March 2020
Super excited to introduce š¹ HAMSTERāour VLA model accepted to #ICLR2025! š. We train a VLM to predict high-level trajectories, executed by a 3D policy like RVT-2. This decomposition enables better generalization and efficient use of off-domain cheap data. Check out.
š Meet š¹HAMSTER, our new hierarchical Vision-Language-Action (VLA) framework for robot manipulation!. š¹ High-level VLM for perception & reasoning.š¹ Low-level 3D policy for precise control.š¹ Bridged by 2D paths for trajectory planning.HAMSTER learns from cost-effective
1
1
22
4. Flowing from Words to Pixels. An insight that seems so simple in hindsight. For conditional generation, instead of starting from noise, why not flow directly from source to the target distri.? I'll be watching closely if this becomes the norm. Great Work by @Qihao Liu et al.
0
0
2
2. Reconstructing Animals and the Wild. This work generates complete scenes from natural images, trained with just synthetic Infinigen data. While working on Infinigen, I never thought it could be used so creatively. Fantastic work by Peter @Michael_J_Black @silvia_zuffi
2
0
2
Thatās a wrap for #CVPR2025! Here's a š§µ of some really cool works š. 1. Let Humanoids Hike!. Great work @ky_lin0305 and Stella Xu. They drove home the point that we can't treat locomotion and navigation as separate. The ultimate test: Can your robot complete a hike on its own?
1
0
12
Headed to Nashville for #CVPR2025! . Excited to present our robot learning work, "3D-MVP," and connect with fellow researchers. I'll be sharing my thoughts on the most exciting papers. Should be a fun week of learning and meeting friends, old and new! #Computervision #robotics
0
0
10
RT @Ishika_S_: VLAs have the potential to generalize over scenes and tasks, but require a ton of data to learn robust policies. We introduā¦.
0
23
0
RT @NVIDIARobotics: š That's a wrap on #ICLR2025. Shout out to all the amazing research in #robotics, machine vision, and more. Missed itā¦.
0
20
0
3D-MVP has been accepted to #CVPR2025! . Our research demonstrates that 3D pretraining significantly outperforms other pretraining approaches in robotics. By integrating this pretraining into the Robotic View Transformer (RVT), we enhance downstream policy performance.
Exciting News! Our new paper "3D-MVP" is out! We propose a novel approach for 3D multi-view pretraining using masked autoencoders, leveraging Robotic View Transformer (RVT) to improve generalization for downstream robotics tasks.
0
6
36
RT @abhishekunique7: Over the last few months, weāve been thinking about how to learn from āoff-domainā data - data from non-robot sourcesā¦.
0
17
0
A very well deserved recognition to the pioneers of neural networks! The Nobel prize is timely and marks the acknowledgment of the transformative power of AI.
BREAKING NEWS.The Royal Swedish Academy of Sciences has decided to award the 2024 #NobelPrize in Physics to John J. Hopfield and Geoffrey E. Hinton āfor foundational discoveries and inventions that enable machine learning with artificial neural networks.ā
0
0
7
Thanks for the invitation! I am looking forward to presenting @MontrealRobots.
Join us tomorrow for this week's robot learning seminar, in which @imankitgoyal from @nvidia will present his talk titled "View Transformers for 3D Manipulation in Robotics". See you there at 11:30AM ET! . #Manipulation #multiViewLearning #transformers
0
1
38