
Yixuan Wang
@YXWangBot
Followers
1K
Following
1K
Media
44
Statuses
182
CS Ph.D. student @Columbia & Intern @AIatMeta | Prev. Boston Dynamics AI Institute, Google X #Vision #Robotics #Learning
New York, USA
Joined October 2019
🤔Active robot exploration is critical but hard – long-horizon, large space, and complex occlusions. How can robot explore like human?.🤖Introducing CuriousBot, which interactively explores and builds actionable 3D relational object graph. 🔗👇Threads(1/9)
2
9
39
RT @HaochenShi74: ToddlerBot 2.0 is released🥳! Now Toddy can also do cartwheels🤸! We have added so many features since our first release in….
0
44
0
RT @suning_huang: 🚀 Excited to share our #CoRL2025 paper! See you in Korea 🇰🇷!🎉. We present ParticleFormer, a Transformer-based 3D world m….
0
19
0
RT @YunzhuLiYZ: I was really impressed by the UMI gripper (@chichengcc et al.), but a key limitation is that **force-related data wasn’t ca….
0
24
0
It is soooooo awesome to see UMI + Tactile comes to life! I am very impressed how quickly the whole hardware + software system is built. Meanwhile, they even collected lots of the data in the wild! Very amazing work!!!.
Tactile interaction in the wild can unlock fine-grained manipulation! 🌿🤖✋. We built a portable handheld tactile gripper that enables large-scale visuo-tactile data collection in real-world settings. By pretraining on this data, we bridge vision and touch—allowing robots to:
1
5
39
RT @RussTedrake: TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: .
0
107
0
RT @shivanshpatel35: 🚀 Introducing RIGVid: Robots Imitating Generated Videos!.Robots can now perform complex tasks—pouring, wiping, mixing—….
0
33
0
RT @YunzhuLiYZ: Had a great time yesterday giving three invited talks at #RSS2025 workshops—on foundation models, structured world models,….
0
12
0
Just arrived at LA and excited to be at RSS! I will present CodeDiffuser at following sessions:.- Presentation on June 22 (Sun.) 5:30 PM - 6:30 PM.- Poster on June 22 (Sun.) 6:30 PM - 8:00 PM. I will also present CuriousBot at.- FM4RoboPlan Workshop on June 21 (Sat.) 9:40 - 10:10.
🤖 Does VLA models really listen to language instructions? Maybe not 👀.🚀 Introducing our RSS paper: CodeDiffuser -- using VLM-generated code to bridge the gap between **high-level language** and **low-level visuomotor policy**.🎮 Try the live demo: (1/9)
0
1
13
RT @YunzhuLiYZ: We’ve been exploring 3D world models with the goal of finding the right recipe that is both:.(1) structured—for sample effi….
0
13
0
RT @robo_kat: How can we achieve both common sense understanding that can deal with varying levels of ambiguity in language and dextrous ma….
0
7
0
Two releases in a row from our lab today 😆. One problem I was always pondering on is how to use structured representation while making it scalable. Super excited that Kaifeng's work pushes this direction forward and I cannot wait to see what's more in the future!!.
Can we learn a 3D world model that predicts object dynamics directly from videos? . Introducing Particle-Grid Neural Dynamics: a learning-based simulator for deformable objects that trains from real-world videos. Website: ArXiv:
0
0
9
RT @YunzhuLiYZ: **Steerability** remains one of the key issues for current vision-language-action models (VLAs). Natural language is often….
0
24
0
RT @Haoyu_Xiong_: It is cool to see that you can steer your low-level policy with foundation models. Check out new work from @YXWangBot.
0
2
0
RT @RoboPapers: Ep#10 with @RogerQiu_42 on Humanoid Policy ~ Human Policy . Co-hosted by @chris_j_paxton & @micoolc….
0
6
0
This work is done with awesome Yitong and Guang! Thanks to the amazing collaborators Dale, Paarth, Kuni, @huan_zhang12, and Katherine for their supports and contributions! Also a huge thanks for my incredible advisor @YunzhuLiYZ for the support and guidance as always!.
0
1
2