Yixuan Wang Profile
Yixuan Wang

@YXWangBot

Followers
1K
Following
1K
Media
44
Statuses
182

CS Ph.D. student @Columbia & Intern @AIatMeta | Prev. Boston Dynamics AI Institute, Google X #Vision #Robotics #Learning

New York, USA
Joined October 2019
Don't wanna be here? Send us removal request.
@YXWangBot
Yixuan Wang
7 months
🤔Active robot exploration is critical but hard – long-horizon, large space, and complex occlusions. How can robot explore like human?.🤖Introducing CuriousBot, which interactively explores and builds actionable 3D relational object graph. 🔗👇Threads(1/9)
2
9
39
@YXWangBot
Yixuan Wang
2 days
RT @HaochenShi74: ToddlerBot 2.0 is released🥳! Now Toddy can also do cartwheels🤸! We have added so many features since our first release in….
0
44
0
@grok
Grok
8 days
Join millions who have switched to Grok.
221
456
3K
@YXWangBot
Yixuan Wang
25 days
RT @suning_huang: 🚀 Excited to share our #CoRL2025 paper! See you in Korea 🇰🇷!🎉. We present ParticleFormer, a Transformer-based 3D world m….
0
19
0
@YXWangBot
Yixuan Wang
1 month
RT @YunzhuLiYZ: I was really impressed by the UMI gripper (@chichengcc et al.), but a key limitation is that **force-related data wasn’t ca….
0
24
0
@YXWangBot
Yixuan Wang
1 month
It is soooooo awesome to see UMI + Tactile comes to life! I am very impressed how quickly the whole hardware + software system is built. Meanwhile, they even collected lots of the data in the wild! Very amazing work!!!.
@binghao_huang
Binghao Huang
1 month
Tactile interaction in the wild can unlock fine-grained manipulation! 🌿🤖✋. We built a portable handheld tactile gripper that enables large-scale visuo-tactile data collection in real-world settings. By pretraining on this data, we bridge vision and touch—allowing robots to:
1
5
39
@YXWangBot
Yixuan Wang
2 months
RT @RussTedrake: TRI's latest Large Behavior Model (LBM) paper landed on arxiv last night! Check out our project website: .
0
107
0
@YXWangBot
Yixuan Wang
2 months
RT @shivanshpatel35: 🚀 Introducing RIGVid: Robots Imitating Generated Videos!.Robots can now perform complex tasks—pouring, wiping, mixing—….
0
33
0
@YXWangBot
Yixuan Wang
2 months
RT @YunzhuLiYZ: Had a great time yesterday giving three invited talks at #RSS2025 workshops—on foundation models, structured world models,….
0
12
0
@YXWangBot
Yixuan Wang
2 months
Just arrived at LA and excited to be at RSS! I will present CodeDiffuser at following sessions:.- Presentation on June 22 (Sun.) 5:30 PM - 6:30 PM.- Poster on June 22 (Sun.) 6:30 PM - 8:00 PM. I will also present CuriousBot at.- FM4RoboPlan Workshop on June 21 (Sat.) 9:40 - 10:10.
@YXWangBot
Yixuan Wang
2 months
🤖 Does VLA models really listen to language instructions? Maybe not 👀.🚀 Introducing our RSS paper: CodeDiffuser -- using VLM-generated code to bridge the gap between **high-level language** and **low-level visuomotor policy**.🎮 Try the live demo: (1/9)
0
1
13
@YXWangBot
Yixuan Wang
2 months
RT @YunzhuLiYZ: We’ve been exploring 3D world models with the goal of finding the right recipe that is both:.(1) structured—for sample effi….
0
13
0
@YXWangBot
Yixuan Wang
2 months
RT @robo_kat: How can we achieve both common sense understanding that can deal with varying levels of ambiguity in language and dextrous ma….
0
7
0
@YXWangBot
Yixuan Wang
2 months
RT @kaiwynd: Check out the cool results and demo!.
0
2
0
@YXWangBot
Yixuan Wang
2 months
Two releases in a row from our lab today 😆. One problem I was always pondering on is how to use structured representation while making it scalable. Super excited that Kaifeng's work pushes this direction forward and I cannot wait to see what's more in the future!!.
@kaiwynd
Kaifeng Zhang
2 months
Can we learn a 3D world model that predicts object dynamics directly from videos? . Introducing Particle-Grid Neural Dynamics: a learning-based simulator for deformable objects that trains from real-world videos. Website: ArXiv:
0
0
9
@YXWangBot
Yixuan Wang
2 months
RT @YunzhuLiYZ: **Steerability** remains one of the key issues for current vision-language-action models (VLAs). Natural language is often….
0
24
0
@YXWangBot
Yixuan Wang
2 months
RT @Haoyu_Xiong_: It is cool to see that you can steer your low-level policy with foundation models. Check out new work from @YXWangBot.
0
2
0
@YXWangBot
Yixuan Wang
2 months
RT @RoboPapers: Ep#10 with @RogerQiu_42 on Humanoid Policy ~ Human Policy . Co-hosted by @chris_j_paxton & @micoolc….
0
6
0
@YXWangBot
Yixuan Wang
2 months
This work is done with awesome Yitong and Guang! Thanks to the amazing collaborators Dale, Paarth, Kuni, @huan_zhang12, and Katherine for their supports and contributions! Also a huge thanks for my incredible advisor @YunzhuLiYZ for the support and guidance as always!.
0
1
2
@YXWangBot
Yixuan Wang
2 months
Links:.🌐 Website: 🖥️ Code: 📺 Video: We will present the paper on June 22 (Friday) from 5:30 PM to 6:30 PM, and poster from 6:30 PM to 8:00 PM. See you in LA! (9/9)
1
1
6
@YXWangBot
Yixuan Wang
2 months
Our tasks involve contact-rich manipulation and multi-object interactions, enabled by the visuomotor policy learned from demonstrations. To stow the book, the robot first squeezes in the book, pushes other books aside to find space, and finally inserts the book. (8/9)
1
0
1
@YXWangBot
Yixuan Wang
2 months
Geometric relations are also important for manipulation tasks. Our system can also capture geometric relations, as evidenced by the following battery packing task. The user can specify the target battery by its relative location, such as “frontmost” or “right column”. (7/9)
1
0
1
@YXWangBot
Yixuan Wang
2 months
Moreover, our framework can understand more fine-grained semantic information, such as names on the book cover, and select the right book instance to stow books. (6/9)
1
0
1