suvir_m Profile Banner
Suvir Mirchandani Profile
Suvir Mirchandani

@suvir_m

Followers
537
Following
411
Media
13
Statuses
43

PhD Student @StanfordAILab | Prev: Student Researcher @GoogleAI, AI Resident @MetaAI, BS/MS @Stanford CS

Joined August 2014
Don't wanna be here? Send us removal request.
@suvir_m
Suvir Mirchandani
2 months
RT @Haoyu_Xiong_: Your bimanual manipulators might need a Robot Neck 🤖🦒. Introducing Vision in Action: Learning Active Perception from Huma….
0
87
0
@suvir_m
Suvir Mirchandani
2 months
So excited for @siddk to begin this new chapter! Working with Sidd has been a blast -- one of the most brilliant and kindest people I have had the privilege to learn from. Future PhD students, apply to work with Sidd!!.
@siddkaramcheti
Siddharth Karamcheti
2 months
Thrilled to share that I'll be starting as an Assistant Professor at Georgia Tech (@ICatGT / @GTrobotics / @mlatgt) in Fall 2026. My lab will tackle problems in robot learning, multimodal ML, and interaction. I'm recruiting PhD students this next cycle – please apply/reach out!
Tweet media one
Tweet media two
1
0
9
@suvir_m
Suvir Mirchandani
2 months
RT @jenngrannen: Meet ProVox: a proactive robot teammate that gets you 🤖❤️‍🔥. ProVox models your goals and expectations before a task start….
0
14
0
@suvir_m
Suvir Mirchandani
2 months
RT @oshaikh13: What if LLMs could learn your habits and preferences well enough (across any context!) to anticipate your needs?. In a new p….
0
92
0
@suvir_m
Suvir Mirchandani
3 months
RT @verityw_: Embodied chain-of-thought reasoning (ECoT) is a powerful way to improve robot generalization & performance. But why is this t….
0
11
0
@suvir_m
Suvir Mirchandani
3 months
RT @perryadong: Robotic models are advancing rapidly—but how do we scale their improvement? 🤖. We propose a recipe for batch online RL (tra….
0
15
0
@suvir_m
Suvir Mirchandani
6 months
Data quality can have a big impact on the performance of behavior cloning methods. But how can we measure the quality of demonstrations? One way is to score demonstrations via mutual information estimators:. Check out @JoeyHejna's thread 👇.
@JoeyHejna
Joey Hejna
6 months
Behavior cloning. clones behaviors, so naturally data quality directly affects performance. However, there aren't great ways of measuring how "good" or "bad" different demonstrations are. Our recent work seeks to address this problem using estimators of mutual information. đź§µ
0
1
11
@suvir_m
Suvir Mirchandani
6 months
Human video data can be easier to collect than robot demonstrations—but extracting actions for training robot policies is challenging. RAD uses language reasoning extracted from human videos to boost generalization in reasoning-based policies. Check out @jadenvclark's thread 👇.
@jadenvclark
Jaden Clark
6 months
How can we leverage human video data to train generalist robot policies? 🤖. Enter RAD: Reasoning through Action-Free Data, a new way to train robot policies using both robot and human video data via action reasoning.
0
0
5
@suvir_m
Suvir Mirchandani
9 months
This was a fun project to be a part of! Led by @jenngrannen and @siddkaramcheti (on the faculty job market!); w/ @percyliang and @DorsaSadigh. Paper: Website:
0
0
0
@suvir_m
Suvir Mirchandani
9 months
Introducing Vocal Sandbox 💬🤖: a framework for enabling human-robot collaboration by building up & adapting robot behaviors at different levels of abstraction via multiple teaching modalities. Check out @jenngrannen's thread below 👇.
@jenngrannen
Jenn Grannen
9 months
Introducing 🆚Vocal Sandbox: a framework for building adaptable robot collaborators that learns new 🧠high-level behaviors and 🦾low-level skills from user feedback in real-time. ✅. Appearing today at @corl_conf as an Oral Presentation (Session 3, 11/6 5pm). 🧵(1/6)
1
0
5
@suvir_m
Suvir Mirchandani
9 months
But learning from mixed-quality crowdsourced data also presents challenges and opportunities for future work! We’ll be presenting more about RoboCrowd at the Workshop on Mastering Robot Manipulation in a World of Abundant Data @corl_conf on Saturday 11/9. (4/5)
Tweet media one
1
2
9
@suvir_m
Suvir Mirchandani
9 months
We validate various incentive mechanisms for shaping user interactions into useful data—and further, see evidence that the crowdsourced data can be useful for pre-training policies that are then fine-tuned on expert data. (3/5)
1
0
4
@suvir_m
Suvir Mirchandani
9 months
We instantiate the framework in a system based on ALOHA (@tonyzzhao et al. 2023). We deployed the system for two weeks near a cafe on Stanford campus. Over 200 users voluntarily came up to the robot and interacted with it, giving us over 800 episodes (2/5)
1
0
3
@suvir_m
Suvir Mirchandani
9 months
We propose RoboCrowd, a framework for collecting robot demonstrations via crowdsourcing, gamification, and incentive design. Our key idea is to align user incentives like material rewards 🍬, intrinsic interest 🤔, and social comparison 🎮 with getting useful robot data 🧵(1/5)
4
18
108
@suvir_m
Suvir Mirchandani
9 months
We’re presenting our “negative results” at @corl_conf -- stop by our poster in Session 4 (4:00-5:30pm) this Friday 11/8. (6/7)
Tweet media one
1
0
2
@suvir_m
Suvir Mirchandani
9 months
And even on simpler tasks in real/simulation with minimal environment challenges, we found across a variety of methods that autonomous IL could be mildly helpful to performance, but reallocating our effort to collecting more data for IL was often simpler and more efficient. (5/7).
1
0
2
@suvir_m
Suvir Mirchandani
9 months
In our experience: not really. We document a variety of challenges in setting up *truly autonomous* data collection. Addressing these challenges required effort on either instrumenting environments or shaping tasks to become less useful/realistic. (4/7)
Tweet media one
1
0
2