RobobertoMM Profile Banner
Roberto Profile
Roberto

@RobobertoMM

Followers
2K
Following
495
Media
19
Statuses
172

Assistant CS Professor at UT Austin. Former Stanford and TUBerlin. Researching at the intersection of vision, learning and robotics 🏳️‍🌈

Joined August 2019
Don't wanna be here? Send us removal request.
@ChengshuEricLi
Chengshu Li
6 days
We are excited to release MoMaGen, a data generation method for multi-step bimanual mobile manipulation. MoMaGen turns 1 human-teleoped robot trajectory into 1000s of generated trajectories automatically.🚀 Website: https://t.co/DYKvqY4bII arXiv: https://t.co/lDffi0FXHl
1
35
157
@ShivinDass
Shivin Dass
20 days
A little late to this but excited to share that DataMIL won the best paper at the Data workshop at #CoRL! If you haven't already, check it out! 👇
@ShivinDass
Shivin Dass
6 months
Ever wondered which data from large datasets (like OXE) actually helps when training/tuning a policy for specific tasks? We present DataMIL, a framework for measuring how each training sample influences policy performance, hence enabling effective data selection 🧵
1
4
13
@NaveenManwani17
naveen manwani
30 days
🚨CoRL 2025 Best Poster Award 🏆 Paper Alert 🚨 ➡️Paper Title: Mash, Spread, Slice! Learning to Manipulate Object States via Visual Spatial Progress 🌟Few pointers from the paper 🎯Most robot manipulation focuses on changing the kinematic state of objects: picking, placing,
0
3
7
@JiahengHu1
Jiaheng Hu
1 month
Excited that SPARTA ( https://t.co/AVZTpsbfSw) won the best poster award at the CoRL RINO workshop! Big congrats to the project lead Priyanka, who worked so hard on this project, as well as to the rest of the co-authors @ShivinDass @sagnikmjr @RobobertoMM and Kristen!
1
2
19
@thegautamkamath
Gautam Kamath
1 month
📢 Call for Community Activities #AAAI2026 We invite submissions of proposals for including and open activities that help broaden community participation in the AI field. October 4: Submission Deadline October 18: Acceptance Notifications @RobobertoMM @marucabrera27 @RealAAAI
1
5
10
@RobobertoMM
Roberto
1 month
Simple but *so effective idea*! And it can be used with any feature data selector. Great work led by @sateeshk21 . Do not miss it at #CoRL2025 (Spotlight 4 & Poster 2 on Sept 29)!
@sateeshk21
Sateesh Kumar
1 month
Which data is best for training few-shot imitation policies for robot manipulation? Some think it’s the data that looks similar, or has similar motion, or comes with related language labels. They are all right AND wrong: depending on the task, sometimes this similarity helps but
0
1
7
@rutavms
Rutav
1 month
Intelligent humanoids should have the ability to quickly adapt to new tasks by observing humans Why is such adaptability important? 🌍 Real-world diversity is hard to fully capture in advance 🧠 Adaptability is central to natural intelligence We present MimicDroid 👇 🌐
7
40
121
@RobobertoMM
Roberto
2 months
Honored to give an Early Career Invited Talk at #IJCAI today. See you at 11:30am in room 520C!
1
3
21
@RobobertoMM
Roberto
4 months
It was time to improve our evaluations in robot learning! We introduce a methodology based on anonymous A/B testing: fairer, stronger, community-driven. Awesome work by @KarlPertsch @pranav_atreya @tonyh_lee and an incredible crowdsourcing team. Upload and test your model! 🚀
@KarlPertsch
Karl Pertsch
4 months
We’re releasing the RoboArena today!🤖🦾 Fair & scalable evaluation is a major bottleneck for research on generalist policies. We’re hoping that RoboArena can help! We provide data, model code & sim evals for debugging! Submit your policies today and join the leaderboard! :) 🧵
0
3
21
@ArpitBahety
Arpit Bahety
4 months
Imagine a future where robots are part of our daily lives — How can end users teach robots new tasks by directly showing them, just like teaching another person? 🧵👇
3
17
44
@huihan_liu
Huihan Liu
4 months
Meet Casper👻, a friendly robot sidekick who shadows your day, decodes your intents on the fly, and lends a hand while you stay in control! Instead of passively receiving commands, what if a robot actively sense what you need in the background, and step in when confident? (1/n)
6
40
159
@RobobertoMM
Roberto
5 months
🚨RL training for contact-rich tasks with a mobile manipulator IN THE REAL WORLD?!🤯 We're not crazy—just equipped with the right action space! SLAC learns a safe, effective action space via unsupervised RL in sim, enabling real-world RL training in minutes. Check it out!🚀
@JiahengHu1
Jiaheng Hu
5 months
Real-world RL, where robots learn directly from physical interactions, is extremely challenging — especially for high-DoF systems like mobile manipulators. 1⃣ Long-horizon tasks and large action spaces lead to difficult policy optimization. 2⃣ Real-world exploration with
0
2
23
@JiahengHu1
Jiaheng Hu
5 months
Excited to be in ATL for #ICRA2025 to present 🔥FLaRe: fine-tuning large transformer policies with #RL, 15:25 Tuesday @ room 410! I will also be attending the 📷Doctoral Consortium on Monday to talk about my research on self-improving robots. Happy to meet old and new friends!
@JiahengHu1
Jiaheng Hu
1 year
🚀 Despite efforts to scale up Behavior Cloning for Robots, large-scale BC has yet to live up to its promise. How can we break through the performance plateau? Introducing 🔥FLaRe: fine-tuning large-scale robot policies with Reinforcement Learning. https://t.co/iRC1NTgoFI 🧵
2
8
35
@RobobertoMM
Roberto
5 months
Loved working on this with our MIT/Stanford/OpenAI collaborators. It brings "The Bitter Lesson" to data curation: skip the hand-tuned heuristics (visual similarity, motion...) and let the data speak for itself! Datamodels is a fascinating framework 🤯
@ShivinDass
Shivin Dass
6 months
Ever wondered which data from large datasets (like OXE) actually helps when training/tuning a policy for specific tasks? We present DataMIL, a framework for measuring how each training sample influences policy performance, hence enabling effective data selection 🧵
0
4
39
@EliasEskin
Elias Stengel-Eskin
6 months
Extremely excited to announce that I will be joining @UTAustin @UTCompSci in August 2025 as an Assistant Professor! 🎉 I’m looking forward to continuing to develop AI agents that interact/communicate with people, each other, and the multimodal world. I’ll be recruiting PhD
92
65
458
@RobobertoMM
Roberto
6 months
So happy for @JiahengHu1 ! He has been rocking it, with outstanding work that pushes the limits of what robot learning can achieve in mobile manipulation and other domains. And one of my first Ph.D. students! Congratulations! 🦾🦾🦾🦾
@JiahengHu1
Jiaheng Hu
6 months
I'm honored to be awarded the 2025 Two Sigma PhD fellowship, and extremely grateful to my two amazing advisors @RobobertoMM @PeterStone_TX ! Looking forward to continuing to advance the field of RL and Robotics.
0
0
5
@xf1280
Fei Xia
8 months
✨Super excited to share what the team has been working on! ♊️🤖 Gemini Robotics is a family of frontier models that are dexterous, interactive, and general. It builds on top of Gemini's world understanding, enhancing it's spatial/embodied reasoning, and producing robot
@GoogleDeepMind
Google DeepMind
8 months
Meet Gemini Robotics: our latest AI models designed for a new generation of helpful robots. 🤖 Based on Gemini 2.0, they bring capabilities such as better reasoning, interactivity, dexterity and generalization into the physical world. 🧵 https://t.co/EXRJrmxGxl
8
14
126
@RobobertoMM
Roberto
8 months
Giving a talk as New Faculty Highlight at AAAI tomorrow morning (9:30am)! https://t.co/gzhs00f4Qt Come if you want to get an overview of some of the works from the lab
aaai.org
The Thirty-Ninth AAAI Conference on Artificial Intelligence will be held in Philadelphia at the Pennsylvania Convention Center in 2025.
0
4
18
@RobobertoMM
Roberto
9 months
Tired of guessing what tasks people want robots to do for them? Check our study! We correlate time spent and emotions people felt while performing tasks with the desire to automate them, comparing between different groups. And with an online tool for you to play with the data!
@ruchira_ray
Ruchira (she/they)
10 months
🤔What tasks do we want robots to handle? Are these preferences based on saved time or feelings we associate with the tasks? Introducing Why Automate This?—a study exploring automation preferences across social groups, using feelings & time-spent as key factors. 👇 (1/5)
0
0
9
@duke_zzwang
Zizhao Wang
1 year
In multi-object env, why do most Unsupervised Skill Discovery methods fail to learn complex skills like tool use? Because they simply maximize state coverage. Introducing our solution SkiLD: Skill Discovery Guided by Factor Interactions (NeurIPS24) https://t.co/buo3qSdI1O
1
12
64