
Karl Pertsch
@KarlPertsch
Followers
3K
Following
538
Media
98
Statuses
382
Robot Foundation Models @ UC Berkeley & Stanford & @physical_int | Postdoc w/ Sergey Levine & Chelsea Finn | Prev. Intern @ Google Brain, Meta AI | PhD @ USC.
Joined July 2015
RT @RobobertoMM: It was time to improve our evaluations in robot learning! We introduce a methodology based on anonymous A/B testing: faire….
0
3
0
RT @abhishekunique7: Check out some of our new work on distributed robot evaluation led by @KarlPertsch, @pranav_atreya and @tonyh_lee! Hop….
0
6
0
RT @tonyh_lee: 🚀 We just launched RoboArena — a real-world evaluation platform for robot policies!.Think Chatbot Arena, but for robotics.….
0
15
0
RT @pranav_atreya: In robotics benchmarks are rarely shared. New eval setups are created for each new project, a stark difference from eval….
0
21
0
Thanks to my co-leads @pranav_atreya @tonyh_lee!.And thanks to the many collaborators from across the robotics community who agreed to help out with evals! .@moo_jin_kim @prodarhan @dineshjayaraman @RobobertoMM @GlenBerseth @abhishekunique7 @YoungwoonLee @percyliang.
1
0
6
RT @kvablack: In LLM land, a slow model is annoying. In robotics, a slow model can be disastrous! Visible pauses at best, dangerously jerky….
0
55
0
RT @polkirichenko: Join us at #CVPR2025 Demographic Diversity in Computer Vision workshop tomorrow!.📅 Wednesday, June 11, 9am-6pm.📍 room 21….
0
20
0
Check out Danny's paper on a single-stage VLA recipe that trains fast, has fast inference, and follows language commands well. ⚡️⚡️⚡️.The key: combine FAST tokens + flow-matching expert, and make sure those pesky diffusion gradients don't mess up your beautiful VLM backbone! :).
How to build vision-language-action models that train fast, run fast & generalize? In our new paper, we formalize & analyze the approach of our π-0.5 model & further improve it with a single stage recipe. Blog: Paper:
0
2
34
Here's the link to our original Embodied CoT work: Also lots of other works have since shown that grounded reasoning can help generalization (eg Gemini robotics, HAMSTER. ) -- I think we still have only scratched the surface on these approaches!.
Excited to release our work on Embodied Chain-of-Thought Reasoning today!. We can boost performance of vision-language-action models like OpenVLA by a large margin without any additional robot training data!. The key: simply think before you act!. 1/.
0
0
3
Our embodied CoT work (ECoT) showed that policies generalize better when allowed to reason step-by-step, at the expense of slower inference. Will's new work investigates *why* ECoT policies work better & develops "ECoT-Lite" recipes that run much faster & still generalize well!👇.
Embodied chain-of-thought reasoning (ECoT) is a powerful way to improve robot generalization & performance. But why is this the case, and how can that inform the design of learned robot policies?.We investigate these questions in our latest work!.1/6
1
1
18