Haozhi Qi
@HaozhiQ
Followers
4K
Following
2K
Media
70
Statuses
381
Amazon FAR | Incoming Assistant Professor @UChicagoCS Prev: Ph.D. @berkeley_ai
San Francisco, CA
Joined February 2019
I will join UChicago CS @UChicagoCS as an Assistant Professor in late 2026, and I’m recruiting PhD students in this cycle (2025 - 2026). My research focuses on AI & Robotics - including dexterous manipulation, humanoids, tactile sensing, learning from human videos, robot
26
101
643
Huge thanks to my brilliant collaborators: Jessica Yin, @WiYoungsun, Sayantan Kundu, Mike Lambeta, William Yang, @ChanghaoWang78, @TingfanW1208, @JitendraMalikCV, and @t_hellebrekers! Check out the website ( https://t.co/gh76bWyTha) for the full details, code, and design files!
jessicayin.github.io
OSMO is an open-source wearable tactile glove with 12 three-axis sensors for human-to-robot skill transfer. Train robot policies from human demonstrations without robot data.
1
0
12
In our real-world wiping task, our tactile-aware policy achieved a 72% success rate, significantly outperforming vision-only baselines. The vision-only policies frequently failed due to contact-related issues (e.g., inconsistent pressure or lost grip on the sponge) that tactile
1
0
6
The most exciting part: We trained a robot policy for a wiping task exclusively on human demonstrations collected with OSMO - without any real robot data! By using the same OSMO glove on both the human and the Psyonic Ability Hand, we minimize the visual and tactile "embodiment
1
0
5
We designed OSMO to be "in-the-wild" compatible: - It integrates with state-of-the-art hand tracking methods, including wearable AR/VR headsets (Quest 3, Aria Gen 2, Apple Vision Pro) and generic RGB videos. - The glove's appearance (low-profile electronics, beige fabric) is
1
0
8
OSMO's design is all about bridging the gap between human and robot dexterity! It has 12 three-axis magnetic tactile sensors across the fingertips and palm, measuring both shear and normal forces (what vision can't see!). It preserves full human dexterity, allowing for natural,
1
1
8
Human videos are great for robot learning, but they critically lack the sense of touch needed for complex manipulation tasks! Introducing OSMO, an open-source tactile glove that captures rich contact signals during human demonstrations for direct, successful transfer to robots.
5
31
165
Thank you, @YiMaTweets! It’s been an honor to be your student! I wouldn’t be here without your guidance.
Congratulations, Haozhi. Very proud of you! Wish this to be an exciting and rewarding journey for you.
0
0
22
Simulation has been a central theme of my PhD, but one thing has always troubled me (and many colleagues): complex physics and rich sensing are incredibly hard to simulate. In this work, we explore a different path: learn a simplified task in simulation, then use that
dexscrew.github.io
We propose a framework for learning dexterous manipulation skills from imperfect simulations, demonstrating effectiveness on nut-bolt fastening and screwdriving tasks.
How do we get a robot to use a screwdriver 🪛 and fasten a nut 🔩 ? Introducing DexScrew, our sim-to-real framework that enables a dexterous hand 🖐️ to autonomously perform complex and contact-rich tasks even when we cannot accurately simulate them. We open-source the
0
7
62
I'm recruiting multiple PhD students this cycle to join me at Harvard University and the Kempner Institute! My interests span vision and intelligence, including 3D/4D, active perception, memory, representation learning, and anything you're excited to explore! Deadline: Dec 15th.
24
152
912
Generative models (diffusion/flow) are taking over robotics 🤖. But do we really need to model the full action distribution to control a robot? We suspected the success of Generative Control Policies (GCPs) might be "Much Ado About Noising." We rigorously tested the myths. 🧵👇
15
90
510
Can we quickly improve a pre-trained robot policy by learning from real world human corrections? Introducing Compliant Residual DAgger (CR-DAgger), a system that improves policies performance to close to 100% on challenging contact-rich manipulation problems, using as few as
4
51
214
Sim-to-real learning for humanoid robots is a full-stack problem. Today, Amazon FAR is releasing a full-stack solution: Holosoma. To accelerate research, we are open-sourcing a complete codebase covering multiple simulation backends, training, retargeting, and real-world
19
131
573
Thank you @ken0324! I'm also super excited to join!
I’m super excited for Haozhi to join @UChicagoCS and lead cutting-edge research on robotic manipulation. If you’re interested in this area, consider applying to work with his lab!
0
0
9
Thank you @plopesresearch! Looking forward to working together!
Robotics/ML/HCI PHDs: Come work with us at @UChicagoCS in @HaozhiQ 's group, solving great challenges in dexterous robotic hands, tactile sensing and more!
0
1
4
Thank you @irmakkguzey! It was a truly great and productive summer! Learned a lot from you.
Haozhi is an absolutely great person to work with and I'm very excited for him. It was a joy collaborating with him and chatting throughout the summer. If you’re applying to PhD programs, make sure to check him out!
0
0
10
Thank you @mangahomanga! It’s an honor to work with you!
Haozhi is amazing and has great research taste! Big win for @UChicagoCS
0
0
5