Mustafa Mukadam
@mukadammh
Followers
2K
Following
973
Media
36
Statuses
275
Robotics and AI researcher @amazon | Prev: Research lead, robot manipulation @AIatMeta, PhD @GTrobotics
Joined December 2021
Sparsh-X is the culmination of our multi year work on building a touch foundation model SSL pre-trained for multimodal tactile (image, audio, motion, pressure) Work led by @carohiguerarias, @akashshrm02 and will be presented as oral @corl_conf
How can we leverage the full spectrum of touch for more robust robot manipulation? 🦾Meet Sparsh-X, a backbone for multisensory touch representation across image, audio, motion, and pressure. 🧵1/6 👇
1
1
23
100s of robots manipulating 1M+ unique objects in dense clutter in real fullfilment centers that ship to customers we are working on multimodal representations, behavior cloning, reinforcement learning, all with the focus on what actually scales in real deployment conditions
0
0
1
Bored of working on toy tasks in the lab? We are solving robot manipulation at massive real world scales with Vulcan at @amazon I am at #CoRL2025 and my team is looking for PhD research interns, postodcs, and scientists https://t.co/cXGrLDIGEi
aboutamazon.com
Built on advances in robotics, engineering, and physical AI, Vulcan is making our workers’ jobs easier and safer while moving orders more efficiently.
1
2
16
Flying out for CoRL from Pittsburgh today! It's already exciting to hear robotics conversations around. Looking forward to see both new and familiar faces. I will be presenting Sparsh-X (Oral presentation session 3) and Sparsh-skin (Poster session 3)!
Robots need touch for human-like hands to reach the goal of general manipulation. However, approaches today don’t use tactile sensing or use specific architectures per tactile task. Can 1 model improve many tactile tasks? 🌟Introducing Sparsh-skin: https://t.co/DgTq9OPMap 1/6
0
9
46
Introducing CLAMP: : a device, dataset, and model that bring large-scale, in-the-wild multimodal haptics to real robots. Haptics / Tactile data is more than just force or surface texture, and capturing this multimodal haptic information can be useful for robot manipulation.
1
19
41
We have completed the SLAM Handbook "From Localization and Mapping to Spatial Intelligence" and released it online: https://t.co/AnKa398nyw . The handbook will be published by Cambridge University Press. [1/n]
4
76
292
We have completed editing the SLAM Handbook "From Localization and Mapping to Spatial Intelligence" and released it. The 3-part handbook will be published by Cambridge University Press. Enjoy reading online for now!
6
23
210
We're hiring interns (and full-times) all year long! Please email me if interested.
41
86
2K
I am particularly excited about how Sparsh-X enables easy sim-to-real transfer of proprioception-only dexterous manipulation policies through tactile adaptation and a few real rollouts https://t.co/YvlzXENJu4
We put tactile adaptation with Sparsh-X to the test on the in-hand rotation policy! Results show enhanced stability, a 90% reduction in vertical slip, and remarkable robustness against wrist pose perturbations & changes in object properties. 🏆 🧵5/6
0
1
5
Exciting work on multi-purpose tactile representations, enabling precise insertion tasks and more! Led by friends in our FAIR Robotics team @AIatMeta
How can we leverage the full spectrum of touch for more robust robot manipulation? 🦾Meet Sparsh-X, a backbone for multisensory touch representation across image, audio, motion, and pressure. 🧵1/6 👇
0
1
18
Sparsh-X is our new 'multi-sensory' rep model showing the power of encoding multiple tactile modalities for dexterity. We saw performance for plug insertion go up to 90% success with all modalities, and higher robustness with sim-to-real policies adapted to use Sparsh-X!
How can we leverage the full spectrum of touch for more robust robot manipulation? 🦾Meet Sparsh-X, a backbone for multisensory touch representation across image, audio, motion, and pressure. 🧵1/6 👇
0
4
18
Excited to release AlgoTune!! It's a benchmark and coding agent for optimizing the runtime of numerical code 🚀 https://t.co/bdR630y0dL 📚 https://t.co/vSnV3eUgVs 🤖 https://t.co/krJ7XDrJFA with @OfirPress @ori_press @PatrickKidger @b_stellato @ArmanZharmagam1 & many others 🧵
3
41
194
We demo’d Amazon grasp model at RSS this year. We performed over 600 grasps over one day at roughly 80-90% SR 1. On an open item set (people gave random often adversarial items), 2. In a random scene fully outdoor throughout the day 3. On a new embodiment (different from
2
7
68
So you’ve trained your favorite diffusion/flow based policy, but it’s just not good enough 0-shot. Worry not, in our new work DSRL - we show how to *steer* pre-trained diffusion policies with off-policy RL, improving behavior efficiently enough for direct training in the real
6
25
197
📢Life is a sequence of bets – and I’ve picked my next: @MyolabAI It’s incredibly ambitious, comes with high risk, & carries unbounded potential. But it’s a version of the #future I deeply believe in. I believe: ➡️AI will align strongly with humanity - coz it maximizes its own
myolab.ai
Building Human-Embodied Intelligence to Empower Humans.
All forms of intelligence co-emerged with a body, except AI We're building a #future where AI evolves as your lifelike digital twin to assist your needs across health, sports, daily life, creativity, & beyond... https://t.co/QL3o9YxZYz ➡️ Preview your first #HumanEmbodiedAI
17
21
108
We introduce Dexterity Gen (DexGen), a foundation controller that enables unprecedented dexterous manipulation capabilities. For the first time, it allows human teleoperation of tasks such as using a pen, screwdriver, and syringe. Developed by @berkeley_AI and @MetaAI. A Thread.
0
0
0
This was a key feature in enabling DexterityGen, our teleop that can support tasks like using a screw driver Led by @zhaohengyin, now open source
Just open-sourced Geometric Retargeting (GeoRT) — the kinematic retargeting module behind DexterityGen. Includes tools for importing custom hands. Give it a try: https://t.co/MxSuitRaDM A software by @berkeley_ai and @AIatMeta. More coming soon.
1
1
7
Everyday human data is robotics’ answer to internet-scale tokens. But how can robots learn to feel—just from videos?📹 Introducing FeelTheForce (FTF): force-sensitive manipulation policies learned from natural human interactions🖐️🤖 👉 https://t.co/CZcG87xYn5 1/n
11
39
221
Very interesting work! And great to see self-supervised learning being used for tactile data. This is critical to scaling tactile to the level that vision has scaled.
Robots need touch for human-like hands to reach the goal of general manipulation. However, approaches today don’t use tactile sensing or use specific architectures per tactile task. Can 1 model improve many tactile tasks? 🌟Introducing Sparsh-skin: https://t.co/DgTq9OPMap 1/6
2
3
29