Physical Intelligence Profile
Physical Intelligence

@physical_int

Followers
24K
Following
43
Media
38
Statuses
61

Physical Intelligence (Pi), bringing AI into the physical world.

San Francisco, CA
Joined March 2024
Don't wanna be here? Send us removal request.
@physical_int
Physical Intelligence
3 months
RT @kvablack: In LLM land, a slow model is annoying. In robotics, a slow model can be disastrous! Visible pauses at best, dangerously jerky….
0
62
0
@physical_int
Physical Intelligence
3 months
To learn more about RTC, check our blog post, as well as the full-length research paper that we prepared with detailed results:
pi.website
Physical Intelligence is bringing general-purpose AI into the physical world.
1
4
44
@grok
Grok
22 days
Introducing Grok Imagine.
2K
4K
28K
@physical_int
Physical Intelligence
3 months
Quantitatively, RTC enables high success rates even as we introduce artificially high inference delays. Even if we add 200ms of additional delays on top of what the model needs, RTC performance remains stable, while naive methods drop off sharply.
Tweet media one
1
1
40
@physical_int
Physical Intelligence
3 months
RTC not only makes motions smoother, but it actually allows the robot to move with more precision and speed, improving the performance of our models without any training-time changes at all!
1
0
25
@physical_int
Physical Intelligence
3 months
The main idea in RTC is to treat generating a new action chunk as an “inpainting” problem: the actions that will be executed while the robot “thinks” are treated as fixed, while new actions are inferred via flow matching.
1
2
35
@physical_int
Physical Intelligence
3 months
Our models need to run in real time on real robots, but inference with big VLAs takes a long time. We developed Real-Time Action Chunking (RTC) to enable real-time inference with flow matching for the π0 and π0.5 VLAs! More in the thread👇
7
78
675
@physical_int
Physical Intelligence
3 months
RT @GoddenThomas: Quick life update; I've joined the hardware team at @physical_int and I'm moving to San Francisco!.
0
6
0
@physical_int
Physical Intelligence
3 months
To learn more, check out our blog post here:
pi.website
Physical Intelligence is bringing general-purpose AI into the physical world.
0
4
47
@physical_int
Physical Intelligence
3 months
At runtime, π-0.5 with knowledge insulation runs much faster than π-0-FAST, giving us the best of all worlds: fast training, fast robot control, and better generalization.
1
3
55
@physical_int
Physical Intelligence
3 months
This works really well: π-0.5 with knowledge insulation trains just as fast as π-0-FAST, 5x-7x faster than π-0, and still retains excellent performance, dexterity, and compute-efficient inference at runtime, outperforming our prior models.
Tweet media one
Tweet media two
1
1
32
@physical_int
Physical Intelligence
3 months
Our knowledge insulation method uses discrete tokens to train the VLM backbone, so that it learns appropriate representations, even as the action expert trains to take those representations and turn them into continuous actions.
Tweet media one
1
2
27
@physical_int
Physical Intelligence
3 months
The trouble is that in practice, the action expert is cold started, and in the beginning of VLA training the gradients from the action expert corrupt the VLM backbone, causing it to forget web-scale pretraining and slow down training.
Tweet media one
1
1
34
@physical_int
Physical Intelligence
3 months
The VLM consists of a language model (LLM) and a vision encoder. The vision encoder serves as a kind of “visual cortex”, and the LLM as a kind of “prefrontal cortex”. To turn it into a VLA, we add a “motor cortex”: an action expert that can output continuous action chunks.
Tweet media one
1
2
31
@physical_int
Physical Intelligence
3 months
We figured out how to train VLAs with diffusion outputs much faster (7.5x faster), inheriting better language following from the VLM, and leading to better results. The key: protect the VLM backbone during training with knowledge insulation. Let’s talk about what we learned👇
6
103
816
@physical_int
Physical Intelligence
4 months
To learn more about π-0.5 (pronounced “pi oh five”), check out our blog post here: A research paper about π-0.5 describing the model, training, and experiments is here:
Tweet card summary image
pi.website
Our latest generalist policy, π0.5, extends π0 and enables open-world generalization. Our new model can control a mobile manipulator to clean up an entirely new kitchen or bedroom.
2
10
103
@physical_int
Physical Intelligence
4 months
97.6% of the π-0.5 data does not come from the mobile robots we use in our experiments. Most comes from other robots: cross-embodiment data in the lab, non-mobile (static) robots in the wild. It also includes web data. Each piece of data is important for good results.
Tweet media one
1
5
82
@physical_int
Physical Intelligence
4 months
π-0.5 performs hierarchical inference, inferring high-level semantic subtasks ("pick up the plate") followed by actions. It uses a co-training recipe with data from other robots, high-level commands, verbal instructions, and multimodal data from the web.
Tweet media one
3
4
74
@physical_int
Physical Intelligence
4 months
We got a robot to clean up homes that were never seen in its training data! Our new model, π-0.5, aims to tackle open-world generalization. We took our robot into homes that were not in the training data and asked it to clean kitchens and bedrooms. More below⤵️
54
261
2K
@physical_int
Physical Intelligence
5 months
At Pi, we seek to develop generalizable and broadly capable physical intelligence. With models that can control more robots and perform more skills, generalist cross-embodiment policies will continue to become more and more useful, practical and powerful.
0
1
23
@physical_int
Physical Intelligence
5 months
Receive menu from customer
1
0
29