Minghuan Liu Profile
Minghuan Liu

@ericliuof97

Followers
661
Following
483
Media
17
Statuses
169

Postdoc at @utaustin RPL Lab. Prev: Bytedance Seed Robotics; @ucsd; @sjtu1896. Robot Learning, Embodied AI.

Joined September 2016
Don't wanna be here? Send us removal request.
@ericliuof97
Minghuan Liu
2 months
๐Ÿš€ Want to build a 3D-aware manipulation policy, but troubled by the noisy depth perception? Want to train your manipulation policy in simulation, but tired of bridging the sim2real gap by degenerating geometric perception, like adding noise? Now these notorious problems are gone
6
43
214
@ericliuof97
Minghuan Liu
5 days
Mask Robot
@UnitreeRobotics
Unitree
5 days
Unitree Introducing | Unitree H2 Destiny Awakening!๐Ÿฅณ Welcome to this world โ€” standing 180cm tall and weighing 70kg. The H2 bionic humanoid - born to serve everyone safely and friendly.
0
0
0
@dwarkesh_sp
Dwarkesh Patel
8 days
The @karpathy interview 0:00:00 โ€“ AGI is still a decade away 0:30:33 โ€“ LLM cognitive deficits 0:40:53 โ€“ RL is terrible 0:50:26 โ€“ How do humans learn? 1:07:13 โ€“ AGI will blend into 2% GDP growth 1:18:24 โ€“ ASI 1:33:38 โ€“ Evolution of intelligence & culture 1:43:43 - Why self
504
3K
18K
@ericliuof97
Minghuan Liu
7 days
I recently realized that the morphology of whole-body robots (beyond table top) shapes how we design teleoperation interfaces โ€” and those interfaces dictate how efficiently we can collect data. So maybe the question isnโ€™t just about better teleopโ€ฆ Itโ€™s whether we should design
1
0
4
@googleaidevs
Google AI Developers
9 days
With Gemini CLI's new pseudo-terminal (PTY) support, you can run complex, interactive commands like vim, top, or git rebase -i directly within the CLI without having to exit, keeping everything in context.
77
229
2K
@moritz_reuss
Moritz Reuss
11 days
VLAs have become the fastest-growing subfield in robot learning. So where are we now? After reviewing ICLR 2026 submissions and conversations at CoRL, I wrote an overview of the current state of VLA research with some personal takes: https://t.co/OMMdB1MHtS
Tweet card summary image
mbreuss.github.io
Comprehensive analysis of Vision-Language-Action models at ICLR 2026 - discrete diffusion, reasoning VLAs, and benchmark insights.
10
100
524
@ericliuof97
Minghuan Liu
7 days
Very cool work!
@bingyikang
Bingyi Kang
9 days
How can an AI model learn the underlying dynamics of a visual scene? We're introducing Trajectory Fields, a new way to represent video in 4D! It models the path of each pixel as a continuous 3D trajectory, which is parameterized by a B-spline function of time. This unlocks
0
0
1
@sainingxie
Saining Xie
9 days
thereโ€™s only one right answer here, the @ylecun definition, and everyone should be able to recite it word for word
@JustinLin610
Junyang Lin
9 days
what is a world model?
61
204
2K
@kunlei15
Kun Lei
9 days
Introducing RL-100: Performant Robotic Manipulation with Real-World Reinforcement Learning. https://t.co/tZ0gz6OTdb 7 real robot tasks, 900/900 successes. Up to 250 consecutive trials in one task, running 2 hours nonstop without failure. High success rate against physical
15
68
358
@ericliuof97
Minghuan Liu
8 days
Interesting...
@imankitgoyal
Ankit Goyal
9 days
What's the right architecture for a VLA? VLM + custom action heads (ฯ€โ‚€)? VLM with special discrete action tokens (OpenVLA)? Custom design on top of the VLM (OpenVLA-OFT)? Or... VLM with ZERO modifications? Just predict action as text. The results will surprise you. VLA-0:
0
0
3
@ericliuof97
Minghuan Liu
14 days
๐Ÿ™Œ
@xbpeng4
Jason Peng
16 days
Implementing motion imitation methods involves lots of nuisances. Not many codebases get all the details right. So, we're excited to release MimicKit! https://t.co/7enUVUkc3h A framework with high quality implementations of our methods: DeepMimic, AMP, ASE, ADD, and more to come!
0
0
2
@ericliuof97
Minghuan Liu
14 days
After carefully watching the video I guess they are transferred mainly through wrist camera (all robots usetvery similar wrist camera setup) along with relative actions (aligned with wrist cam obs)?
@bousmalis
Konstantinos Bousmalis
24 days
You have to watch this! For years now, I've been looking for signs of nontrivial zero-shot transfer across seen embodiments. When I saw the Alohas unhang tools from a wall used only on our Frankas I knew we had it! Gemini Robotics 1.5 is the first VLA to achieve such transfer!!
1
3
33
@AnkaHeChen
Anka He Chen
21 days
Penetration-free for free, you need OGC. I released the code of our SIGGRAPH 2025 paper: Offset Geometric Contact, where we made real-time, penetration free simulation possible, with @JerryHsu32 @zihengliu @milesmacklin @YinYang24414350 @cem_yuksel Page: https://t.co/LsfX0lbYiJ
61
423
4K
@zzlccc
Zichen Liu
19 days
GEMโค๏ธTinker GEM, an environment suite with a unified interface, works perfectly with Tinker, the API by @thinkymachines that handles the heavy lifting of distributed training. In our latest release of GEM, we 1. supported Tinker and 5 more RL training frameworks 2. reproduced
5
33
285
@linkevin0
Kevin Lin
27 days
Tired of collecting robot demos? ๐Ÿš€ Introducing CP-Gen: geometry-aware data generation for robot learning. From a single demo, CP-Gen generates thousands of new demonstrations to train visuomotor policies that transfer zero-shot sim-to-real across novel geometries and poses.
3
34
123
@willyayy
will yeh แฏ…/acc
29 days
1x's NEO Humanoid has the LOWEST latency VR teleoperation I've ever seen! It matches nearly instantly!
39
177
2K
@stepjamUK
Stephen James
1 month
๐—”๐—ณ๐˜๐—ฒ๐—ฟ ๐Ÿญ๐Ÿฌ+ ๐˜†๐—ฒ๐—ฎ๐—ฟ๐˜€ ๐—ถ๐—ป ๐—ฟ๐—ผ๐—ฏ๐—ผ๐˜ ๐—น๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด, from my PhD at Imperial to Berkeley to building the Dyson Robot Learning Lab, one frustration kept hitting me: ๐—ช๐—ต๐˜† ๐—ฑ๐—ผ ๐—œ ๐—ต๐—ฎ๐˜ƒ๐—ฒ ๐˜๐—ผ ๐—ฟ๐—ฒ๐—ฏ๐˜‚๐—ถ๐—น๐—ฑ ๐˜๐—ต๐—ฒ ๐˜€๐—ฎ๐—บ๐—ฒ ๐—ถ๐—ป๐—ณ๐—ฟ๐—ฎ๐˜€๐˜๐—ฟ๐˜‚๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ ๐—ผ๐˜ƒ๐—ฒ๐—ฟ ๐—ฎ๐—ป๐—ฑ
15
84
712
@JieWang_ZJUI
Jie Wang
2 months
@GRASPlab @physical_int ๐Ÿงต6/ PI0 is actually a FPV player: It mainly relies on wrist camera. In fact, it still works even if the side-view camera is blocked. 3rd camera view contains more variations and change, the neural network may focus more on what the gripper see to execute tasks.
1
1
17
@rutavms
Rutav
1 month
Intelligent humanoids should have the ability to quickly adapt to new tasks by observing humans Why is such adaptability important? ๐ŸŒ Real-world diversity is hard to fully capture in advance ๐Ÿง  Adaptability is central to natural intelligence We present MimicDroid ๐Ÿ‘‡ ๐ŸŒ
7
40
121