Ralf Römer
@ralfroemer99
Followers
79
Following
6
Media
3
Statuses
12
PhD Student in Robot Learning @TUM
Munich, Germany
Joined March 2023
7/8 Safety: VLAs still fail in many situations. Offline safety alignment (e.g., SafeVLA) and combining online failure detection (e.g., FIPER, SAFE) with data-efficient human-in-the-loop learning (e.g., Compliant Residual DAgger, APO) hold a lot of promise to improve VLA safety.
1
0
0
6/8 Inference Speed: Not really a bottleneck anymore for VLAs thanks to new tokenization (e.g., BEAST), caching (e.g., VLA-Cache, EfficientVLA), and fast-sampling sampling methods (e.g., Two-Steps Diffusion Policy, MeanFlow).
1
0
0
5/8 Learning from Videos: This is progressing (e.g., EgoBridge, Object-centric 3D Motion Field, KungfuBot), and it seems to be easier for locomotion than for manipulation, where the embodiment gap is often bigger.
1
0
0
4.5/8 Currently, world models are mainly used in two ways in robotics: Data generation for policy training (BOOM, Newt, RoboScape) and safe policy eval (AutoVLA, ReSim). Incorporating world modeling objectives into VLA training also helps learn better representations (DreamVLA).
1
0
0
4/8 World Models: They have become quite powerful (e.g., MindJourney, PlayerOne) and are starting to become real-time capable. In autonomous driving, they already work well at a larger scale, as showcased by several companies, but things are harder for manipulation due to contact
1
0
0
3/8 RL for VLAs: This is becoming a hot topic. RL for flow matching has been studied a lot recently (ReinFlow, CQN-AS), but sample efficiency remains a challenge. For RL fine-tuning of VLAs on hard real-world tasks, a strong base policy and targeted exploration will be crucial.
1
0
0
2/8 Imitation Learning: Learning from teleoperation data is too inefficient to achieve long-term, general autonomy on its own. This is not particularly surprising, as even for LLMs, SFT (imitation) on all internet data has plateaued, so that many people are now working on RL.
1
0
0
Embodied AI & Robotics research at #NeurIPS2025 - 8 takeaways. 1/8 Continual Learning: The ability to learn from experience is still largely missing in embodied agents. A key problem is that continual deep learning (no forgetting or plasticity loss) does not yet work well.
1
0
5
I’ll be at NeurIPS this week, presenting our work on runtime failure prediction for generative policies. If you’re around and want to chat about research, please reach out! 🕒Friday, 5 December, 4:30 pm 📍Poster 2204 🌐 Website: https://t.co/fD8cH1OKmd
#NeurIPS2025 #NeurIPS
0
0
3