Jason Ma Profile
Jason Ma

@JasonMa2020

Followers
2,092
Following
1,015
Media
61
Statuses
481

PhD @Penn and @GoogleDeepMind . Apple Scholars in AI/ML. Prev: @NVIDIAAI , @MetaAI , @Harvard . Foundation Models for Robotics

Palo Alto, CA
Joined August 2018
Don't wanna be here? Send us removal request.
Explore trending content on Musk Viewer
Pinned Tweet
@JasonMa2020
Jason Ma
25 days
Introducing DrEureka🎓, our latest effort pushing the frontier of robot learning using LLMs! DrEureka uses LLMs to automatically design reward functions and tune physics parameters to enable sim-to-real robot learning. DrEureka can propose effective sim-to-real configurations
24
112
586
@JasonMa2020
Jason Ma
7 months
Super excited to share Eureka, our "spin" on how to use LLMs to teach low-level dexterity skills! Eureka is an open-ended reward design agent that can write and evolve superhuman reward functions for a large suite of robots and tasks, including challenging pen spinning tricks!
@DrJimFan
Jim Fan
7 months
Can GPT-4 teach a robot hand to do pen spinning tricks better than you do? I'm excited to announce Eureka, an open-ended agent that designs reward functions for robot dexterity at super-human level. It’s like Voyager in the space of a physics simulator API! Eureka bridges the
188
1K
6K
14
20
180
@JasonMa2020
Jason Ma
2 years
Excited to share VIP, a self-supervised visual reward and representation pre-trained on diverse human videos! VIP’s frozen reward and rep. can solve diverse unseen robot tasks using TrajOpt, online RL, and enables real-world few-shot offline RL! 🧵:
2
40
170
@JasonMa2020
Jason Ma
1 year
Excited to share our #ICML2023 paper ✨LIV✨! Extending VIP, LIV is at once a pre-training, fine-tuning, and (zero-shot!) multi-modal reward method for (real-world!) language-conditioned robotic control. Project: Code & Model: 🧵:
2
44
163
@JasonMa2020
Jason Ma
3 months
Humbled to share that I was selected as an Apple Scholar in AIML PhD Fellowship! Very grateful to Apple, my advisors @dineshjayaraman @obastani as well as all my mentors and collaborators for their support!
22
5
157
@JasonMa2020
Jason Ma
7 months
Excited to share my first paper as an "advisor" :D We show that pre-trained visual representations enable a simple, fast, no-training subgoal decomposition method for long-horizon robotic manipulation! Paper: Website: (🧵1/n)
3
20
129
@JasonMa2020
Jason Ma
7 months
I am attending #CORL2023 and presenting two new papers at various workshops! Excited to make new friends and catch up! Please reach out if you are attending and would like to chat about anything robot learning :)
4
16
119
@JasonMa2020
Jason Ma
16 days
This is so impressive! I can't imagine the amount of progress we will unlock as a community with low-cost, highly capable robots. Congrats to the Unitree Team!
@UnitreeRobotics
Unitree
16 days
Unitree Introducing | Unitree G1 Humanoid Agent | AI Avatar Price from $16K 🤩 Unlock unlimited sports potential(Extra large joint movement angle, 23~34 joints) Force control of dexterous hands, manipulation of all things Imitation & reinforcement learning driven #Unitree #AI
309
752
3K
2
7
53
@JasonMa2020
Jason Ma
25 days
Learning policies in simulation and transferring to the real world (or Sim-To-Real in short) is a promising strategy for robots to learn complex skills. However, humans need to tune the simulator carefully so that the policies work robustly in the real world: this is difficult,
2
8
50
@JasonMa2020
Jason Ma
11 months
Due to popular requests, we have now uploaded our pre-trained LIV model on HuggingFace for easier downloads! This is my first time doing it, and the experience was quite smooth @_akhaliq
@JasonMa2020
Jason Ma
1 year
Excited to share our #ICML2023 paper ✨LIV✨! Extending VIP, LIV is at once a pre-training, fine-tuning, and (zero-shot!) multi-modal reward method for (real-world!) language-conditioned robotic control. Project: Code & Model: 🧵:
2
44
163
1
17
49
@JasonMa2020
Jason Ma
10 months
We are presenting LIV today at #ICML2023 ! Exhibit Hall 1, #827 2:00pm - 3:30pm HST The future of robotics is multi-modal, and LIV demonstrates how multi-modal value pre-training from diverse human videos can bootstrap language-conditioned robot skill learning. See you there!
@JasonMa2020
Jason Ma
1 year
Excited to share our #ICML2023 paper ✨LIV✨! Extending VIP, LIV is at once a pre-training, fine-tuning, and (zero-shot!) multi-modal reward method for (real-world!) language-conditioned robotic control. Project: Code & Model: 🧵:
2
44
163
1
6
41
@JasonMa2020
Jason Ma
5 months
Honored to see Eureka on this list along side many amazing works!
@NVIDIAAIDev
NVIDIA AI Developer
6 months
👀 Discover the top 10 #NVIDIAresearch projects of the year. ✨ From Neuralangelo's high-fidelity neural surface reconstruction to Magic3D's text-to-3D content creation, these projects push the boundaries of innovation in #AI .
0
11
52
1
2
40
@JasonMa2020
Jason Ma
10 months
I am attending #ICML2023 next week in Hawaii! Excited to make new friends and re-connect with old ones! Please reach out if you are attending and would like to chat about anything related to research or ML! My particular interests include foundation models, RL, and robotics!
0
0
34
@JasonMa2020
Jason Ma
9 months
We are organizing Workshop on Goal-Conditioned Reinforcement Learning (GCRL) at #NeurIPS 2023! Submission Deadline: October 4th, 2023 Website:
1
4
32
@JasonMa2020
Jason Ma
3 months
Big congrats to my mentors and collaborators @DrJimFan and @yukez on the new group! Embodied AI and robotics research just kicked up a gear ;)
@DrJimFan
Jim Fan
3 months
Career update: I am co-founding a new research group called "GEAR" at NVIDIA, with my long-time friend and collaborator Prof. @yukez . GEAR stands for Generalist Embodied Agent Research. We believe in a future where every machine that moves will be autonomous, and robots and
Tweet media one
241
471
4K
1
1
31
@JasonMa2020
Jason Ma
25 days
At a technical level, DrEureka, following our prior work Eureka (), uses LLM-guided evolutionary search to generate safety-aware reward functions in code that can be used to train policies in sim. Then, leveraging LLMs’ capability as hypothesis generators,
3
2
28
@JasonMa2020
Jason Ma
8 months
Excited to share some of my recent works on pre-training for robotics with the MILA community!
@MontrealRobots
REAL - Robotics and Embodied AI Lab
8 months
Hello! We have @JasonMa2020 from UPenn giving a talk at this week's robot learning seminar (Thursday 11:30am EST online). Hope to see you all there! Title: Foundation Reward Models for General Robot Skill Acquisition #Robotics #MachineLearning
0
1
14
0
3
27
@JasonMa2020
Jason Ma
1 year
Thanks @_akhaliq ! LIV is now on arXiv: Check it out if you are interested in the space of (RL-based) vision-language pre-training for robotics! Happy to answer any questions about the paper :)
@_akhaliq
AK
1 year
LIV: Language-Image Representations and Rewards for Robotic Control paper page: Language-Image Value (LIV) is a unified pre-training, fine-tuning, and reward learning algorithm for language-conditioned visual manipulation. LIV can perform zero-shot
0
41
184
0
7
25
@JasonMa2020
Jason Ma
7 months
Happy to announce that UVD is announced as the best paper at the CORL LEAP Workshop!
Tweet media one
@JasonMa2020
Jason Ma
7 months
I am attending #CORL2023 and presenting two new papers at various workshops! Excited to make new friends and catch up! Please reach out if you are attending and would like to chat about anything robot learning :)
4
16
119
2
2
25
@JasonMa2020
Jason Ma
25 days
DrEureka is co-led by @willjhliang and me with collaborators from @Penn and @NVIDIA : @johnnywang_16 , @sam_wang23 , and our advisors @yukez @DrJimFan @obastani @dineshjayaraman Check out our project website for the paper and more videos: Code:
4
1
24
@JasonMa2020
Jason Ma
1 year
We have added example code for visualizing **animated** zero-shot VIP reward curve on robot videos! Try it out (on your own robot videos) here:
@JasonMa2020
Jason Ma
2 years
Excited to share VIP, a self-supervised visual reward and representation pre-trained on diverse human videos! VIP’s frozen reward and rep. can solve diverse unseen robot tasks using TrajOpt, online RL, and enables real-world few-shot offline RL! 🧵:
2
40
170
0
9
21
@JasonMa2020
Jason Ma
1 year
Check out our #L4DC paper on learning policy-aware dynamics model for reinforcement learning! The idea is very simple: focus model learning on the current policy’s visitation distribution. We theoretically show why this is desirable and extend the dual RL paradigm to MBRL,
@kausiksivakumar
Kausik Sivakumar
1 year
Excited to share our our #L4DC2023 paper that introduces "Transition Occupancy Matching"(TOM) TOM learns a dynamics model that keeps up with the improving policy, facilitating continued progress Paper 📰: Code 💻: 🧵
1
2
12
0
1
21
@JasonMa2020
Jason Ma
2 years
I am attending #NeurIPS2022 next week to present several works on offline RL and pre-training for robotics! Would love to meet people and discuss anything, in particular, RL and robot learning topics! DM me if you are attending and want to chat or grab coffee 😀
1
0
21
@JasonMa2020
Jason Ma
2 months
We at @GRASPlab hosted @ericjang11 last week for a thought-provoking talk on Humanoid robots and 1X! The full recording is up now, check it out!
@xiao_ted
Ted Xiao
2 months
Nice talk on the technical approach to intelligent humanoids at 1X! As usual for @ericjang11 ’s spicy takes, I agree strongly with 60%, ambivalent on 20%, and disagree with 20%. Highly recommend a watch! 💯
2
11
52
1
1
21
@JasonMa2020
Jason Ma
6 months
Jim has been a fantastic mentor for me! Apply if you are interested in foundation models for decision making and AI Agents!
1
1
20
@JasonMa2020
Jason Ma
25 days
This was a very fun project and we learned a lot about how to use LLMs to enable robot skill learning! There are many challenges and potential future directions. For example, how to combine DrEureka with real-world execution feedback and using vision to provide feedback on reward
2
0
19
@JasonMa2020
Jason Ma
6 months
Come to our workshop on Goal-Conditioned RL tomorrow at #NeurIPS2023 !
@gcrl_workshop
Goal-Conditioned RL Workshop
6 months
Check out the #NeurIPS2023 workshop on Goal Conditioned Reinforcement Learning: * Tomorrow (Friday) 900 -- 1830 CST * Great speakers: @jeffclune @r_mirsky @olexandr @ybisk @SusanMurphylab1 * Program:
0
2
17
0
2
18
@JasonMa2020
Jason Ma
1 year
Reward learning is a fundamental challenge in RL. In VIP, we address this by pre-training a value function on action-free human videos, and the pre-trained VIP value function can zero-shot transfer to unseen robot tasks! Find out more about VIP at the Deep RL workshop tomorrow!
@hausman_k
Karol Hausman
1 year
Deep RL workshop @NeurIPSConf starts on Friday! Hear invited talks from @tobigerstenberg on counterfactual simulation of causal judgments, @j_foerst on opponent-shaping in games, @IMordatch on sequence modeling, @yayitsamyzhang on learning generalist agents. Don't miss out! 🧵
Tweet media one
1
20
81
0
1
18
@JasonMa2020
Jason Ma
10 days
sim2real for manipulation tasks is really hard! Great to see this work come out
@YunfanJiang
Yunfan Jiang
11 days
Does your sim2real robot falter at critical moments 🤯? Want to help but unsure how, all you can do is reward tuning in sim 😮‍💨? Introduce 𝐓𝐑𝐀𝐍𝐒𝐈𝐂 for manipulation sim2real. Robots learned in sim can accomplish complex tasks in real, such as furniture assembly. 🤿🧵
16
40
176
2
2
17
@JasonMa2020
Jason Ma
7 months
UVD code is open-source now: Get your long-horizon demonstrations segmented in few seconds! We support various SOTA pre-trained reps (VIP, R3M, LIV, VC-1, ...) as well as policy backbones (MLP, GPT).
@JasonMa2020
Jason Ma
7 months
Excited to share my first paper as an "advisor" :D We show that pre-trained visual representations enable a simple, fast, no-training subgoal decomposition method for long-horizon robotic manipulation! Paper: Website: (🧵1/n)
3
20
129
0
3
15
@JasonMa2020
Jason Ma
25 days
We highlight quadruped yoga ball walking. This task is particularly hard because (1) it is a novel task for which the LLM could not have seen human-generated reward functions or DR, and (2) simulation cannot model the deformable surface of an air-inflated ball, making good
3
0
16
@JasonMa2020
Jason Ma
2 months
@ericjang11
Eric Jang
2 months
I'll be giving a talk at @GRASPlab on Wednesday, 3/27 on the robot learning we're doing at @1x_tech . If you're a researcher at Penn working on similar things I'd love to visit your labs and see what you're working on as well! Please DM
0
3
37
0
0
15
@JasonMa2020
Jason Ma
11 months
@HaqueIshfaq We are organizing a workshop on goal-conditioned RL! Please do consider submitting/participating if relevant :) more details to follow.
1
1
15
@JasonMa2020
Jason Ma
1 year
Giving contributed talks on VIP at the Offline RL, Foundation Model for Decision Making, and Deep RL workshops at #NeurIPS2022 . Come check out how we can pre-train a value function on passive human data and zero-shot transfer to robotics manipulation!
@Vikashplus
Vikash Kumar ✈️ICRA2024
1 year
#VIP : Self-supervised pre-trained visual reward and representation for robotics 🎯 DeepRL, OfflineRL, SSL workshops 🔗 @shagunsodhani @dineshjayaraman @obastani @Vikashplus @yayitsamyzhang @JasonMa2020 ⏯️⏯️
1
0
3
0
0
15
@JasonMa2020
Jason Ma
25 days
The DrEureka policy, as you have seen in the uncut videos, is quite robust in the real world. It can successfully stay balanced over curbs, terrain changes, and even when the ball is being kicked!
1
0
14
@JasonMa2020
Jason Ma
7 months
This is my internship project at @NVIDIAAI ! I had a blast working on it and learned a lot from the experience. I am really grateful to my mentors @AnimaAnandkumar @DrJimFan @yukez for their guidance on the project!
2
0
14
@JasonMa2020
Jason Ma
5 months
@tonyzzhao @zipengfu @chelseabfinn Congrats Tony!! Very impressive :)
1
0
11
@JasonMa2020
Jason Ma
25 days
We even tried to deploy the policy while the yoga ball was being deflated. Here is a ~1-minute uncut video of the robot balancing well until it eventually loses control😅
1
0
13
@JasonMa2020
Jason Ma
1 year
We ( @yayitsamyzhang ) are presenting VIP at #ICLR tomorrow! Talk: Oral 1 Track 5: Reinforcement Learning Poster Session: MH1-2-3-4 #118 Looking forward to having discussions with you there!
@JasonMa2020
Jason Ma
2 years
Excited to share VIP, a self-supervised visual reward and representation pre-trained on diverse human videos! VIP’s frozen reward and rep. can solve diverse unseen robot tasks using TrajOpt, online RL, and enables real-world few-shot offline RL! 🧵:
2
40
170
0
1
13
@JasonMa2020
Jason Ma
7 months
This looks fantastic!
@RL_Conference
RL_Conference
7 months
Thrilled to announce the first annual Reinforcement Learning Conference @RL_Conference , which will be held at UMass Amherst August 9-12! RLC is the first strongly peer-reviewed RL venue with proceedings, and our call for papers is now available: .
Tweet media one
4
95
242
0
0
12
@JasonMa2020
Jason Ma
1 year
Check out our #AISTATS2023 paper on fair exploration in RL! Led by @wanqiao_xu
@hamsabastani
Hamsa Bastani
1 year
The exploration-exploitation tradeoff in #RL raises concerns about exploration -- who does it impact and how much? In a recent #AISTATS paper, we show how to effectively "spread out" exploration across episodes (individuals) w/ only a small cost to regret:
1
8
34
0
0
12
@JasonMa2020
Jason Ma
25 days
Beyond yoga ball walking, we also benchmarked DrEureka on Quadruped Locomotion and Dexterous Cube Rotation, two known tasks where there are pre-existing human-designed reward functions and DR configurations. We find DrEureka configurations to match or outperform human-designed
1
0
11
@JasonMa2020
Jason Ma
1 year
A very cool way of leveraging human videos for robotics manipulation!
@mangahomanga
Homanga Bharadhwaj
1 year
How can robots learn manipulation *just* by watching videos of humans in different unstructured settings? In our new paper, we develop a framework enabling zero-shot coarse robot manipulation from passive human videos (a 🧵) w/ Abhinav Gupta, @shubhtuls , @Vikashplus 1/N
5
37
198
1
1
11
@JasonMa2020
Jason Ma
1 year
One visual representation for a wide variety of manipulation and navigation tasks! Check out our recent work on building a Visual Cortex for Embodied AI.
@DhruvBatraDB
Dhruv Batra
1 year
Contemporary discussion (hype?) about LLMs and “pausing AGI development” seems oblivious of Moravec’s paradox. We’ve hypothesized since the 80s — that the hardest problems in AI involve sensorimotor control, not abstract thought or reasoning. It
22
153
811
0
2
10
@JasonMa2020
Jason Ma
9 months
Looking forward to this talk!
@GRASPlab
GRASP Laboratory
9 months
The GRASP SFI series is back for the Fall Semester! Please join us TODAY from 3pm - 4pm EST as Dr. Jim Fan presents "Generalist Agents in Open-Ended Worlds". For more info and how to join, please visit: #GRASP #GRASPLab #GRASPSFI
Tweet media one
0
4
10
0
0
7
@JasonMa2020
Jason Ma
8 months
@MontrealRobots @dineshjayaraman Thanks for having me! I really enjoyed the conversation with everyone
0
0
7
@JasonMa2020
Jason Ma
7 months
Please check out the paper for more details and experimental results! This is a fun collaboration with @ZCCZHANG @YunshuangL @obastani @abhishekunique7 @dineshjayaraman @LucaWeihs Also check out @ZCCZHANG 's original thread on the project:
@ZCCZHANG
Zichen "Charles" Zhang
7 months
How can pre-trained visual representations help solve long-horizon manipulation? 🤔 Introducing Universal Visual Decomposer (UVD), an off-the-shelf method for identifying subgoals from videos - NO extra data, training, cost, or task knowledge required. (🧵1/n)
1
17
87
1
2
7
@JasonMa2020
Jason Ma
7 months
With UVD's recursive decomposition⏮️, it can seamlessly build upon any standard visuomotor policy training⏭️, in Sim & Real, across IL & RL!
1
2
6
@JasonMa2020
Jason Ma
7 months
2. Human interference during deployment UVD can also enable agents to auto-skip sub-stages preemptively finished by humans and can reset to redo certain stages during deployment. (🧵5/n)
1
2
6
@JasonMa2020
Jason Ma
1 year
Check out some recent works from our lab on versatile goal-based task specification for robot learning!
@EpisodeYang
Ge Yang
1 year
Happening now! @dineshjayaraman from @PennEngineers is giving a talk at the Embodied Intelligence Seminar @MIT_CSAIL on Polyglot Robots: Versatile Goal-Based Task Specification for Robot Learning Streaming at: , cohosted with @du_yilun
Tweet media one
0
1
8
0
0
6
@JasonMa2020
Jason Ma
22 days
0
0
2
@JasonMa2020
Jason Ma
1 year
Detailed Thread on VC-1, a first step towards a unified and single visual representation for all Embodied AI tasks
@_kainoa_
Franziska Meier
1 year
A visual cortex is the region of the brain that (together with the motor cortex) enables an organism to convert vision into movement. We present an artificial visual cortex — the module in an AI system that enables an artificial agent to convert camera input into actions. 🧵👇
1
51
204
0
1
6
@JasonMa2020
Jason Ma
7 months
@EugeneVinitsky You are too kind Eugene! I very much look forward to learning about your neat work soon ;)
0
0
5
@JasonMa2020
Jason Ma
7 months
Pre-trained reprs, e.g VIP (), can produce smooth embeddings on videos of atomic tasks. Given this, UVD discovers subgoals by recursively detecting phase shifts in the embedding space. This simple idea works well across human or robot videos! (🧵2/n)
1
2
5
@JasonMa2020
Jason Ma
2 years
Shagun is a great mentor! Highly recommend working with him :)
@shagunsodhani
Shagun Sodhani
2 years
We are hiring a #research #intern at FAIR ( @MetaAI ) to work in areas related to #RL , #hierarchical RL, #modular #networks , and #world #models . Location: Montreal / New York / Remote. You can dm me your questions and resume!
4
34
287
0
0
5
@JasonMa2020
Jason Ma
1 year
Cool work on RL based pre-training on observational data! If you are interested in this line of work, also check out our work VIP (), which directly obtains a visual representation and universal value function from human videos!
@its_dibya
Dibya Ghosh
1 year
"Ask not what representation learning can do for RL, ask what RL can do for representation learning" -- JFK? In our new paper: a useful way to pre-train representations on video data for decision-making agents is basically to run RL on the video!
1
5
67
0
1
5
@JasonMa2020
Jason Ma
2 months
@DrJimFan Congrats Jim! Killing it with project naming per always ;)
1
0
2
@JasonMa2020
Jason Ma
1 year
Cool work that explores how interaction can be used to verify task completion in the real world!
@edward_s_hu
Edward Hu
1 year
Really thrilled that our paper was picked a best paper finalist at CORL22! If you’re attending, come to our oral on Saturday at 11:20am! We present Interactive Reward Function (IRF) policies, which interact with the world to provide rewards for training task policies. (1/3)
1
5
43
0
0
5
@JasonMa2020
Jason Ma
7 months
Importantly, UVD enables long-horizon compositional generalization across: 1. New task sequences Plug-and-play UVD significantly boosts OoD results across various visual backbones and policies, while maintaining in-domain (InD) performance. (🧵4/n)
Tweet media one
Tweet media two
1
2
5
@JasonMa2020
Jason Ma
3 months
@DrJimFan @yukez My Eureka project done in an internship at the group was also selected as one of the "Top 10 NVIDIA Research Projects of 2023"! This speaks to the mission-focused nature of the group and how even a junior research can make substantial impacts!
@DrJimFan
Jim Fan
3 months
Eureka was named the "Top 10 NVIDIA Research Projects of 2023":
8
12
122
0
0
4
@JasonMa2020
Jason Ma
7 months
@mihdalal Nice work, congrats!!
0
0
2
@JasonMa2020
Jason Ma
1 year
@natolambert In my own VIP work (), we explore how fixed pre-trained models fare against increasing compute budget in downstream trajectory optimization:
Tweet media one
0
0
4
@JasonMa2020
Jason Ma
1 year
@GMartius @nico_guertler @_sebastianblaes @manuelwuethrich @robo_challenge Super cool work! This seems like a really useful and challenging benchmark for testing real-world offline RL. We have also done some work on offline goal-conditioned RL for real-world dexterous manipulation that you may be interested!
0
2
4
@JasonMa2020
Jason Ma
9 months
Finally, big thanks to the organizing team: @ben_eysenbach @TongzhouWang @IshanDurugkar @TheAndiPenguin @yayitsamyzhang We are super excited for the workshop and look forward to all the great discussions that will take place!
0
0
3
@JasonMa2020
Jason Ma
6 months
@notmahi Super cool work!! Congrats @notmahi @LerrelPinto
0
0
3
@JasonMa2020
Jason Ma
2 years
The code and pre-trained model are open-sourced: It is also hosted on @torchRL as an encoder option! It is super easy to load VIP for your new task: from vip import load_vip vip = load_vip() vip.eval()
Tweet media one
1
0
3
@JasonMa2020
Jason Ma
7 months
@sytelus This is excellent analysis -- thank you Shital!
0
0
3
@JasonMa2020
Jason Ma
6 months
@AnimaAnandkumar Congratulations, Anima! It was a pleasure working and learning from you!
0
0
1
@JasonMa2020
Jason Ma
4 years
@hima_lakkaraju Congrats Alexis !!
0
0
3
@JasonMa2020
Jason Ma
7 months
Our second paper is UVD: UVD is a simple, fast, no-training subgoal decomposition method for long-horizon robotic manipulation using PVRs (e.g., VIP, R3M)! @ZCCZHANG @YunshuangL will be giving the workshop spotlight!
@ZCCZHANG
Zichen "Charles" Zhang
7 months
How can pre-trained visual representations help solve long-horizon manipulation? 🤔 Introducing Universal Visual Decomposer (UVD), an off-the-shelf method for identifying subgoals from videos - NO extra data, training, cost, or task knowledge required. (🧵1/n)
1
17
87
0
0
3
@JasonMa2020
Jason Ma
1 year
Very cool paper with an awesome codebase! Kudos to the authors!
@chichengcc
Cheng Chi
1 year
What if the form of visuomotor policy has been the bottleneck for robotic manipulation all along? Diffusion Policy achieves 46.9% improvement vs prior StoA on 11 tasks from 4 benchmarks + 4 real world tasks! (1/7) website : paper:
9
100
533
0
0
3
@JasonMa2020
Jason Ma
2 years
VIP enables a simple and practical real-world few-shot offline RL pipeline: just do reward-weighted regression (RWR) with VIP’s reward and the representation! With VIP, offline RL is as simple as BC but far more effective.
1
1
3
@JasonMa2020
Jason Ma
1 year
LIV was a fun and rewarding project with my advisors at @MetaAI and @PennEngineers : @Vikashplus @yayitsamyzhang @obastani @dineshjayaraman Please check out the paper for more results and in-depth analysis!
1
0
2