Aravind Rajeswaran
@aravindr93
Followers
3K
Following
1K
Media
34
Statuses
265
Building AI agents. Developer of DAPG, Decision Transformers, R3M, OpenEQA, and Implicit MAML. PhD from @uwcse and BTech from @iitmadras
San Francisco Bay Area
Joined May 2015
To my friends at FAIR and anyone who might be impacted by layoffs, please stay strong. The AI world out there is exciting and dynamic. FAIR was great, but all good things must come to an end. Happy to talk anytime and share leads for interesting opportunities. Get in touch!
0
0
25
I've been thinking about doing this for a while. And no better time than now in the aftermath of the H1B issue. Indian AI startup landscape is exploding. For anyone looking for a job at some of the best companies, I'd love put my network to work. - We've curated 20 amazing
docs.google.com
We’ve curated 20 of the most exciting growth-stage AI startups in India. If you’re exploring opportunities, fill out this quick form and we’ll try to connect you directly with relevant founders. An...
18
30
243
📢Life is a sequence of bets – and I’ve picked my next: @MyolabAI It’s incredibly ambitious, comes with high risk, & carries unbounded potential. But it’s a version of the #future I deeply believe in. I believe: ➡️AI will align strongly with humanity - coz it maximizes its own
myolab.ai
Building Human-Embodied Intelligence to Empower Humans.
All forms of intelligence co-emerged with a body, except AI We're building a #future where AI evolves as your lifelike digital twin to assist your needs across health, sports, daily life, creativity, & beyond... https://t.co/QL3o9YxZYz ➡️ Preview your first #HumanEmbodiedAI
17
21
109
Excited to share my project from FAIR! As a side quest to 3D localization, we also ended up developing 3D-JEPA, the first large scale SSL method for 3D. I'm optimistic 3D-JEPA and Locate3D will play a major role in robotics and improving spatial AI. Try the model for your apps!
Introducing Meta Locate 3D: a model for accurate object localization in 3D environments. Learn how Meta Locate 3D can help robots accurately understand their surroundings and interact more naturally with humans. You can download the model and dataset, read our research paper,
2
2
54
Really exciting VLA demo from @Figure_robot ! The hericahical architecture they use here is very similar to Latent Code as Bridges (LCB) from @YideShentu et. al. Cool to see them scale this type approach scaled to such success! https://t.co/K9LpQc4tFo 🧵
Meet Helix, our in-house AI that reasons like a human Robotics won't get to the home without a step change in capabilities Our robots can now handle virtually any household item:
5
20
198
Interested in knowing the difference between "world" models and "word" models? Come talk to us at the #OpenEQA poster today at #CVPR (Arch 4A-E Poster #179) https://t.co/9EC3coHeJd
Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ https://t.co/vBFG7Z58kQ
0
6
13
New work with @YideShentu! We present Latent Code as Bridges🌉 or LCB. We leverage LLMs for robotics by finetuning the LLM to output latent embeddings to control the lower level policy. This hierarchical approach enables end2end training while maintaining the advantages of LLMs!
2
40
139
Wouldn't it be cool to just talk to your robot 🤖 or smart glass agent 🕶️ in natural language? ✅ Hey 🕶️ do you remember where I left my badge? ✅ Hey 🤖 I'm hungry and driving back from work, do I have any fruits left? We're (@AIatMeta) working on this Embodied Question
3
14
53
Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ https://t.co/vBFG7Z58kQ
39
253
1K
Sad to see the saga unfold at @OpenAI, I was one of the early interns, and still have several friends and mentors there. I'll resist the urge to make elevator pitches for different labs at this stage. Folks, let the poor souls catch a break! Crisis calls for judgement and
0
0
6
🚀 Excited to announce the open-sourcing of #𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲! It brings together numerous strands of work from talented collaborators over the years! ✅ Dexterous hands ✅ Quadrupeds on randomized surfaces ✅ Musculoskeletal models ✅ Visually rich domains ✅ Large
📢#𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲 - a unified robot learning framework ✅Designed for genralizn first robot-learning era ✅Diverse (500 envs, 8 domain) ✅Single flag for Sim<>Real ✅TeleOper Support ✅Multi-(Skill x Task) realworld dataset ✅pip install robohive https://t.co/CWSipPv8M7 🧵👇
0
5
14
📢#𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲 - a unified robot learning framework ✅Designed for genralizn first robot-learning era ✅Diverse (500 envs, 8 domain) ✅Single flag for Sim<>Real ✅TeleOper Support ✅Multi-(Skill x Task) realworld dataset ✅pip install robohive https://t.co/CWSipPv8M7 🧵👇
3
32
193
🤖 Excited to share MoDem-v2, our latest work on decoder-free (JEPA-style) world models. 💡 Provides a simple three-step recipe to extend world model-style algorithms (e.g., TD-MPC) to real-world robots. 🙌 Fun collaboration with @palanc_, @ncklashansen, and @Vikashplus. Thanks
Check out MoDem-V2 - our latest work on Visuo-Motor World Models for Real-World Robot Manipulation! 🧵🔽 Website: https://t.co/Vkx4nwsjZR Paper: https://t.co/gtEATtfepw
1
4
22
#𝗥𝗼𝗯𝗼𝗔𝗴𝗲𝗻𝘁 -- A universal multi-task agent on a data-budget 💪 with 12 non-trivial skills 💪 can generalize them across 38 tasks 💪& 100s of novel scenarios! 🌐 https://t.co/37YbJP05Zt w/ @mangahomanga @jdvakil @m0hitsharma, Abhinav Gupta, @shubhtuls
6
67
252
LfS revisited for control: ✅ A well-tuned LfS baseline is competitive with pre-trained visual rep. (PVR) ✅ But fine-tuning with aug improves PVR results ✅ And better benchmarks are needed! 📢 Tue 2-3.30pm HST, Hall-1 #711
https://t.co/5yv1UjuldD
I'll be presenting our work on pretrained representations for control vs. learning from scratch at #ICML2023🏝️ next week! Reach out or stop by poster 711 on Tue (session 2) for a chat about the future of RL! https://t.co/qNRf4YwjIE Details below 👇 1/n
0
0
1
🌟 MTM is a new SSL paradigm for decision making 🚀 Builds on decision transformer, BERT, and MAE 🔥 A single MTM transformer is versatile and can function as a world model, policy, representation, and more! 🔍🧵 📢 Thu 1.30-3pm HST, Hall-1 #402
https://t.co/Au200XqpcX
Introducing Masked Trajectory Modeling (MTM), a new general-purpose framework for sequential decision making. A single transformer trained with MTM can exhibit multiple capabilities by simply choosing different masking patterns at inference time. Accepted at ICML 2023. 🧵👇
1
1
4
✈️ Just landed in Hawaii 🌴 to present two cool projects at #ICML2023 🚀 Masked Trajectory Models (w/ @philippswu, @arjunmajum, @kevinleestone, @yixin_lin_ , @IMordatch, @pabbeel) 📚 LfS Revisited (w/ @ncklashansen, @haosu_twitr, @HarryXu12, @xiaolonw et al.) Details in 🧵👇
3
2
25
Excited to share our #ICML2023 paper ✨LIV✨! Extending VIP, LIV is at once a pre-training, fine-tuning, and (zero-shot!) multi-modal reward method for (real-world!) language-conditioned robotic control. Project: https://t.co/HZuZnKOqPL Code & Model: https://t.co/J30MOankDW 🧵:
2
45
161