Aravind Rajeswaran Profile
Aravind Rajeswaran

@aravindr93

Followers
3K
Following
1K
Media
34
Statuses
265

Building AI agents. Developer of DAPG, Decision Transformers, R3M, OpenEQA, and Implicit MAML. PhD from @uwcse and BTech from @iitmadras

San Francisco Bay Area
Joined May 2015
Don't wanna be here? Send us removal request.
@aravindr93
Aravind Rajeswaran
28 days
To my friends at FAIR and anyone who might be impacted by layoffs, please stay strong. The AI world out there is exciting and dynamic. FAIR was great, but all good things must come to an end. Happy to talk anytime and share leads for interesting opportunities. Get in touch!
0
0
25
@aakrit
Aakrit Vaish
2 months
I've been thinking about doing this for a while. And no better time than now in the aftermath of the H1B issue. Indian AI startup landscape is exploding. For anyone looking for a job at some of the best companies, I'd love put my network to work. - We've curated 20 amazing
Tweet card summary image
docs.google.com
We’ve curated 20 of the most exciting growth-stage AI startups in India. If you’re exploring opportunities, fill out this quick form and we’ll try to connect you directly with relevant founders. An...
18
30
243
@Vikashplus
Vikash Kumar
5 months
📢Life is a sequence of bets – and I’ve picked my next: @MyolabAI It’s incredibly ambitious, comes with high risk, & carries unbounded potential. But it’s a version of the #future I deeply believe in. I believe: ➡️AI will align strongly with humanity - coz it maximizes its own
Tweet card summary image
myolab.ai
Building Human-Embodied Intelligence to Empower Humans.
@MyolabAI
myolab.ai
5 months
All forms of intelligence co-emerged with a body, except AI We're building a #future where AI evolves as your lifelike digital twin to assist your needs across health, sports, daily life, creativity, & beyond... https://t.co/QL3o9YxZYz ➡️ Preview your first #HumanEmbodiedAI
17
21
109
@aravindr93
Aravind Rajeswaran
7 months
Excited to share my project from FAIR! As a side quest to 3D localization, we also ended up developing 3D-JEPA, the first large scale SSL method for 3D. I'm optimistic 3D-JEPA and Locate3D will play a major role in robotics and improving spatial AI. Try the model for your apps!
@AIatMeta
AI at Meta
7 months
Introducing Meta Locate 3D: a model for accurate object localization in 3D environments. Learn how Meta Locate 3D can help robots accurately understand their surroundings and interact more naturally with humans. You can download the model and dataset, read our research paper,
2
2
54
@philippswu
Philipp Wu
9 months
Really exciting VLA demo from @Figure_robot ! The hericahical architecture they use here is very similar to Latent Code as Bridges (LCB) from @YideShentu et. al. Cool to see them scale this type approach scaled to such success! https://t.co/K9LpQc4tFo 🧵
@Figure_robot
Figure
9 months
Meet Helix, our in-house AI that reasons like a human Robotics won't get to the home without a step change in capabilities Our robots can now handle virtually any household item:
5
20
198
@aravindr93
Aravind Rajeswaran
1 year
Interested in knowing the difference between "world" models and "word" models? Come talk to us at the #OpenEQA poster today at #CVPR (Arch 4A-E Poster #179) https://t.co/9EC3coHeJd
@AIatMeta
AI at Meta
2 years
Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ https://t.co/vBFG7Z58kQ
0
6
13
@philippswu
Philipp Wu
2 years
New work with @YideShentu! We present Latent Code as Bridges🌉 or LCB. We leverage LLMs for robotics by finetuning the LLM to output latent embeddings to control the lower level policy. This hierarchical approach enables end2end training while maintaining the advantages of LLMs!
2
40
139
@aravindr93
Aravind Rajeswaran
2 years
Wouldn't it be cool to just talk to your robot 🤖 or smart glass agent 🕶️ in natural language? ✅ Hey 🕶️ do you remember where I left my badge? ✅ Hey 🤖 I'm hungry and driving back from work, do I have any fruits left? We're (@AIatMeta) working on this Embodied Question
3
14
53
@AIatMeta
AI at Meta
2 years
Today we’re releasing OpenEQA — the Open-Vocabulary Embodied Question Answering Benchmark. It measures an AI agent’s understanding of physical environments by probing it with open vocabulary questions like “Where did I leave my badge?” More details ➡️ https://t.co/vBFG7Z58kQ
39
253
1K
@aravindr93
Aravind Rajeswaran
2 years
🚀 Exciting news! The Cortex team at #FAIR is looking for #PostDocs to join our research journey! Our mission is to create foundation models for all manner of Embodied AI agents - think robots 🤖, smart glasses 🕶️, and beyond! 🤝 Our commitment? Open-source innovation! Just like
2
12
45
@aravindr93
Aravind Rajeswaran
2 years
Sad to see the saga unfold at @OpenAI, I was one of the early interns, and still have several friends and mentors there. I'll resist the urge to make elevator pitches for different labs at this stage. Folks, let the poor souls catch a break! Crisis calls for judgement and
0
0
6
@aravindr93
Aravind Rajeswaran
2 years
🚀 Excited to announce the open-sourcing of #𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲! It brings together numerous strands of work from talented collaborators over the years! ✅ Dexterous hands ✅ Quadrupeds on randomized surfaces ✅ Musculoskeletal models ✅ Visually rich domains ✅ Large
@Vikashplus
Vikash Kumar
2 years
📢#𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲 - a unified robot learning framework ✅Designed for genralizn first robot-learning era ✅Diverse (500 envs, 8 domain) ✅Single flag for Sim<>Real ✅TeleOper Support ✅Multi-(Skill x Task) realworld dataset ✅pip install robohive https://t.co/CWSipPv8M7 🧵👇
0
5
14
@Vikashplus
Vikash Kumar
2 years
📢#𝗥𝗼𝗯𝗼𝗛𝗶𝘃𝗲 - a unified robot learning framework ✅Designed for genralizn first robot-learning era ✅Diverse (500 envs, 8 domain) ✅Single flag for Sim<>Real ✅TeleOper Support ✅Multi-(Skill x Task) realworld dataset ✅pip install robohive https://t.co/CWSipPv8M7 🧵👇
3
32
193
@aravindr93
Aravind Rajeswaran
2 years
🤖 Excited to share MoDem-v2, our latest work on decoder-free (JEPA-style) world models. 💡 Provides a simple three-step recipe to extend world model-style algorithms (e.g., TD-MPC) to real-world robots. 🙌 Fun collaboration with @palanc_, @ncklashansen, and @Vikashplus. Thanks
@palanc_
Patrick Lancaster
2 years
Check out MoDem-V2 - our latest work on Visuo-Motor World Models for Real-World Robot Manipulation! 🧵🔽 Website: https://t.co/Vkx4nwsjZR Paper: https://t.co/gtEATtfepw
1
4
22
@Vikashplus
Vikash Kumar
2 years
#𝗥𝗼𝗯𝗼𝗔𝗴𝗲𝗻𝘁 -- A universal multi-task agent on a data-budget 💪 with 12 non-trivial skills 💪 can generalize them across 38 tasks 💪& 100s of novel scenarios! 🌐 https://t.co/37YbJP05Zt w/ @mangahomanga @jdvakil @m0hitsharma, Abhinav Gupta, @shubhtuls
6
67
252
@aravindr93
Aravind Rajeswaran
2 years
LfS revisited for control: ✅ A well-tuned LfS baseline is competitive with pre-trained visual rep. (PVR) ✅ But fine-tuning with aug improves PVR results ✅ And better benchmarks are needed! 📢 Tue 2-3.30pm HST, Hall-1 #711 https://t.co/5yv1UjuldD
@ncklashansen
Nicklas Hansen
2 years
I'll be presenting our work on pretrained representations for control vs. learning from scratch at #ICML2023🏝️ next week! Reach out or stop by poster 711 on Tue (session 2) for a chat about the future of RL! https://t.co/qNRf4YwjIE Details below 👇 1/n
0
0
1
@aravindr93
Aravind Rajeswaran
2 years
🌟 MTM is a new SSL paradigm for decision making 🚀 Builds on decision transformer, BERT, and MAE 🔥 A single MTM transformer is versatile and can function as a world model, policy, representation, and more! 🔍🧵 📢 Thu 1.30-3pm HST, Hall-1 #402 https://t.co/Au200XqpcX
@philippswu
Philipp Wu
3 years
Introducing Masked Trajectory Modeling (MTM), a new general-purpose framework for sequential decision making. A single transformer trained with MTM can exhibit multiple capabilities by simply choosing different masking patterns at inference time. Accepted at ICML 2023. 🧵👇
1
1
4
@aravindr93
Aravind Rajeswaran
2 years
✈️ Just landed in Hawaii 🌴 to present two cool projects at #ICML2023 🚀 Masked Trajectory Models (w/ @philippswu, @arjunmajum, @kevinleestone, @yixin_lin_ , @IMordatch, @pabbeel) 📚 LfS Revisited (w/ @ncklashansen, @haosu_twitr, @HarryXu12, @xiaolonw et al.) Details in 🧵👇
3
2
25
@JasonMa2020
Jason Ma
2 years
Excited to share our #ICML2023 paper ✨LIV✨! Extending VIP, LIV is at once a pre-training, fine-tuning, and (zero-shot!) multi-modal reward method for (real-world!) language-conditioned robotic control. Project: https://t.co/HZuZnKOqPL Code & Model: https://t.co/J30MOankDW 🧵:
2
45
161