Vivek Myers Profile
Vivek Myers

@vivek_myers

Followers
1K
Following
3K
Media
32
Statuses
147

PhD student @Berkeley_AI | reinforcement learning | 🦋 @ https://t.co/KcNnhdfr5m

Joined December 2019
Don't wanna be here? Send us removal request.
@vivek_myers
Vivek Myers
7 days
RT @esfrankel: Tomorrow, I'm excited to present "Finite-Time Convergence Rates in Stochastic Stackelberg Games with Smooth Algorithmic Agen….
0
2
0
@vivek_myers
Vivek Myers
7 days
RT @ajwagenmaker: How can we train a foundation model to internalize what it means to “explore”?. Come check out our work on “behavioral ex….
0
50
0
@vivek_myers
Vivek Myers
12 days
RT @qiyang_li: Everyone knows action chunking is great for imitation learning. It turns out that we can extend its success to RL to better….
0
65
0
@vivek_myers
Vivek Myers
28 days
RT @ajwagenmaker: Diffusion policies have demonstrated impressive performance in robot control, yet are difficult to improve online when 0-….
0
59
0
@vivek_myers
Vivek Myers
1 month
RT @siddkaramcheti: Thrilled to share that I'll be starting as an Assistant Professor at Georgia Tech (@ICatGT / @GTrobotics / @mlatgt) in….
0
27
0
@vivek_myers
Vivek Myers
1 month
RT @seohong_park: Q-learning is not yet scalable. I wrote a blog post about my thoughts on scalable RL algorithms.….
0
188
0
@vivek_myers
Vivek Myers
1 month
RT @chongyiz1: 1/ How should RL agents prepare to solve new tasks? While prior methods often learn a model that predicts the immediate next….
0
33
0
@vivek_myers
Vivek Myers
1 month
RT @seohong_park: New paper on unsupervised pre-training for RL!. The idea is to learn a flow-based future prediction model for each "inten….
0
20
0
@vivek_myers
Vivek Myers
2 months
RT @YifeiZhou02: 📢 New Preprint: Self-Challenging Agent (SCA) 📢. It’s costly to scale agent tasks with reliable verifiers. In SCA, the key….
0
37
0
@vivek_myers
Vivek Myers
2 months
RT @verityw_: Embodied chain-of-thought reasoning (ECoT) is a powerful way to improve robot generalization & performance. But why is this t….
0
11
0
@vivek_myers
Vivek Myers
3 months
How can we learn the reward functions of multiple agents from unlabeled data? We model interactions between learned models of the agent objectives (marginalized 𝑄 functions) when performing inverse RL to learn mixed cooperative/competitive environments. See @ebiyik_'s thread ↓.
@ebiyik_
Erdem Bıyık
3 months
In another ICRA 2025 paper, we developed a multi-agent inverse reinforcement learning method. As opposed to existing solutions, our method does not make assumptions about the reward structure -- it can be an environment that is neither fully cooperative nor fully competitive.
Tweet media one
0
0
11
@vivek_myers
Vivek Myers
3 months
RT @amyxlu: Gave my PhD dissertation talk! 🧬 Tried my best to make AI for drug discovery & protein design accessible for ML folks:. Beyond….
0
72
0
@vivek_myers
Vivek Myers
3 months
RT @ebiyik_: In another ICRA 2025 paper, we developed a multi-agent inverse reinforcement learning method. As opposed to existing solutions….
0
2
0
@vivek_myers
Vivek Myers
3 months
Presenting w/ Cathy Ji and @ben_eysenbach . Thread:
@vivek_myers
Vivek Myers
6 months
Reinforcement learning should be able to improve upon behaviors seen when training. In practice, RL agents often struggle to generalize to new long-horizon behaviors. Our new paper studies *horizon generalization*, the degree RL algorithms generalize to reaching distant goals. 1/
0
0
0
@vivek_myers
Vivek Myers
3 months
How can agents trained to reach (temporally) nearby goals generalize to attain distant goals?. Come to our #ICLR2025 poster tomorrow to discuss 𝘩𝘰𝘳𝘪𝘻𝘰𝘯 𝘨𝘦𝘯𝘦𝘳𝘢𝘭𝘪𝘻𝘢𝘵𝘪𝘰𝘯!. 🗓️Saturday, April 26, 2025.🕙10:00 – 12:30 AM.📍Hall 3 + Hall 2B #637
Tweet media one
1
1
8
@vivek_myers
Vivek Myers
3 months
RT @m_bortkiewicz: Excited to present JaxGCRL at ICLR 2025 (spotlight):. 📍Hall 3 + Hall 2B, Poster #422.🗓️Friday, April 25.🕒3:00 PM – 5:00….
0
3
0
@vivek_myers
Vivek Myers
3 months
GitHub: See @m_bortkiewicz's thread for details:.
Tweet card summary image
github.com
Online Goal-Conditioned Reinforcement Learning in JAX. ICLR 2025 Spotlight. - MichalBortkiewicz/JaxGCRL
@m_bortkiewicz
Michał Bortkiewicz
10 months
I am excited to share our recent work with @WladekPalucki , @vivek_myers, @Taddziarm , @tomArczewski, @LukeKucinski, and @ben_eysenbach!. Accelerating Goal-Conditioned Reinforcement Learning Algorithms and Research . Webpage:
0
0
2
@vivek_myers
Vivek Myers
3 months
Working on goal-conditioned RL? 𝙹𝚊𝚡𝙶𝙲𝚁𝙻 offers 10+ GPU-accelerated environments (up to 22× faster) with strong GCRL baselines—ideal for rapid iteration and benchmarking. Now on PyPI!.𝚙𝚒𝚙 𝚒𝚗𝚜𝚝𝚊𝚕𝚕 𝚓𝚊𝚡𝚐𝚌𝚛𝚕
Tweet media one
Tweet media two
1
0
17
@vivek_myers
Vivek Myers
3 months
RT @cassidy_laidlaw: We built an AI assistant that plays Minecraft with you. Start building a house—it figures out what you’re doing and ju….
0
217
0
@vivek_myers
Vivek Myers
4 months
RT @a_lidayan: 🚨Our new #ICLR2025 paper presents a unified framework for intrinsic motivation and reward shaping: they signal the value of….
0
31
0