Yonathan Efroni
@EfroniYonathan
Followers
896
Following
2K
Media
3
Statuses
129
Assistant Professor@TAU | AA-I Technologies
Tel Aviv
Joined October 2020
AI agents thrive on context. But with too much context, they go off the rails. This is why we’re going to see subagents for particular tasks or roles in a workflow. And this also means there’s a ton of opportunity for building these deep, domain specific agents.
The AGI mental model once was a single monolithic AI system that could do all your tasks. However, the future likely looks like many specialist subagents with deep expertise, orchestrated together – @Levie, @stevesi, @martin_casado & @eriktorenberg discuss on the @a16z podcast
40
59
334
In era of pretraining, what mattered was internet text. You'd primarily want a large, diverse, high quality collection of internet documents to learn from. In era of supervised finetuning, it was conversations. Contract workers are hired to create answers for questions, a bit
Introducing the Environments Hub RL environments are the key bottleneck to the next wave of AI progress, but big labs are locking them down We built a community platform for crowdsourcing open environments, so anyone can contribute to open-source AGI
264
882
7K
*very* excited to share a new *efficient* method for learning *marginally stable* and NONLINEAR dynamical systems, w. brilliant students Evan Dogariu and Anand Brahmbhatt @AnandBrahm15501: https://t.co/gwVAaPVVst more info in thread
arxiv.org
We study the fundamental problem of learning a marginally stable unknown nonlinear dynamical system. We describe an algorithm for this problem, based on the technique of spectral filtering, which...
7
28
207
Are frontier AI models really capable of “PhD-level” reasoning? To answer this question, we introduce FormulaOne, a new reasoning benchmark of expert-level Dynamic Programming problems. We have curated a benchmark consisting of three tiers, in increasing complexity, which we call
102
404
4K
It seems GPT‑OSS is very prone to hallucinations … check out our RLCR paper to see how we trained reasoning models to know what they don't know. Website 🌐 and code 💻 out today! https://t.co/YqLu92enIy 🚀
5
58
404
AI Research Agents for ML Achieves state-of-the-art on MLE-bench lite! Using AI to automate the training of ML models is one of the most exciting and promising areas of research today. Lots of cool ideas in this paper:
11
130
609
@seohong_park Yeah! We wondered the same in casual inference / offline RL. We found earliest disagreement times, a type of adaptive time scale, are a useful concept for continuous or finely discretized times. Could benefit from some large scale experiments, lots to do! https://t.co/uWC2wwG8Sx
openreview.net
Problems in fields such as healthcare, robotics, and finance requires reasoning about the value both of what decision or action to take and when to take it. The prevailing hope is that artificial...
0
1
13
Key to research success: ambition in vision, but pragmatism in execution. You must be guided by a long-term, ambitious goal that addresses a fundamental problem, rather than chasing incremental gains on established benchmarks. Yet, your progress should be grounded by tractable
48
241
2K
It was a dream come true to teach the course I wish existed at the start of my PhD. We built up the algorithmic foundations of modern-day RL, imitation learning, and RLHF, going deeper than the usual "grab bag of tricks". All 25 lectures + 150 pages of notes are now public! 🧵
8
92
706
I now have a draft of my introduction to CTDE in (cooperative) MARL. It is meant to introduce new graduate students (that already know a bit about RL) to the area. Check it out and let me know your thoughts! https://t.co/6PyE9eMXlL
arxiv.org
Multi-agent reinforcement learning (MARL) has exploded in popularity in recent years. Many approaches have been developed but they can be divided into three main types: centralized training and...
1
7
52
Modern scaling law research often feels like this: 1. Train a few models 2. Plot metrics on a log-log scale 3. Fit a line 4. Call it a new law Maybe it’s time to ask: are we uncovering principles, or just describing artifacts?🤔
7
8
135
The most rewarding aspect of scientific adventure is why we do it ? How we get there and what we achieve through understanding the governing laws of the world is less important.
0
1
4
Tomorrow it is time again for a great seminar! Join us to hear out one of Jeongyeol's latest findings.
0
4
8
@chrodan @YishayMansour @MehryarMohri tl;dr: a fun project that required us to rethink of a new framework w/ Ben Kretzu, @danielrjiang, @bhandari_jalaj, @ZheqingZhu and @karen_ullrich
0
0
3
we actually started by asking this question in the multi-armed / tabular RL, and after spending some time on it realized it has been explored already by @chrodan, @YishayMansour, @MehryarMohri:
proceedings.mlr.press
Reward design is one of the most critical and challenging aspects when formulating a task as a reinforcement learning (RL) problem. In practice, it often tak...
1
1
5
accepted to #ICML25🍁 we asked 🤔 how can we improve gradient-descent in the presence of multiple aligned or similar objectives?🤔 this becomes increasingly important when having access to multiple reward functions / datasets / tasks
Aligned Multi-Objective Optimization (A-🐮) has been accepted at #ICML2025! 🎉 We explore optimization scenarios where objectives align rather than conflict, introducing new scalable algorithms with theoretical guarantees. #MachineLearning #AIResearch #Optimization #MLCommunity
1
6
36
Meet us (but not me sadly) at the poster session: https://t.co/sJdUpeNleu
#ICLR2025 (Also, much more interesting things to explore in MARL and offline MARL imo)
💫Accepted to ICLR25! 💫 We investigate a special MARL structure in which agents weakly interact. This, we show, makes MARL much more tractable. Led by @zhan_wenhao in his summer internship + it was a delight working on this, and expect to see cool extensions ahead!
1
5
32
Hiring researchers and engineers for a stealth, applied research company with a focus on RL x foundation models. Folks on the team already are leading RL / learning researchers. If you think you'd be good at the research needed to get things working in practice, email me
10
34
518
There are multiple postdoc positions available as part of an exciting new AI-agent initiative at Columbia that tackles challenges at the frontier of agentic systems and sequential decision-making. I am not very active here so please help me spread the word!
1
19
56