EhsanImanii Profile Banner
Ehsan Imani Profile
Ehsan Imani

@EhsanImanii

Followers
1K
Following
4K
Media
3
Statuses
204

RecSys/Post-Training @Spotify | RLAI grad

Toronto, Ontario
Joined November 2018
Don't wanna be here? Send us removal request.
@akhil_bagaria
Akhil Bagaria
25 days
My RL team at @amazon NYC is looking for summer 2026 PhD interns! We apply RL to Amazon's supply chain and do publishable, open-ended research in meta learning, multi-agent RL, constrained RL, exploration, and LLMs+RL. Interested? Email your CV to akhilbg@amazon.com by Nov 30th!
4
27
259
@xai
xAI
1 month
Introducing Grok 4.1, a frontier model that sets a new standard for conversational intelligence, emotional understanding, and real-world helpfulness. Grok 4.1 is available for free on https://t.co/AnXpIEOPEb, https://t.co/53pltyq3a4 and our mobile apps. https://t.co/Cdmv5CqSrb
Tweet card summary image
x.ai
Grok 4.1 is now available to all users on grok.com, 𝕏, and the iOS and Android apps. It is rolling out immediately in Auto mode and can be selected explicitly as “Grok 4.1” in the model picker.
2K
2K
13K
@Mononofu
Julian Schrittwieser
1 month
Very excited that our AlphaProof paper is finally out! It's the final thing I worked on at DeepMind, very satisfying to be able to share the full details now - very fun project and awesome team! https://t.co/OuWDemzAt4
19
101
1K
@korymath
Kory Mathewson
1 month
I'm looking for a student researcher to join me at @GoogleDeepMind Montreal to work on getting the best generative AI into the hands of the best creative people. Your work will push models past their current frontiers and directly benefit creative engagements. Are you
22
82
637
@finbarrtimbers
finbarr
1 month
Come work with me on OLMo! https://t.co/YiWhZtrcvz
job-boards.greenhouse.io
Seattle, WA
@natolambert
Nathan Lambert
1 month
We're starting to hire for our 2026 Olmo interns! Looking for excellent students to do research to help build our best models (primarily enrolled in Ph.D. with experience or interest in any area of the language modeling pipeline).
1
4
98
@luisenp
Luis Pineda
2 months
After 7 years at FAIR, I've been affected by the recent AI layoffs. If you are interested in robotics learning, let's chat :)
61
54
1K
@EthanHe_42
Ethan He
2 months
Recommendation system with AI will be a complete game changer. This is nothing like traditional algorithms that recommend based on heuristics. @grok will read every post, understand you even better than yourself, and recommend contents you like. If your content is genuinely
132
62
912
@Devvrit_Khatri
Devvrit
2 months
Wish to build scaling laws for RL but not sure how to scale? Or what scales? Or would RL even scale predictably? We introduce: The Art of Scaling Reinforcement Learning Compute for LLMs
9
103
553
@InfiniAILab
Infini-AI-Lab
2 months
🤔Can we train RL on LLMs with extremely stale data? 🚀Our latest study says YES! Stale data can be as informative as on-policy data, unlocking more scalable, efficient asynchronous RL for LLMs. We introduce M2PO, an off-policy RL algorithm that keeps training stable and
4
41
228
@elonmusk
Elon Musk
2 months
To the degree that people are seeing improvements in their feed, it is not due to the actions of specific individuals changing heuristics, but rather increasing use of Grok and other AI tools. We will post the updated algorithm, including model weights, later this week. You
@XFreeze
X Freeze
2 months
The improvement in the X feed over the past year is incredible X team has finally got to to the bottom and wiped spam & engagement bait now or extremely rare Can’t even recall the last time a low-quality post showed up Huge credit to the X team.... these changes are truly
48K
10K
87K
@RichardSSutton
Richard Sutton
2 months
More on LLMs, RL, and the bitter lesson, on the Derby Mill podcast.
@professor_ajay
Ajay Agrawal
2 months
7
18
238
@SFResearch
Salesforce AI Research
2 months
📣 Webscale-RL: Automated Data Pipeline for Scaling RL Data to Pretraining Levels 📣 RL for LLMs faces a critical data bottleneck: existing RL datasets are <10B tokens while pretraining uses >1T tokens. Our Webscale-RL pipeline solves this by automatically converting pretraining
1
7
36
@pouria_mahdavi
Pouria Mahdavinia
3 months
(1/n) Neural networks trained with Adam record a map of their learning journey—the optimizer's second moments (exp_avg_sq). These approximate the Hessian (curvature) at training's end. We used this insight to uncover why model merging works & built stronger merging methods 🧵
3
18
59
@apsarathchandar
Sarath Chandar
3 months
At @ChandarLab, we are happy to announce the third edition of our assistance program to provide feedback for members of communities underrepresented in AI who want to apply to high-profile graduate programs. Want feedback? Details: https://t.co/0K6AA4gV0U. Deadline: Nov 01! cc:
2
28
86
@kchonyc
Kyunghyun Cho
3 months
when you give up on this nebulous idea and illusion of prestige, you will finally find peace and freedom. submit to TMLR and JMLR.
10
30
549
@shayan_
Shayan
3 months
We just released a model that’s both fast and smart, beating all models of the same size by a huge margin. It’s available for free. Give it a try!
@xai
xAI
3 months
Introducing Grok 4 Fast, a multimodal reasoning model with a 2M context window that sets a new standard for cost-efficient intelligence. Available for free on https://t.co/AnXpIEOhOD, https://t.co/53pltypvkw, iOS and Android apps, and OpenRouter. https://t.co/3YZ1yVwueV
35
11
366
@MLciosek
Kamil Ciosek
3 months
For anyone worried their LLM might be making stuff up, we made a budget‐friendly truth serum (semantic entropy + Bayesian). See for yourself: https://t.co/gq8oFP5Eqr Paper:
0
7
4
@arshtvk
Arash Tavakoli
4 months
We're hiring Research Scientists in Tech Research @RiotGames! If you're passionate about shaping the future of games with AI and have research experience in RL, IL, or generative methods, we'd love to hear from you. DM me if interested.
1
3
20