Jonathan Lai Profile
Jonathan Lai

@_JLai

Followers
518
Following
2K
Media
1
Statuses
26

Post training @GoogleDeepMind, Gemini Reasoning, training algorithms, RL, opinions are my own

Joined November 2012
Don't wanna be here? Send us removal request.
@YiTayML
Yi Tay
3 months
Excited to share that I'll be hosting some of the world's best AI researchers and engineers for our @GoogleDeepMind Gemini event next week in Singapore 🇸🇬! Join @JeffDean, @quocleix, @benoitschilling, @melvinjohnsonp and @denny_zhou for a day of technical conversations, panels
20
39
336
@_JLai
Jonathan Lai
5 months
Huge congrats @prateeky2806 and all!! 🎉 It was great to work with everyone here!
@tuvllms
Tu Vu
5 months
Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern @prateeky2806 and our awesome co-authors @_JLai, @alexandraxron, @manaalfar, @mohitban47, and @TsendeeMTS 🎉!!
0
0
5
@tuvllms
Tu Vu
5 months
Excited to share that our paper on model merging at scale has been accepted to Transactions on Machine Learning Research (TMLR). Huge congrats to my intern @prateeky2806 and our awesome co-authors @_JLai, @alexandraxron, @manaalfar, @mohitban47, and @TsendeeMTS 🎉!!
@prateeky2806
Prateek Yadav
1 year
Ever wondered if model merging works at scale? Maybe the benefits wear off for bigger models? Maybe you considered using model merging for post-training of your large model but not sure if it generalizes well? cc: @GoogleAI @GoogleDeepMind @uncnlp 🧵👇 Excited to announce my
2
21
90
@tuvllms
Tu Vu
8 months
🚨 New paper 🚨 Excited to share my first paper w/ my PhD students!! We find that advanced LLM capabilities conferred by instruction or alignment tuning (e.g., SFT, RLHF, DPO, GRPO) can be encoded into model diff vectors (à la task vectors) and transferred across model
14
94
445
@ankesh_anand
Ankesh Anand
8 months
shoutout to the believers!
43
70
2K
@TsendeeMTS
Tsendsuren
8 months
Gemini-ийн шинэ загварыг туршаад үзээрэй. Код бичих дээр нилээн сайжирсан байгаа.
@GoogleDeepMind
Google DeepMind
8 months
Think you know Gemini? 🤔 Think again. Meet Gemini 2.5: our most intelligent model 💡 The first release is Pro Experimental, which is state-of-the-art across many benchmarks - meaning it can handle complex problems and give more accurate responses. Try it now →
0
8
18
@_JLai
Jonathan Lai
8 months
A historic elo margin on LMSYS and also crushed almost all reasoning and STEM benchmarks!! So proud of this team!!
@arena
lmarena.ai
8 months
BREAKING: Gemini 2.5 Pro is now #1 on the Arena leaderboard - the largest score jump ever (+40 pts vs Grok-3/GPT-4.5)! 🏆 Tested under codename "nebula"🌌, Gemini 2.5 Pro ranked #1🥇 across ALL categories and UNIQUELY #1 in Math, Creative Writing, Instruction Following, Longer
0
1
4
@sundarpichai
Sundar Pichai
8 months
1/ Gemini 2.5 is here, and it’s our most intelligent AI model ever. Our first 2.5 model, Gemini 2.5 Pro Experimental is a state-of-the-art thinking model, leading in a wide range of benchmarks – with impressive improvements in enhanced reasoning and coding and now #1 on
306
961
7K
@OfficialLoganK
Logan Kilpatrick
10 months
We are rolling out a new Gemini 2.0 Flash Thinking update: - Exp-01-21 variant in AI Studio and API for free - 1 million token context window - Native code execution support - Longer output token generation - Less frequent model contradictions Try it
Tweet card summary image
aistudio.google.com
The fastest path from prompt to production with Gemini
219
324
3K
@_JLai
Jonathan Lai
11 months
Try our new experimental thinking model at https://t.co/jcF922lj4j !! Appreciate any and all feedback
@JeffDean
Jeff Dean
11 months
Introducing Gemini 2.0 Flash Thinking, an experimental model that explicitly shows its thoughts. Built on 2.0 Flash’s speed and performance, this model is trained to use thoughts to strengthen its reasoning. And we see promising results when we increase inference time
2
0
20
@_JLai
Jonathan Lai
1 year
Prateek is an amazing researcher!! Definitely hire him!
@prateeky2806
Prateek Yadav
1 year
I'm on the job market! Please reach out if you are looking to hire someone to work on - RLHF - Efficiency - MoE/Modular models - Synthetic Data - Test time compute - other phases of pre/post-training. If you are not hiring then I would appreciate a retweet! More details👇
0
0
2
@prateeky2806
Prateek Yadav
1 year
I'm on the job market! Please reach out if you are looking to hire someone to work on - RLHF - Efficiency - MoE/Modular models - Synthetic Data - Test time compute - other phases of pre/post-training. If you are not hiring then I would appreciate a retweet! More details👇
8
59
235
@_JLai
Jonathan Lai
1 year
🚨✨ Thrilled to share the first study on model merging at large scales by our intern @prateeky2806 @GoogleAI @GoogleDeepMind For larger LLMs merging is an efficient alternative to multitask learning, that can preserve the majority of in-domain performance, while significantly
@prateeky2806
Prateek Yadav
1 year
Ever wondered if model merging works at scale? Maybe the benefits wear off for bigger models? Maybe you considered using model merging for post-training of your large model but not sure if it generalizes well? cc: @GoogleAI @GoogleDeepMind @uncnlp 🧵👇 Excited to announce my
0
2
21
@MLB
MLB
10 years
#BatFlip forever on loop. #OwnOctober
243
9K
9K