XingyouSong Profile Banner
Richard Song Profile
Richard Song

@XingyouSong

Followers
2K
Following
488
Media
87
Statuses
270

Research Scientist @GoogleDeepmind working on Gemini thinking and AutoML. Ex: @OpenAI, @citsecurities, @MSFTResearch.

Joined February 2018
Don't wanna be here? Send us removal request.
@XingyouSong
Richard Song
3 days
Seeing text-to-text regression work for Google’s massive compute cluster (billion $$ problem!) was the final result to convince us we can reward model literally any world feedback. Paper: Code: Just train a simple encoder-decoder
Tweet media one
11
69
496
@XingyouSong
Richard Song
3 days
Thanks so much for the repost, @_akhaliq!!.
@_akhaliq
AK
3 days
Google presents Performance Prediction for Large Systems via Text-to-Text Regression
Tweet media one
0
4
50
@XingyouSong
Richard Song
2 months
Thank you both @shaohua0116 (National Taiwan University) and @MediaTek (MediaTek Advanced Research Center) for hosting my talk on LLM reward modeling and regression and applications in e.g. chip design!. Slides here:
Tweet media one
Tweet media two
1
1
19
@XingyouSong
Richard Song
2 months
RT @yidingjiang: @MinqiJiang I think it also shows how bad the existing exploration methods are. RL works now because pretraining did the h….
0
3
0
@XingyouSong
Richard Song
2 months
I’ve done way better jobs as an AC than what we got for ICML…. esp. in borderline cases, I personally spend hours reading the papers and reviews in detail, overriding mediocre or wrong reviews if necessary. Papers are the lifeline of early-stage scientists, and I wouldn’t dare.
@roydanroy
Dan Roy
2 months
I’ve found the laziest AC or maybe they found me. Look people: if you get invited to be an AC at a top conference like… I don’t know, ICML, and then go M.I.A. during the entire process until you have ChatGPT write your decision summary, and then you don’t even bother to check.
0
1
32
@XingyouSong
Richard Song
2 months
Equator Red 🇸🇬 @iclr_conf
Tweet media one
0
0
8
@XingyouSong
Richard Song
2 months
Exactly why we spent last year on LLM regressors -> reward models to simulate expensive world feedback, not just humans. "To discover new ideas that go far beyond existing human knowledge, it is instead necessary to use grounded rewards: signals that arise from the environment
Tweet media one
@RichardSSutton
Richard Sutton
3 months
David Silver really hits it out of the park in this podcast. The paper "Welcome to the Era of Experience" is here:
2
46
315
@XingyouSong
Richard Song
2 months
Will be in Singapore for ICLR next week if you want to catch up!
Tweet media one
0
0
10
@XingyouSong
Richard Song
3 months
RT @ankesh_anand: MathArena results for gemini-2.5-pro
Tweet media one
0
52
0
@XingyouSong
Richard Song
3 months
AIME is nearly completely solved (92% on AIME 2024) 🤯 Congratulations to everybody! Very happy to be working in Gemini Thinking and Reasoning - it's been really fun so far.
@sundarpichai
Sundar Pichai
3 months
1/ Gemini 2.5 is here, and it’s our most intelligent AI model ever. Our first 2.5 model, Gemini 2.5 Pro Experimental is a state-of-the-art thinking model, leading in a wide range of benchmarks – with impressive improvements in enhanced reasoning and coding and now #1 on
1
0
7
@XingyouSong
Richard Song
4 months
Congratulations to Gemini Thinking! 🥳.
@jack_w_rae
Jack Rae
4 months
Google Deep Research is noticeably improved today! Cool to see people's experience so far. Why is it better? A bunch of product development from the team, and the underlying model updating from 1.5 Pro --> 2.0 Flash Thinking.
0
1
82
@XingyouSong
Richard Song
4 months
🔥🔥🔥🔥🔥.
@suchenzang
Susan Zhang
4 months
overall, this is decent work for a single intern's project, but i find the claims to be too grand for what is actually presented. publication pressure seems to force writing a 35 page paper on a new codebase, and the substance of the codebase could've been an add-on to SWE-Bench
Tweet media one
Tweet media two
0
0
4
@XingyouSong
Richard Song
4 months
These are exactly the type of superhuman AI problems which can benefit from the AutoML community's expertise :).
@DrJimFan
Jim Fan
4 months
The coolest autonomous coding agent I've seen recently: use AI to write better CUDA kernels to accelerate AI. AutoML is so back! The highest leverage thing you can do with your compute resources is to increase the future productivity of the same compute. It aligns all the stars
Tweet media one
0
0
5
@XingyouSong
Richard Song
5 months
In conclusion: We hope this paper can be a valuable starting point and reference for understanding next-token-prediction over numbers. What originally started out as an investigation, turned out to be very deep with connections to error-correcting codes, risk minimization
Tweet media one
0
0
3
@XingyouSong
Richard Song
5 months
It's an open question of what are better tokenizations of numbers. We tried basic error-correcting codes (e.g. majority voting over repeats) and found they can reduce the error. We could consider e.g. 𝑝-adic expansions, or even the Stern–Brocot tree as well.
Tweet media one
Tweet media two
1
0
1