m_e_sander Profile Banner
Michael Eli Sander Profile
Michael Eli Sander

@m_e_sander

Followers
3K
Following
412
Media
44
Statuses
213

Research Scientist at Google DeepMind

Paris
Joined February 2021
Don't wanna be here? Send us removal request.
@m_e_sander
Michael Eli Sander
1 year
🚨🚨New ICML 2024 Paper: How do Transformers perform In-Context Autoregressive Learning?. We investigate how causal Transformers learn simple autoregressive processes or order 1. with @RGiryes, @btreetaiji, @mblondel_ml and @gabrielpeyre 🙏
Tweet media one
2
42
154
@m_e_sander
Michael Eli Sander
5 months
RT @mblondel_ml: Distillation is becoming a major paradigm for training LLMs but its success and failure modes remain quite mysterious. Our….
0
11
0
@m_e_sander
Michael Eli Sander
5 months
RT @mblondel_ml: Really proud of these two companion papers by our team at GDM:. 1) Joint Learning of Energy-based Models and their Partiti….
0
24
0
@m_e_sander
Michael Eli Sander
7 months
RT @RednasTom: I am in NeurIPS week :). Friday, Presenting our spotlight work: Watermarking Makes LLMs Radioactive ☢️ (.
0
1
0
@m_e_sander
Michael Eli Sander
7 months
RT @sarah_perrin_: ♟️Mastering Board Games by External and.Internal Planning with Language Models♟️. I'm happy to finally share https://t.c….
0
12
0
@m_e_sander
Michael Eli Sander
8 months
RT @SibylleMarcotte: Merci pour l’opportunité d’avoir échangé sur mes recherches et mes expériences !.Merci à mes directeurs de thèse @gabr….
0
6
0
@m_e_sander
Michael Eli Sander
8 months
RT @RednasTom: ☢️ Some news about radioactivity ☢️.- We got a Spotlight at Neurips! 🥳 and we will be in Vancouver with @pierrefdz to prese….
0
2
0
@m_e_sander
Michael Eli Sander
8 months
RT @RednasTom: 🔒Image watermarking is promising for digital content protection. But images often undergo many modifications—spliced or alt….
0
7
0
@m_e_sander
Michael Eli Sander
9 months
RT @fpedregosa: Six years at Google today! 🎉 From 🇨🇦 to 🇨🇭, optimizing everything in sight. Grateful for the incredible journey and amazin….
0
1
0
@m_e_sander
Michael Eli Sander
9 months
RT @SibylleMarcotte: 🏆Didn't get the Physics Nobel Prize this year, but really excited to share that I've been named one of the #FWIS2024 @….
0
20
0
@m_e_sander
Michael Eli Sander
9 months
🥳🥳.Thrilled to share that I've joined Google DeepMind as a Research Scientist. Super excited for what's to come!
Tweet media one
118
56
4K
@m_e_sander
Michael Eli Sander
9 months
RT @PierreMari0n: 🚨New paper alert🚨: How does Transformer retrieve information which is sparsely concentrated in f….
0
5
0
@m_e_sander
Michael Eli Sander
9 months
0
21
0
@m_e_sander
Michael Eli Sander
10 months
RT @JulesSamaran: After a very constructive back and forth with editors and reviewers of @NatureComms, scConfluence has now been published….
0
10
0
@m_e_sander
Michael Eli Sander
11 months
RT @gabrielpeyre: "Transformers are Universal In-context Learners": in this paper, we show that deep transformers with a fixed embedding di….
0
318
0
@m_e_sander
Michael Eli Sander
11 months
RT @gjhuizing: 🎉 New preprint! STORIES learns a differentiation potential from spatial transcriptomics profiled at….
0
22
0
@m_e_sander
Michael Eli Sander
11 months
RT @jkobject: 🚨🚨 AI in Bio release 🧬  . Very happy to share my work on a Large Cell Model for Gene Network Inference. It is for now just a….
0
20
0
@m_e_sander
Michael Eli Sander
1 year
RT @mblondel_ml: We uploaded a v2 of our book draft "The Elements of Differentiable Programming" with many improvements (~70 pages of new c….
0
123
0
@m_e_sander
Michael Eli Sander
1 year
Come and see us today at 1:30 pm at spot #411 for our poster session !!
Tweet media one
@m_e_sander
Michael Eli Sander
1 year
🚨🚨New ICML 2024 Paper: How do Transformers perform In-Context Autoregressive Learning?. We investigate how causal Transformers learn simple autoregressive processes or order 1. with @RGiryes, @btreetaiji, @mblondel_ml and @gabrielpeyre 🙏
Tweet media one
0
10
45
@m_e_sander
Michael Eli Sander
1 year
RT @RednasTom: You didn’t believe in Differential Private training for foundation models? We achieved the same performance as non-private M….
0
5
0
@m_e_sander
Michael Eli Sander
1 year
We provide a rationale for our choice of token encoding, and experimentally investigate the generalization of our results to non-commuting context matrices. Our code is available at
Tweet media one
0
0
5