
Michael Eli Sander
@m_e_sander
Followers
3K
Following
412
Media
44
Statuses
213
Research Scientist at Google DeepMind
Paris
Joined February 2021
🚨🚨New ICML 2024 Paper: How do Transformers perform In-Context Autoregressive Learning?. We investigate how causal Transformers learn simple autoregressive processes or order 1. with @RGiryes, @btreetaiji, @mblondel_ml and @gabrielpeyre 🙏
2
42
154
RT @mblondel_ml: Distillation is becoming a major paradigm for training LLMs but its success and failure modes remain quite mysterious. Our….
0
11
0
RT @mblondel_ml: Really proud of these two companion papers by our team at GDM:. 1) Joint Learning of Energy-based Models and their Partiti….
0
24
0
RT @RednasTom: I am in NeurIPS week :). Friday, Presenting our spotlight work: Watermarking Makes LLMs Radioactive ☢️ (.
0
1
0
RT @sarah_perrin_: ♟️Mastering Board Games by External and.Internal Planning with Language Models♟️. I'm happy to finally share https://t.c….
0
12
0
RT @SibylleMarcotte: Merci pour l’opportunité d’avoir échangé sur mes recherches et mes expériences !.Merci à mes directeurs de thèse @gabr….
0
6
0
RT @RednasTom: ☢️ Some news about radioactivity ☢️.- We got a Spotlight at Neurips! 🥳 and we will be in Vancouver with @pierrefdz to prese….
0
2
0
RT @RednasTom: 🔒Image watermarking is promising for digital content protection. But images often undergo many modifications—spliced or alt….
0
7
0
RT @fpedregosa: Six years at Google today! 🎉 From 🇨🇦 to 🇨🇭, optimizing everything in sight. Grateful for the incredible journey and amazin….
0
1
0
RT @SibylleMarcotte: 🏆Didn't get the Physics Nobel Prize this year, but really excited to share that I've been named one of the #FWIS2024 @….
0
20
0
RT @PierreMari0n: 🚨New paper alert🚨: How does Transformer retrieve information which is sparsely concentrated in f….
0
5
0
RT @JulesSamaran: After a very constructive back and forth with editors and reviewers of @NatureComms, scConfluence has now been published….
0
10
0
RT @gabrielpeyre: "Transformers are Universal In-context Learners": in this paper, we show that deep transformers with a fixed embedding di….
0
318
0
RT @gjhuizing: 🎉 New preprint! STORIES learns a differentiation potential from spatial transcriptomics profiled at….
0
22
0
RT @jkobject: 🚨🚨 AI in Bio release 🧬 . Very happy to share my work on a Large Cell Model for Gene Network Inference. It is for now just a….
0
20
0
RT @mblondel_ml: We uploaded a v2 of our book draft "The Elements of Differentiable Programming" with many improvements (~70 pages of new c….
0
123
0
Come and see us today at 1:30 pm at spot #411 for our poster session !!
🚨🚨New ICML 2024 Paper: How do Transformers perform In-Context Autoregressive Learning?. We investigate how causal Transformers learn simple autoregressive processes or order 1. with @RGiryes, @btreetaiji, @mblondel_ml and @gabrielpeyre 🙏
0
10
45
RT @RednasTom: You didn’t believe in Differential Private training for foundation models? We achieved the same performance as non-private M….
0
5
0