Qu3ntinB Profile Banner
Quentin Bertrand Profile
Quentin Bertrand

@Qu3ntinB

Followers
1K
Following
999
Media
16
Statuses
1K

Researcher at @Inria, affiliated at @Mila_Quebec. Previously, postdoctoral researcher at @Mila_Quebec w/ @SimonLacosteJ and @gauthier_gidel.

Joined January 2021
Don't wanna be here? Send us removal request.
@Yoshua_Bengio
Yoshua Bengio
8 hours
We need innovative technical and societal solutions to mitigate AI risks. I believe liability insurance for AI developers could be an excellent market-based incentive to drive safety standards and accountability, and is an option worth considering. https://t.co/SXD1pRSSz1
Tweet card summary image
ft.com
Turing Prize winner urges governments to require tech groups to cover catastrophic outcomes and fund safety research
5
5
30
@StefanoErmon
Stefano Ermon
10 hours
We just shipped a major Mercury refresh. ⚡ Best-in-class quality at up to 10× lower latency. Still the only commercial diffusion LLM in the world. Try the new model.
@_inception_ai
Inception
10 hours
Mercury is refreshed – with across-the-board improvements in coding, instruction following, math, and knowledge recall. Start building responsive, in-the-flow AI solutions! Read more: https://t.co/QyTVaHAIue
2
8
64
@deedydas
Deedy
10 hours
This Stanford professor just raised a $50M Seed and have built a 10x faster and 10x cheaper AI coding model with the performance of Gemini Flash / Haiku. Inception Labs’ Mercury model can implement games like Connect 4 from scratch in ~2s. The speed feels magical, like going
17
28
307
@Mila_Quebec
Mila - Institut québécois d'IA
2 days
Being part of Mila means joining the world’s largest community of academic researchers in deep learning. Submit your supervision request now for the MSc or PhD for Fall 2026. https://t.co/r01eLcXtZw
0
4
8
@SuryaGanguli
Surya Ganguli
1 day
Very grateful to @schmidtsciences for being awarded an #AI2050 senior fellowship. And honored to be part of this select 2025 cohort of 7 senior fellows. This award will support our work on a deeper scientific basis for understanding and improving how artificial intelligence
@schmidtsciences
Schmidt Sciences
1 day
We're excited to welcome 28 new AI2050 Fellows! This 4th cohort of researchers are pursuing projects that include building AI scientists, designing trustworthy models, and improving biological and medical research, among other areas. https://t.co/8oY7xdhxvF
8
4
60
@weijie444
Weijie Su
2 days
Why and how does gradient/matrix orthogonalization work in Muon for training #LLMs? We introduce an isotropic curvature model to explain it. Take-aways: 1. Orthogonalization is a good idea, "on the right track". 2. But it might not be optimal. [1/n]
3
14
127
@bose_joey
Joey Bose
2 days
Come do a PhD with me 😀! Promise of fun science and great coffee ☕
@giladturok
Gilad
3 days
I like the way @joeybos lays out his vision for PhD supervision! Seems intense and rewarding.
30
70
731
@mathusmassias
Mathurin Massias
2 days
🌀New paper on the generation phases of Flow Matching https://t.co/tzG2kPVGsE Are FM & diffusion models nothing else than denoisers trained at every noise level? In theory yes, *if trained optimally*. But in practice, do all noise level matter equally?
6
101
643
@sivareddyg
Siva Reddy
3 days
🚨🚨 The common myth that value alignment happens at the preference optimization (RLHF) stage is incorrect and have mislead years of research 💣. Mehar did a meticulous job showing that LLMs acquire values during SFT, not during preference optimization. Your SFT is probably the
@bhatia_mehar
Mehar Bhatia
3 days
🚨How do LLMs acquire human values?🤔 We often point to preference optimization. However, in our new work, we trace how and when model values shift during post-training and uncover surprising dynamics. We ask: How do data, algorithms, and their interaction shape model values?🧵
2
3
20
@yoavartzi
Yoav Artzi
3 days
Pushed a big update to LM-class (v2025.2) -- this second version makes a much more mature resource Many refinements of lecture slides + significant improvements to the assignments Many thanks to @ch272h @HuaYilun and @shankarpad8 for their work on the assignments
1
5
21
@MistralAI
Mistral AI
3 days
Full stack devs, SWEs, MLEs, forward deployed engineers, research engineers, applied scientists: we are hiring! Join us and tackle cutting-edge challenges including physical AI, time series, material sciences, cybersecurity and many more. Positions available in Paris, London,
Tweet card summary image
jobs.lever.co
Job openings at Mistral AI
31
92
1K
@sedielem
Sander Dieleman
6 days
I contemplated whether I should post this, because it seems kind of obvious. But it's often taken for granted, so we might underestimate the impact: e.g. these days, diffusion papers don't usually show samples without guidance anymore (figures from GLIDE https://t.co/2wOdFfRHCK)
@sedielem
Sander Dieleman
6 days
Generative modelling used to be about capturing the training data distribution. Interestingly, this stopped being the case when we started actually using them🤔 We tweak temps, use classifier-free guidance and post-train to get a distribution better than the training data.
3
15
153
@mariannearr
Marianne Arriola
9 days
🚨In our NeurIPS paper, we bring encoder-decoders back.. for diffusion language models! ⚡️Encoder-decoders make diffusion sampling fast: a small (fast) decoder denoises tokens progressively and a large (slower) encoder represents clean context.
8
36
242
@jm_alexia
Alexia Jolicoeur-Martineau
7 days
Recordings for the talk on "Tiny Recursive Models" are up. https://t.co/6iCRHcX4VL
@jm_alexia
Alexia Jolicoeur-Martineau
10 days
I will give a presentation on "Tiny Recursion Models" tomorrow at 1pm in the Mila Agora (6650 Saint-Urbain, Montreal). Its open to everyone, feel free to come by!
14
86
651
@osclsd
Oscar Davis
8 days
Introducing Generalised Flow Maps 🎉 A stable, few-step generative model on Riemannian manifolds 🪩 📚 Read it at: https://t.co/iCTHedwCxf 💾 Code: https://t.co/MeukcthFN2 @msalbergo @nmboffi @mmbronstein @bose_joey
3
22
112
@ErnestRyu
Ernest Ryu
9 days
Preprint on using ChatGPT to resolve a 42-year-old open problem (point convergence of Nesterov’s accelerated gradient method) is out. Mathematical results are complete, though still need to expand the discussion of historical context & prior work. (1/2) https://t.co/Dmd9huMjXS
Tweet card summary image
arxiv.org
The Nesterov accelerated gradient method, introduced in 1983, has been a cornerstone of optimization theory and practice. Yet the question of its point convergence had remained open. In this work,...
12
68
469
@yoavartzi
Yoav Artzi
9 days
I am potentially recruiting a postdoctoral fellow through this program. If interested, name me as a mentor, and ping me to let me know that you are applying! The process includes some sort of interview, so I can try to squeeze a few of these in advance (it will help a lot!)
@yoavartzi
Yoav Artzi
9 days
.@Cornell is recruiting for multiple postdoctoral positions in AI as part of two programs: Empire AI Fellows and Foundational AI Fellows. Positions are available in NYC and Ithaca. Deadline for full consideration is Nov 20, 2025! https://t.co/Cp5710BauU
1
7
18
@giannis_daras
Giannis Daras
9 days
Deep Image Prior is back
@kwangmoo_yi
Kwang Moo Yi
10 days
Luo et al., "Self-diffusion for Solving Inverse Problems" Pretty much a deep image prior for denoising models. Without ANY data, with a single image, you can train a denoiser via diffusion training, and it just magically learns to solve inverse problems.
1
6
55
@jm_alexia
Alexia Jolicoeur-Martineau
10 days
I will post the recordings after, don't worry.
9
2
116
@jm_alexia
Alexia Jolicoeur-Martineau
10 days
I will give a presentation on "Tiny Recursion Models" tomorrow at 1pm in the Mila Agora (6650 Saint-Urbain, Montreal). Its open to everyone, feel free to come by!
20
18
279