max_nlp Profile Banner
Max Bartolo Profile
Max Bartolo

@max_nlp

Followers
3K
Following
3K
Media
63
Statuses
797

I lead the Command modelling team at @Cohere and co-chair the @DynabenchAI @MLCommons working group. Prev @DeepMind, @MetaAI / FAIR & @BloomsburyAI.

Joined November 2016
Don't wanna be here? Send us removal request.
@max_nlp
Max Bartolo
8 days
RT @LauraRuis: LLMs can be programmed by backprop 🔎. In our new preprint, we show they can act as fuzzy program interpreters and databases.….
0
51
0
@max_nlp
Max Bartolo
15 days
RT @maximilianmozes: We’re looking for a Research Engineer / Scientist with a focus on Data Analysis and Evaluation to join the post-traini….
0
19
0
@max_nlp
Max Bartolo
17 days
Looking forward to sharing some of our recent research contributions at @MLStreetTalk's first London AI meetup 🤩.
@MLStreetTalk
Machine Learning Street Talk
17 days
We are running our first physical event in London on 14th July! We have Tim Nguyen @IAmTimNguyen from DeepMind and Max Bartolo @max_nlp from Cohere and Enzo Blindow (VP of Data, Research & Analytics) at @Prolific joining us. Not many seats for the first one.
0
3
20
@max_nlp
Max Bartolo
28 days
RT @MoritzLaurer: Kudos to @cohere for releasing 6 proper research papers in May alone, while publications of other western labs increasing….
0
13
0
@max_nlp
Max Bartolo
1 month
RT @_xjdr: the command-a paper is one of my top 5 papers of the year for sure.
0
18
0
@max_nlp
Max Bartolo
1 month
Another side finding was that in some cases, incoherent preambles also led to improved performance. This has exciting implications for other conditioning token sequences such as reasoning traces.
0
0
2
@max_nlp
Max Bartolo
1 month
Can LLMs be incentivised to generate token sequences (in this case preambles) that condition downstream models to improve performance when judged by reward models? Yes! ✅.
@LisaAlazraki
Lisa Alazraki
1 month
Thrilled to share our new preprint on Reinforcement Learning for Reverse Engineering (RLRE) 🚀. We demonstrate that human preferences can be reverse engineered effectively by pipelining LLMs to optimise upstream preambles via reinforcement learning 🧵⬇️
Tweet media one
1
5
17
@max_nlp
Max Bartolo
2 months
Massive congrats team Afri-Aya, really great work! 🤩.
@JawardSesay_
Jaward Sesay
2 months
Huge Win Today 🎉🎉 Our team “Afri-Aya” just won this year’s @cohere Aya Expedition. Our work is focusing on 1) curating and evaluating vision dataset then 2) Finetuning the Aya vision model for underrepresented languages in Africa. I represented my beloved Sierra Leone with Krio
Tweet media one
1
1
18
@max_nlp
Max Bartolo
2 months
RT @Cohere_Labs: Join us to mark the end of Expedition Aya, our six-week global open-build challenge designed to accelerate ML research pro….
0
1
0
@max_nlp
Max Bartolo
2 months
RT @Cohere_Labs: Congrats to our Cohere colleagues for their paper “Improving Reward Models with Synthetic Critiques” being presented at NA….
0
2
0
@max_nlp
Max Bartolo
2 months
Recently overheard @iclr_conf: influence functions for LLMs are useless. Poster #208 disagrees 🤔
Tweet media one
1
3
51
@max_nlp
Max Bartolo
2 months
RT @egrefen: At #ICLR2025? Come and see @LauraRuis present these amazing results on how LLMs exploit data in different ways to learn facts….
0
12
0
@max_nlp
Max Bartolo
2 months
If you want to learn more about how LLMs pick up reasoning abilities from procedural knowledge in pretraining, visit poster #208 in Hall 3 at 3pm today @iclr_conf #ICLR #ICLR25 #ICLR2025.
@LauraRuis
Laura Ruis
2 months
Presenting this today 3-530 at poster #208, come say hi 🙋‍♀️.
0
5
33
@max_nlp
Max Bartolo
2 months
RT @221eugene: Attending #ICLR2025 and interested in #LLM, #Alignment, or #SelfImprovement?. Then come by and check out our work from .@coh….
0
11
0
@max_nlp
Max Bartolo
3 months
Really enjoyed giving this talk. Thanks for hosting and for the great questions! @tomhosking you might recognise this slide 😅.
@EvolutionAI
Evolution AI
3 months
Another great London Machine Learning Meetup earlier. Many thanks to Max Bartolo (@max_nlp) (researcher at @cohere) for the fascinating talk on 'Building Robust Enterprise-Ready Large Language Models'. And thanks to @ManGroup and @ArcticDB for hosting.
Tweet media one
Tweet media two
1
0
15
@max_nlp
Max Bartolo
3 months
RT @arduinfindeis: How exactly was the initial Chatbot Arena version of Llama 4 Maverick different from the public HuggingFace version?🕵️….
0
6
0
@max_nlp
Max Bartolo
3 months
RT @mgalle: A year ago we released LBBP - a drop-in replacement of HumanEval that was more challenging and less leaked.Internally we have b….
0
13
0
@max_nlp
Max Bartolo
3 months
RT @sarahookr: Very proud to introduce Kaleidoscope ✨🌿. 🌍 18 languages (Bengali → Spanish).📚 14 subjects (Humanities → STEM).📸 55% requirin….
0
30
0
@max_nlp
Max Bartolo
3 months
RNoPE is a much better name than iRoPE anyway.
@cloneofsimo
Simo Ryu
3 months
it looks like meta's new model's "Key innovaton" : "interleaved no-RoPE attention" for infintie context, is actually the same thing cohere command-a model introduced few days ago! They were literally 5 days apart!
Tweet media one
Tweet media two
0
0
3
@max_nlp
Max Bartolo
3 months
RT @nrehiew_: Cohere's Command A report is an extremely extensive paper on how to train a modern LLM in 2025. But it's a model for very dif….
0
143
0