Kelly Marchisio (St. Denis) Profile
Kelly Marchisio (St. Denis)

@cheeesio

Followers
2K
Following
3K
Media
48
Statuses
686

Multilingualilty Lead @cohere. Formerly: PhD @jhuclsp, Alexa Fellow @amazon, dev @Google, MPhil @cambridgenlp, EdM @hgse 🔑🔑¬🧀 (@kelvenmar20)

Connecticut, USA
Joined June 2019
Don't wanna be here? Send us removal request.
@cheeesio
Kelly Marchisio (St. Denis)
7 days
RT @Cohere_Labs: @weiyinko_ml was one of the earliest members of our Open Science Community and an early collaborator on our open science r….
0
2
0
@cheeesio
Kelly Marchisio (St. Denis)
9 days
“where is the most extensible place to add this fnct call?” 🤖 Memory jogged: “oh, of course that's the best spot.” . I needed code familiarity and years of non-GenAI-influenced experience to reap the time-savings w/o error; but as a time-strapped manager, this was a godsend! 2/2.
0
0
4
@cheeesio
Kelly Marchisio (St. Denis)
9 days
Wow. Coding with GenAI continues to impress, but highlights the need for strong knowledge of the code *without* help. I opened up a once-familiar codebase, but one I haven’t looked at seriously in a year. Instead of hours spent re-familiarizing, I asked an LLM, (1/2).
1
0
7
@cheeesio
Kelly Marchisio (St. Denis)
9 days
RT @davlanade: Excited to announce the call for papers for the Multilingual Representation Learning workshop #EMNLP2025 .
0
11
0
@cheeesio
Kelly Marchisio (St. Denis)
10 days
RT @weiyinko_ml: We're looking for a new member for the multilingual team with a focus on data engineering! Please apply at the link below:.
0
7
0
@cheeesio
Kelly Marchisio (St. Denis)
10 days
RT @mgalle: Make Command speak better & in more languages.
0
2
0
@cheeesio
Kelly Marchisio (St. Denis)
10 days
The Multilingual Team at @cohere is hiring!. If this sounds like you, please apply:.- strong coding skills and a keen eye for detail.- experience working with the challenges & joys of multilingual data. Help us bring AI to the world! 🌏🌍🌎.
2
29
172
@cheeesio
Kelly Marchisio (St. Denis)
16 days
Code-release from our superstar intern, @p_nawrot!.• Write sparse attn patterns in 50 lines, not 5k .• Compatibility w models supported by vLLM, support for TP.• 6 SOTA baselines with optimized implementations + 9 eval tasks .• Research-grade extensibility = rapid prototyping.
@p_nawrot
Piotr Nawrot
16 days
We built sparse-frontier — a clean abstraction that lets you focus on your custom sparse attention implementation while automatically inheriting vLLM’s optimizations and model support. As a PhD student, I've learned that sometimes the bottleneck in research isn't ideas — it's
Tweet media one
1
1
17
@cheeesio
Kelly Marchisio (St. Denis)
25 days
RT @p_nawrot: We release a major improvement upon last year's Dynamic Memory Compression. DMS is better, easier, and faster to train. Futu….
0
2
0
@cheeesio
Kelly Marchisio (St. Denis)
25 days
RT @PontiEdoardo: 🚀 By *learning* to compress the KV cache in Transformer LLMs, we can generate more tokens for the same compute budget.….
0
27
0
@cheeesio
Kelly Marchisio (St. Denis)
1 month
RT @Cohere_Labs: Here are key recommendations to make AI safer & more equitable for everyone:. 🌐 Incentivize the creation of open-access mu….
0
1
0
@cheeesio
Kelly Marchisio (St. Denis)
1 month
RT @Cohere_Labs: Over 7000 languages are spoken worldwide 🌐, but AI safety efforts focus on only a fraction of them. Our latest paper dra….
0
22
0
@cheeesio
Kelly Marchisio (St. Denis)
1 month
RT @p_nawrot: Tomorrow at 6pm CET I'm giving a talk about our latest work on Sparse Attention, at @Cohere_Labs. I plan to describe the fie….
0
3
0
@cheeesio
Kelly Marchisio (St. Denis)
1 month
RT @Cohere_Labs: Our ML Efficiency group is looking forward to welcoming @p_nawrot next week on May 28th, for a session on "The Sparse Fron….
0
4
0
@cheeesio
Kelly Marchisio (St. Denis)
1 month
Welcome, Ruochen! ✨.
@ruochenz_
Ruochen Zhang
1 month
🌟 Little late but I just started my internship @cohere, cooking more multilingual things with the amazing @cheeesio and @SCahyawijaya. Will be in nyc for June and July, hmu!🗽.
1
0
9
@cheeesio
Kelly Marchisio (St. Denis)
2 months
I’m excited to see what you’ve built! 🚀.
@Cohere_Labs
Cohere Labs
2 months
🧑‍⚖️Our Expedition Aya judges are:. @cheeesio, Multilinguality Lead, Cohere.@max_nlp, Command Modelling Team Lead, Cohere.@mziizm, Staff Research Scientist, Cohere Labs. Let’s celebrate this collaborative research and look ahead to what’s next!. Learn more:
0
0
8
@cheeesio
Kelly Marchisio (St. Denis)
2 months
Result of @robinson_n8’s internship on the Cohere multilingual team last year! Check it out!.
@robinson_n8
Nathaniel R. Robinson
2 months
Many LLMs struggle to produce Dialectal Arabic. As practitioners attempt to mitigate this, new evaluation methods are needed. We present AL-QASIDA (Analyzing LLM Quality + Accuracy Systematically In Dialectal Arabic), a comprehensive eval of LLM Dialectal Arabic proficiency (1/7)
Tweet media one
Tweet media two
0
0
11
@cheeesio
Kelly Marchisio (St. Denis)
2 months
This was fun! Excellent work led by @p_nawrot during his internship at @cohere.
@Cohere_Labs
Cohere Labs
2 months
How does sparse attention reshape LLM scaling? 🔍. We’re excited to share this work by former @Cohere intern @p_nawrot, “The Sparse Frontier: Sparse Attention Trade-offs in Transformer LLMs.”.
0
1
13
@cheeesio
Kelly Marchisio (St. Denis)
2 months
RT @seb_ruder: The Sparse Frontier. Efficient sparse attention methods are key to scale LLMs to long contexts. We conduct the largest-scale….
0
29
0
@cheeesio
Kelly Marchisio (St. Denis)
2 months
RT @_akhaliq: The Sparse Frontier. Sparse Attention Trade-offs in Transformer LLMs
Tweet media one
0
33
0