Jay Alammar
@JayAlammar
Followers
48K
Following
3K
Media
528
Statuses
2K
Machine Learning Researcher and writer https://t.co/5GlbofAHs0. O'Reilly Author https://t.co/Fl3uPAZHLg. LLM Builder @Cohere.
Joined April 2020
We're ecstatic to bring you "How Transformer LLMs Work" -- a free course with ~90 minutes of video, code, and crisp visuals and animations that explain the modern Transformer architecture, tokenizers, embeddings, and mixture-of-expert models. @MaartenGr and I have developed a
Announcing How Transformer LLMs Work, created with @JayAlammar and @MaartenGr, co-authors of the beautifully illustrated book, “Hands-On Large Language Models.” This course offers a deep dive into the inner workings of the transformer architecture that powers large language
25
218
1K
Welcoming a new state-of-the-art reranker, Rerank 4! :) It's smarter, faster, and packed with insanely powerful features no other model has. Give it a try on the Cohere API, AWS Sagemaker, and Azure Foundry! Let us know what you think!
Introducing our latest breakthrough in AI search and retrieval: Rerank 4! It’s the most advanced set of reranking models on the market, with best-in-class performance across search relevance, speed, deployment flexibility, multilingual support, and domain-specific understanding.
6
21
235
Congrats @Azaliamirh and @annadgoldie!
Thrilled to share that @annadgoldie and I are launching @RicursiveAI, a frontier lab enabling recursive self-improvement through AIs that design their own chips. Our vision for transforming chip design began with AlphaChip, an AI for layout optimization used to design four
1
0
25
The Illustrated NeurIPS 2025: A Visual Map of the AI Frontier New blog post! NeurIPS 2025 papers are out—and it’s a lot to take in. This visualization lets you explore the entire research landscape interactively, with clusters, summaries, and @cohere LLM-generated explanations
24
213
1K
30 agents. Cameras ordered off. @EricTrump recounts the moment the raid began. Watch the full interview.
0
37
120
This video by @jbhuang0604 manages to cram in all the core pieces of modern attention variants, a perfect refresher if you (like me) need a reminder of the differences between MHA, GQA, MLA, DSA etc :) https://t.co/UrjLHvn1v2
5
84
632
Visualization: https://t.co/1hbBPp2NV5 How it was built:
newsletter.languagemodels.co
Using Cohere's Command A Reasoning and Embed 4 to Visualize the ~6,000 Papers Accepted to NeurIPS 2025
0
3
15
As a good bulk of the AI research community packs its bags to Neurips 2025, some of the year's key research progress is already available for all to read. I built the interactive visualization below to enable you to quickly find and understand the most relevant work to you
8
16
122
Today, Cohere is expanding its partnership with @SAP to meet the increasing demand for sovereign AI technology across Europe and other key global markets. Together, we plan to make our agentic AI platform, North, available on SAP’s infrastructure. Read the blog to learn more -
5
15
79
Sadly I won't be able to attend #NeurIPS2025 this year. But you should attend for me ... A Celebration of All Things Open Source sponsored by @kaggle & @HudsonLabs. https://t.co/enfOHTXh5o
1
1
14
Here it is! The cover of our "An Illustrated Guide to AI Agents" book! Congrats to those who have guessed correctly! We'll randomly draw from people who guessed correctly and reach out to arrange prize delivery. The animal pick is done by @OReillyMedia based on a secret method.
13
143
1K
Excited to see and support the next chapter of @askalphaxiv! Congrats @rajpalleti314 and team!
We just raised a $7M Seed round co-led by @MenloVentures and @haystackvc with participation from @Shakti_VC, @conviction and @upfrontvc 🚀 We're honored to have the support of incredible angels including @ericschmidt, @SebastianThrun, @sarahookr Join us: https://t.co/IKwK8KsG96
1
1
16
The Transformer was built for translation back in 2017. Very proud of the team for setting a new SOTA with Comman A Translate!
Our team presented the process behind Command A Translate, our model that set a new industry standard for secure, enterprise-ready translation at WMT 2025 this month. In our paper, we demonstrate our training and evaluation details, including the data preparation pipeline that
4
11
125
Today we’re releasing Deep Research Tulu (DR Tulu)—the first fully open, end-to-end recipe for long-form deep research, plus an 8B agent you can use right away. Train agents that plan, search, synthesize, & cite across sources, making expert research more accessible. 🧭📚
13
123
667
Post your guesses in a comment. Each person gets two guesses. Contest Rules: - Each participant is allowed two guesses - 3 lucky winners will be randomly selected from those who guessed correctly to receive a free print copy of our book when it hits the shelves (note: we cannot
oreilly.com
What's that animal on the front of your O'Reilly book? You'll find it listed here.
4
0
2
It's time to guess the animal on the cover! Guess the animal on the cover of our book, An Illustrated Guide to AI Agents, for a chance to win a free copy! There's a secret method that assigns the animals of @OReillyMedia books. Even @MaartenGr and I as authors don't even know
66
6
159
It’s time to guess the animal on the cover! Our cover animal is here, and @JayAlammar and I are thrilled to celebrate it with an exciting contest for our upcoming release, An Illustrated Guide to AI Agents! Each person gets two guesses. Which animal do you think we got?
2
2
10
For anyone interested in feature attribution and feature importance for UMAP, this looks like a powerful approach that can provide it for Parametric UMAP.
@ArcadiaScience UMAP is everywhere because it's great at creating visually distinct clusters from high-dimensional data like gene expression. But there's a catch: the nonlinear mapping makes it hard to interpret which features are responsible for those clusters. https://t.co/IE1pSB6enH [2/8]
0
3
17
I'm excited to announce my RLHF Book is now in pre-order for the Manning Early Access Program (MEAP), @ManningBooks, and for this milestone it's 50% off. Excited to land in print in early 2026! Lots of improvements coming soon. Link below & thanks for the support!
42
72
777
Wrote a 1-year retrospective with @a1zhang on KernelBench and the journey toward automated GPU/CUDA kernel generations! Since my labmates (@anneouyang, @simran_s_arora, @_williamhu) and I first started working towards this vision around last year’s @GPU_mode hackathon, we have
11
66
291