mixedbreadai Profile Banner
Mixedbread Profile
Mixedbread

@mixedbreadai

Followers
2K
Following
325
Media
19
Statuses
88

Your fav. AI bakers! We're hiring!

San Francisco, CA
Joined March 2024
Don't wanna be here? Send us removal request.
@mixedbreadai
Mixedbread
2 months
Today, we’re opening up access to the beta version of Mixedbread Search. Mixedbread Search is the first system that lets you experience search over your own documents the way it should be: fast, accurate, multilingual, and truly multimodal. Available immediately.
14
43
176
@joeldierkes
Joel Dierkes
12 days
mgrep is now available for opencode. It makes it 4x faster and uses 3x less tokens. simply run mgrep install-opencode
@aaxsh18
Aamir Shakir
24 days
we just made Claude Code - use 53% fewer tokens - respond 48% faster - give 3.2x better responses just by giving it a better grep
40
59
1K
@isaac_flath
Isaac Flath
18 days
I got better quality, half the tokens, and twice the speed with @mixedbreadai 's mgrep The "trick" is a multi-vector architecture Why only now? Because the arch and quantization details are hard Join @aaxsh18 for a research talk on how it works πŸ‘‡ https://t.co/tDdjimst23
Tweet card summary image
maven.com
Mixedbread showed in their launch how much faster/better semantic search could make Claude code. Cursor also just announced semantic embedding support, and other agents are soon to follow. I got...
1
8
30
@aaxsh18
Aamir Shakir
19 days
opus 4.5 is impressive. anthropic forgot a bar in their report. nano banana just added it for me. the leaderboard: https://t.co/sYZyTEjEMw
4
4
39
@aaxsh18
Aamir Shakir
24 days
we just made Claude Code - use 53% fewer tokens - respond 48% faster - give 3.2x better responses just by giving it a better grep
111
133
3K
@joeldierkes
Joel Dierkes
27 days
> be me, go to SF > never did a hackathon before > find the context engineering one from @Theoryvc > claude code yolo with Mixedbread > won > have a new PC
3
6
41
@danielwsms
Daniel Wasmus
2 months
Mixedbread makes image search feel like magic. No keyword labeling, no manual annotation. Describe what you are looking for and instantly find it.
1
5
17
@mixedbreadai
Mixedbread
2 months
Mixedbread 🀝 Vercel Mixedbread Search is now available natively on the @vercel Marketplace. You can now build with our near-instant multi-modal search from the comfort of your Vercel dashboard. Search doesn't need to be complicated, sometimes everything just works.
@rauchg
Guillermo Rauch
2 months
It's so easy to grow developer products. "Just" build incredible demos that show off the technology's capability. @mixedbreadai cooked so hard here. Instant multi-modal AI search, demo'd with the National Gallery of Art dataset. Everything just works and looks fire.
2
6
33
@mixedbreadai
Mixedbread
2 months
Mixedbread 🀝 Vercel Mixedbread Search is now available natively on the @vercel Marketplace. You can now build with our near-instant multi-modal search from the comfort of your Vercel dashboard. Search doesn't need to be complicated, sometimes everything just works.
@rauchg
Guillermo Rauch
2 months
It's so easy to grow developer products. "Just" build incredible demos that show off the technology's capability. @mixedbreadai cooked so hard here. Instant multi-modal AI search, demo'd with the National Gallery of Art dataset. Everything just works and looks fire.
2
6
33
@mixedbreadai
Mixedbread
2 months
Despite sticking to standard practices and datasets, the resulting models are strong performers. They are some of the best long-context models around, and the 17M one matches ColBERTv2 with 1/6th the parameters and 1/3rd the embed dimension: a perfect edge device retriever!
1
0
4
@mixedbreadai
Mixedbread
2 months
In the process, we also conducted many ablations to understand what makes a ColBERT model better, and what does not. Among other things, we find that the embedding dimension can be lowered quite aggressively while retaining most of the model's performance.
1
0
2
@mixedbreadai
Mixedbread
2 months
So we decided to train one, and produce a report on the current best practices on how to take a language model all the way to being an extremely capable retrieval model. The report compiles common retrieval wisdom in one place, to give a clear view of the full process.
1
0
6
@mixedbreadai
Mixedbread
2 months
Scaling laws are our greatest friend when running experiments: we can test ideas out to figure out which ones are worth scaling. But for ColBERT models, we didn't know which one to use! In 2025, there still were no flash-attention, long-context small ColBERT available!
1
0
8
@mixedbreadai
Mixedbread
2 months
One More (Small) Thing: Introducing mxbai-colbert-edge-v0 17M and 32M. They are are the result of an easily reproducible way to train ColBERT models from scratch. They're strong, too: the 17M variant would rank first on the LongEmbed leaderboard for models under 1B parameters.
5
23
128
@mixedbreadai
Mixedbread
2 months
This is a Fast Rising Science post: what we hope will be a long series of papers and blog posts about narrow, focused studies that help us understand how retrieval works. Paper: https://t.co/iBIKzQM3lI Blog post:
Tweet card summary image
mixedbread.com
Discussing the unique learning constraints introduced by the MaxSim operator, and demonstrating that simple architecture improvements to accommodate for these limitations can increase performance in...
0
1
18
@mixedbreadai
Mixedbread
2 months
The effects are varied, but ultimately, most projection variants are beneficial, and the better ones (which, surprisingly enough, remain very simple) introduce noticeable NDCG@10 gains at almost no cost (other than a negligible increase in parameters).
1
0
8
@mixedbreadai
Mixedbread
2 months
We first highlight the learning properties of MaxSim and the limitations that it comes with. Then, we build on it to propose better architectures for the final projection of ColBERT models, inspired by common deep learning feedforward block designs.
1
0
7
@mixedbreadai
Mixedbread
2 months
Better Projection Variants Improve ColBERT We introduce a rare free lunch: a tiny architectural modification that improves ColBERT performance across the board, without any real tradeoffs. The modification? Better final projections than the currently used linear projection.
5
16
93
@nainglk
Naing Linn Khant
2 months
You can now add AI Search to your @astrodotbuild Starlight 🌟 documentation "With Mixedbread, search results instantly felt more relevant, contextually accurate, and tailored to the query that the user entered." - @imax153(@EffectTS_) Read our guide how you can get started πŸ‘‡
2
6
15