Zach Mueller
@TheZachMueller
Followers
13K
Following
46K
Media
2K
Statuses
20K
Hardware nerd. Usually yelling at NCCL over things
Baltimore, MD
Joined April 2016
Introducing OlmoEarth đ, state-of-the-art AI foundation models paired with ready-to-use open infrastructure to turn Earth data into clear, up-to-date insights within hoursânot years.
10
58
263
Here is an excellent article that explains the differences between Context Parallelism (Ring Attention) and Ulysses Sequence Parallelism (head parallelism) and how the 2 can be combined together for a 2D CP+SP https://t.co/GJT6OuhEUJ
1
14
119
In the Smol Training Playbook, I tried to survey the state of popular post-training frameworks. Let me know if I missed any and I'll add them to the list!
18
13
180
It's that time of the year again and we're coming with another @GPU_MODE competition! This time in collaboration with @nvidia focused on NVFP4. Focused on NVFP4 and B200 GPUs (thanks to @sestercegroup ) we'll release 4 problems over the following 3 months: 1. NVFP4 Batched GEMV
7
11
173
This might be the first KVM I like. Switches between all 3 OSâs (mac, windows, Linux) on separate machines without bugs/issues and can do full display resolutions (144hz via DP etc) smoothly. Iâve tried a few and thoroughly impressed. Iâll have a link below (not affiliated)
2
1
13
Related goal this week to⌠news đ Going to see if I canât manage to get ~30Gbps purely off 2 USB-C -> Ethernet + 2 10GbE ports. Wish me luck
0
0
8
General question: how many TB of storage before you consider yourself a data center? Asking for a friend
3
0
11
Workhorse will hit its final form this week. More news soon đ (And become much less of an abominationTM)
1
0
6
It's the FINAL DAY to sign up for the last cohort of Scratch to Scale! Admission ends at midnight EST, join now while you can: https://t.co/zf7HL3Co8y
0
4
6
đ Qwen3-VL is now available on llama.cpp! Run this powerful vision-language model directly on your personal devicesâfully supported on CPU, CUDA, Metal, Vulkan, and other backends. Weâve also released GGUF weights for all variantsâfrom 2B up to 235B. Download and enjoy! đ đ¤
huggingface.co
45
203
1K
I'm convinced to try it asap, we should all try fp16, look at this plot man. FP16 is like perfect in error reduction. "This is precisely why switching to FP16 provides a fundamental solution. With its 10 mantissa bits, FP16 offers 8 times more precision (2^10 values vs. 2^7
25
43
653
every decommissioned V100 coming out of retirement after hearing that the future of RL is fp16
11
27
441
if you want to contribute to open-source but don't know where to begin and just want to use AI, please don't avoid writing AI comments to GH repository issues, not only you are taking maintainers' time but it's also misleading for other devs same goes for PRs, most
11
8
158
some personal news: i recently joined @NousResearch. excited to learn and do cool stuff with very smart people
22
6
149
Happy Halloween from Reachy Mini! You'll be able to 3D print these skins at home thanks to open-source
31
71
599
Elie, our đ, wrote a banger with other HF folks. I beg, please read if nothing else. If faced with my course and this, please go read this.
Training LLMs end to end is hard. Very excited to share our new blog (book?) that cover the full pipeline: pre-training, post-training and infra. 200+ pages of what worked, what didnât, and how to make it run reliably https://t.co/iN2JtWhn23
2
2
66