Tobias_Writes Profile Banner
Tobias Mann Profile
Tobias Mann

@Tobias_Writes

Followers
1K
Following
2K
Media
483
Statuses
3K

Systems Editor @TheRegister / @SitPub — hiker, animal lover, photographer, blogger, and tech journo. I'm over on Mastodon now at @[email protected]

Denver, CO
Joined June 2017
Don't wanna be here? Send us removal request.
@Tobias_Writes
Tobias Mann
2 days
Compiling code an Raspberry Pi 5 is painfully slow
Tweet media one
1
0
5
@Tobias_Writes
Tobias Mann
2 days
OpenAI's decision to use MXFP4 datatypes as the standard for gpt-oss is a big deal and sets the tone for the rest of the industry. Is it perfect? No, but it's a huge improvement over FP4 or INT4. Find my deep dive @theregister .
theregister.com
Analysis: Decision to use MXFP4 makes models smaller, faster, and more importantly, cheaper for everyone involved
0
0
2
@Tobias_Writes
Tobias Mann
3 days
That's my arm if you're wondering. Not sure why I felt the need to share that, but now you know.
@TheRegister
The Register
3 days
Breathe easy: Apple Watch can read your oxygen levels again
0
1
1
@Tobias_Writes
Tobias Mann
3 days
Main thing confusing me: why give up an optimized training stack built around Nvidia to take a chance on Huawei? Like other than national pride. The 910C doesn't support FP8, so back to BF16. Were they using them for RL?. Latest for @TheRegister .
Tweet card summary image
theregister.com
: Chinese AI model dev still plans to use homegrown silicon for inferencing
0
0
1
@Tobias_Writes
Tobias Mann
3 days
Gentle reminder that GPUs aren't just for AI. In the right hands they can advance life saving science. Boffins at LLNL used the No. 1 ranked supercomputer, El Capitan, to build a real-time tsunami forecast system. My latest for @theregister .
Tweet card summary image
theregister.com
: The world's most powerful known supercomputer stretches its legs with some life-saving science
0
0
2
@Tobias_Writes
Tobias Mann
4 days
Need one of these for review. I want to type -tp 8 into vLLM.
@MaxsunOfficial
Maxsun Official
4 days
🔥 Workstation Project Battlematrix.#MAXSUN W790-112L × #Intel Arc B60 Dual 48G Turbo = 8 ARC B60 Cores & 192GB VRAM. A complete workstation solution for AI, rendering, and simulations — built for industrial-grade performance.
Tweet media one
1
0
3
@Tobias_Writes
Tobias Mann
4 days
Got GPT-OSS 120B running tonight in Llama.cpp. ~60Tok/s with minimal expert offload to DRAM. Had to compile a pull request to make it play nice to Open WebUI, but I have tool calling now. Need to figure out a system prompt cause my god this thing loves tables.
0
1
1
@Tobias_Writes
Tobias Mann
4 days
Meanwhile, if you haven't figured it out already GPT-5 is a huge improvement, just mostly for OpenAI's bottom line. This thing is a big cost cutting measure and not necessarily in a bad way:. My latest for @TheRegister .
Tweet card summary image
theregister.com
Comment: Gotta pay for all those GPUs somehow
0
2
2
@Tobias_Writes
Tobias Mann
4 days
Still need to do some cable management and it remains to be seen whether that PSU will be adequate, but it posted. 24 Zen4 cores.128GB DDR5 5600 RDIMM.2x RTX 3090 TI 24GB w/ NVL. And don’t worry those GPUs are power limited to 250W. The PSU shouldn’t explode.
Tweet media one
1
1
5
@Tobias_Writes
Tobias Mann
4 days
Well of course the Chinese don't want H20s anywhere near their sensitive government workloads. Why give Uncle Sam any more reasons to think kill switches are a good idea. My latest for. @theregister.
Tweet card summary image
theregister.com
: Don't need to give Uncle Sam any more reason to think kill switches are a good idea
0
0
2
@Tobias_Writes
Tobias Mann
5 days
Challenge: go a week with gpt-oss (either 20B or 120B) as your only AI assistant. Could you do it?. I have a feeling it's gonna be painful, but I'm going to try.
0
0
1
@Tobias_Writes
Tobias Mann
5 days
I did a thing…
Tweet media one
0
1
4
@Tobias_Writes
Tobias Mann
5 days
When Llama 4.1?.
0
0
2
@Tobias_Writes
Tobias Mann
10 days
Okay that's a wrap on this thread. Almost time to tune into the GPT-5 launch preso. [Fin].
0
0
0
@Tobias_Writes
Tobias Mann
10 days
But while gpt-oss is a big step up for US open weights models, the best models in this category are still Chinese. Qwen3-Coder and Qwen3-A235B-A22B-Instruct is very, very good for its size, and its just one of many open weights models out of China to choose from. [5].
1
0
0
@Tobias_Writes
Tobias Mann
10 days
Now the ball is in Meta's court. Fingers crossed Llama 4.1 and 4.2 are more impressive than what we got back in April. It'd be great to see Google, Microsoft, and IBM step up their game with more capable Gemma, Phi, and Granite models too. [4].
1
0
0
@Tobias_Writes
Tobias Mann
10 days
But if you need an American made model for regulatory and compliance reasons, gpt-oss is hard to beat. Plus, I'm sure application-specific fine tunes will be more reliable. [3].
1
0
0
@Tobias_Writes
Tobias Mann
10 days
They are not perfect. The gpt-oss-20B halucinates a bit too much and its refusal to admit when its wrong is problematic. [2].
Tweet card summary image
theregister.com
: gpt-oss-20b can't seem to decide who won the election, but tried to convince us that it was Biden
1
0
0
@Tobias_Writes
Tobias Mann
10 days
After having a few days to mess with OpenAI's gpt-oss — specifically the 21B parameter version — I gotta admit, these are the most impressive open models to come out of the US so far this year. A 🧵 [1]
Tweet media one
1
0
0