
John T Davies πΊπ¦πͺπΊπ
@jtdavies
Followers
2K
Following
14
Media
609
Statuses
4K
Entrepreneur, CTO in Gen-AI, investor, father to 3 grown boys, husband to Rachel, astrophysicist, keen photographer, cyclist, ΓΌber-geek, travelled a lot.
West London, England
Joined August 2008
Why not just buy some Mac Minis and use the Qwen3-Coder-30B model on MLX? You have a more scalable cost model.
What if we just rented a single H200 node from @PrimeIntellect and serve Qwen3-Coder-480B-A35B model!. Cost: ~$17k/month. We only need 280 people paying $60/month to cover it. Letβs crowd-power next-gen AI together. Whoβs in?
2
0
2
Spot on Wolfram and nice work!.
π¨ BREAKING: China is no longer catching up; they're setting the pace!. Six Qwen3 models released in one week: from big ones that surpass all open models and nearly all closed AIs to small versions that can run on your laptop - each SOTA and top-tier in its class. I've been
0
0
2
I agree with the points, especially #6.
Ex-Alibaba CTO just made the boldest claim about AI & global power:. βChina is building the future of AI, not Silicon Valley.β. He also revealed why AI by 2030 will look nothing like ChatGPT and how Chinaβs approach is already decades ahead. Here are my top 7 takeaways: π§΅
0
0
1
A few months back OpenAI talked about an open model. Qwen3 has not only filled the top spot, theyβve dominated thinking and instruct with 235B and 30B LLMs along with translate and coding models too. Several releases a week, true AI innovation.
these two days we are sharing the instruct and thinking models of our smaller variant of the 2507 seires, 30a3-2507. fast, but much smarter than before. i like this size, it is just something that i can easily play with, which is also somehow smart enough. btw, i hope we can.
0
0
0
Same here Eric, this may be another nail in their coffin along with GLM 4.5, Qwen3 and Kimi K2. Iβve been a huge fan of sonnet and Opus but things move fast in AI.
Dear @AnthropicAI,.I've been subscribed to your highest level plan since the day you started accepting credit cards. I've finally cancelled my subscription.
0
0
2
Thanks to some background research from @ivanfioravanti, the excellent MLX work of @Prince_Canuma, the incredible model from @JustinLin610 and team at @Alibaba_Qwen, I got one of the world's most powerful LLMs running on my 128GB M4 laptop. Qwen3-235B-A22B-Instruct-2507-3bit-DWQ
3
3
30
So far, three groundbreaking models in a week and another tomorrow. This model alone will help 95% of the worldβs population talk to each other (for free). Probably one of the greatest gifts to humanity in a while!.
π Introducing Qwen3-MT β our most powerful translation model yet!. Trained on trillions of multilingual tokens, it supports 92+ languagesβcovering 95%+ of the worldβs population. πβ¨. π Why Qwen3-MT?.β
Top-tier translation quality.β
Customizable: terminology control, domain
0
0
3
Iβm back in Munich, the organisers of @mlconference ask me last minute if I could teach a 2-day AI course. Stupidly I said yes and then spent all weekend writing exercises. Monday done, time for a beer πΊ
1
1
25
π€£πΊ.
Ever wondered how model quantization (FP16, Q8, Q4) *really* affects performance?. There's an analogy that makes the trade-offs crystal clear. and it involves something you might drink. ππΊ. Kudos to @jtdavies for this brilliant comparison. π. See the image for the full
1
0
2
Please join me on Tuesday evening (21h EU/ 20h UK time) for a chilled out discussion on everything AI, specifically open-source, local & private LLMs. This is run by @AidfulAI and we usually get @WolframRvnwlf on the space too.
1
0
3