
Dominik Lukes
@techczech
Followers
2K
Following
3K
Media
2K
Statuses
14K
Exploring applied epistemology, AI and metaphor. Current work on https://t.co/XxygzOGYN0.
UK
Joined April 2009
There cannot be too many Open Source ASR models. Well done @MistralAI for keeping the engine of progress going here. I remember as recently as 3 years ago people saying that speech recognition was never going to be open because of the computational and data demands.
0
0
0
THIS WAS SUPPOSED TO BE A SHORT QUIP! BUT IT BECAME A TREATISE ON METHOD. OH WELL. This is the big epistemic dilemma of our time: aggregate statistic vs the individual case. You can learn more about the nature of things from a case but it's very easy to overlearn. You can learn.
I always learn a lot more from in-depth analysis of few random cases over dashboards of aggregate statistics across all cases. Both projections can be helpful but the latter is disproportionately pervasive.
0
0
0
The thing people get most wrong about metaphor is that it only explains the unfamiliar with the familiar. Often, the thing we are explaining is more familiar to us than the thing we are comparing it to and our perspective on the source changes more than the perspective on the.
@GregKamradt A model… but it’s much much more malleable and clearly has unique properties on lots of layers (like short vs long vs medium memory, etc). And it’s definitely a reasoning model. We emit thought tokens all the time. And clearly multimodal (even smell and taste, should we give.
0
0
1
More and more, LLM training recipes are starting to resemble designing a training program for humans. Not the actual learning process but the selection or creation and then the sequencing of how the training data is presented. Exactly what a syllabus designer will do.
Scaling up RL is all the rage right now, I had a chat with a friend about it yesterday. I'm fairly certain RL will continue to yield more intermediate gains, but I also don't expect it to be the full story. RL is basically "hey this happened to go well (/poorly), let me slightly.
0
0
2
My initial impressions of Kimi K2 by @Kimi_Moonshot. It took me a while to get to it, there seems to be a new open model from China every day but this one does 'feel' important - a definite frontier model smell:. - o3 feel on agentic choices when it comes to search.- Claude 4.
🚀 Hello, Kimi K2! Open-Source Agentic Model!.🔹 1T total / 32B active MoE model.🔹 SOTA on SWE Bench Verified, Tau2 & AceBench among open models.🔹Strong in coding and agentic tasks.🐤 Multimodal & thought-mode not supported for now. With Kimi K2, advanced agentic intelligence
0
1
3
Common @AnthropicAI - this is just pathetic. You need to do better on sign in. Anything you can do to make this less user hostile @alexalbert__ ? Having to re-login into Claude is starting to fill me with dread.
0
0
0
Great study by @METR_Evals and I appreciate the thoroughness and lack of clickbaitiness. Yes, this is about limits of the usefullness of LLMs in some scenarios but I think the key lesson is one of context and knowledge of the context. This is the bit in the full paper that.
We ran a randomized controlled trial to see how much AI coding tools speed up experienced open-source developers. The results surprised us: Developers thought they were 20% faster with AI tools, but they were actually 19% slower when they had access to AI than when they didn't.
0
0
0
RT @ankitdp_: You asked, and it's here: Gems with Canvas.Make your own custom experts with Gemini Gems and get them to write docs and apps….
0
27
0
LLMs have no mental checklist. Here are some slides with examples on what I mean by it. @AnnaRMills
1
0
4