Ben
@SolidlySheafy
Followers
373
Following
282
Media
0
Statuses
32
Understanding intelligence @tilderesearch // prev math @Penn and @Cambridge_Uni
Joined March 2024
Please apply!! We are very excited about this program and are looking forward to working with folks on some really interesting questions
Today we're very happy to announce that we’re launching the Tilde Fellowship Program to support research in a mechanistic understanding of pre-training science (arch, optimizers, learning dynamics, etc.). Much of modern ML progress has come from scaling models and empirically
0
0
11
We really enjoyed the @thinkymachines post on Manifold Muon from a couple of weeks ago and decided to share some thoughts. Make sure to check out the original post as well:
thinkingmachines.ai
A geometric framework for co-designing neural net optimizers with manifold constraints.
Modern optimizers can struggle with unstable training. Building off of Manifold Muon, we explore more lenient mechanisms for constraining the geometry of a neural network's weights directly through their Gram matrix 🧠 A 🧵… ~1/6~
2
13
214
Have really enjoyed learning from Alec, hope people like the post!
Vignette #2 is here! Join @AlecDewulf to: Learn about circuit complexity theory Derive theoretical capabilities and limitations of transformers Discuss the future of theoretical computer science in architecture design A thread 🧵
0
1
5
Come by our booth! We are giving out shirts to anyone who can solve our challenge problems :)
We'll be at the Berkeley EECS Career Fair this Tuesday & Wednesday with cool custom puzzles (and prizes). @berkeley_ai @UCBerkeley
0
0
8
Really enjoyed reading this!
Today we’re launching our weekly vignette series! 📚 Vignette #1: Attention through a regression lens 📈 • derive attention from regression • analyze the geometry of kernel smoothers • introduce Epanechnikov attention (feature map + recurrent form) A thread 🧵
0
0
5
We believe there is a lot of interesting work that is not traditionally labeled as interpretability, and we also wanted to share a few words about our overall approach.
At Tilde, we believe mechanistic understanding of models is key to enabling entirely new architectures and capabilities. We’ve put together a position piece on what interpretability means to us. A thread 🧵
0
0
8
Hope this is helpful to others!
Mixture‑of‑Experts (MoE) powers many frontier models like R1, K2, & Qwen3 ⚡️ To make frontier-scale MoE models accessible to train, we open-source MoMoE, a hyper-performant MoE implementation built for training and inference, outpacing the fastest existing ones by up to: - 70%
0
0
5
Excited to share soon what we've been working on since :)
We’re excited to announce that Tilde completed an $8M seed round earlier this year, led by Khosla Ventures. Understanding model intelligence is the most important problem in the world, and the key to actualizing the promise that ASI can offer. 🧵 A thread on our approach:
0
0
13
Some work we did on understanding some of the recent sparse attention mechanisms
Sparse attention (MoBA/NSA) trains faster & beats full attention in key tasks. But we’ve had no idea how they truly work…until now. 🔍 We reverse-engineered them to uncover: - Novel attention patterns - Hidden "attention sinks" - Better performance - And more A 🧵… ~1/8~
1
0
6
🎉Excited to announce today that @erikqu_ and I have been accepted into the X25 @ycombinator batch. Erik and I met in 2014 in high school doing coding competitions together. >10 years later, we're building @operative_sh - the world's first web app coding & debugging agent. 🧵
22
14
112
Hopefully useful contribution to the community!
Today, we open-source Activault, a simple, high-throughput, and cost-effective solution to activation data management for accelerating interpretability research on frontier models. A 🧵… ~1/6~ https://t.co/ZMTbtHRUJD
0
0
6
I taught an LLM to optimize proteins. It proposed a better carbon capture enzyme. Introducing Pro-1, an 8b param reasoning model trained using GRPO towards a physics based reward function for protein stability. It takes in a protein sequence + text description + previous
93
332
3K
And if you don’t like graph theory, but do like interpretability, we have plenty of other fun problems so feel free to email us join@tilderesearch.com. We are doing a lot of applied interpretability work like this: https://t.co/WP9MmyLwCv, which was the first application of
tilderesearch.com
0
1
4
It's a fun problem!
Over the past few weeks, we've been using this graph theory problem in interviews and figured we'd open it up to everyone here! https://t.co/GQCg2rk1Q4 If you solve it, we’ll move you directly to the last rounds of our process!
0
0
3
It was so great to work on this with Adam!
I'm excited about our case study! SAEs can be practical for preventing behavior, especially rare behavior, as we only intervene when needed. Other tools, like finetuning or system prompts, have side effects as they modify every forward pass. Some thoughts below.
0
0
6
We used SAEs to outperform baselines on a real-world task! Check it out!
Mechanistic interpretability is fascinating - but can it be useful? In particular, can it beat strong baselines like steering and prompting on downstream tasks that people care about? The answer is, resoundingly, yes. Our new blog post with @a_karvonen, Sieve, dives into the
0
0
5
Unbelievably excited to announce this. We genuinely want to understand the universe, and think building interpretable AGI is the way to get us there :)
We're thrilled to be launching Tilde. We're applying interpretability to unlock deep reasoning and control of models, enabling the next generation of human-AI interaction. By understanding a model's inner mechanisms, we can enhance both its reliability and performance—going
0
1
11
Warmest Birthday wishes to mathematician extraordinaire Pierre Deligne, Fields Medallist and Abel Prize Winner, Permanent Professor at IHES from 1970 to 1984.
0
11
71