immunity Profile Banner
noah Profile
noah

@immunity

Followers
234
Following
2K
Media
252
Statuses
1K

I will take my agony and turn it to exquisite beauty | emergent ventures

Cheltenham, England
Joined July 2010
Don't wanna be here? Send us removal request.
@immunity
noah
3 months
I'm starting a machine learning society at my uni :) We've partnered with the University of Gloucestershire to get members access to the on-site B200 cluster, to start the year we are going to recreate GPT-2 I'm very excited! hope to be able to share more soon
9
0
30
@immunity
noah
4 months
I am going to be in SF till the 13th, if you are interested in talking to me please reach out, my goal for this trip is to meet as many people in ml research and engineering as possible. I have just updated my personal website (noah dot jp dot net) with the relevant information!
@immunity
noah
4 months
I did it. I reached Highway 101 leading into downtown San Francisco with all the AI SaaS ads. This must be what Hannibal felt like after he cleared the Alpine Pass
2
2
29
@immunity
noah
4 months
I did it. I reached Highway 101 leading into downtown San Francisco with all the AI SaaS ads. This must be what Hannibal felt like after he cleared the Alpine Pass
0
0
8
@cvenhoff00
Constantin Venhoff
5 months
🔍 New paper: How do vision-language models actually align visual- and language representations? We used sparse autoencoders to peek inside VLMs and found something surprising about when and where cross-modal alignment happens! Presented at XAI4CV Workshop @ CVPR 🧵 (1/6)
10
44
303
@UpupWang
Shangshang Wang
5 months
Sparse autoencoders (SAEs) can be used to elicit strong reasoning abilities with remarkable efficiency. Using only 1 hour of training at $2 cost without any reasoning traces, we find a way to train 1.5B models via SAEs to score 43.33% Pass@1 on AIME24 and 90% Pass@1 on AMC23.
10
58
502
@victorveitch
Victor Veitch 🔸
5 months
In particular, I think the current wave of SAE skepticism is about as irrational as the wave of SAE hype that preceded it. It is wholly plausible that there are a set of improvements that will lead to massive improvements. Just, those improvements won't be found by throwing
0
2
30
@EkdeepL
Ekdeep Singh
5 months
🚨 New paper alert! Linear representation hypothesis (LRH) argues concepts are encoded as **sparse sum of orthogonal directions**, motivating interpretability tools like SAEs. But what if some concepts don’t fit that mold? Would SAEs capture them? 🤔 1/11
5
62
386
How does this affect LeBron legacy?
0
1
1
@immunity
noah
5 months
hey its me :)
@tylercowen
tylercowen
5 months
Emergent Ventures winners, 43rd cohort:
7
0
20
@immunity
noah
6 months
0
1
9
@iroasmas
Brian Graham
6 months
am i sure the death star is going down? look at my quant. look at him! you notice anything different about him? look at his eyes. i’ll give you a hint—his name’s a fucking number!! he doesn’t even speak english—it’s all beep-boop shit!! yeah, i’m sure.
27
351
6K
@GoodfireAI
Goodfire
6 months
We created a canvas that plugs into an image model’s brain. You can use it to generate images in real-time by painting with the latent concepts the model has learned. Try out Paint with Ember for yourself 👇
39
99
924
@immunity
noah
6 months
Dont forget learning dynamics of fine tuning and what SFT does to latent space vs DPO either, i think something is brewing…
0
0
3
@immunity
noah
6 months
A lot of alpha in this graph remember “Large Language Models are Zero-Shot Reasoners” as well? I love it when a plan comes together
@xuandongzhao
Xuandong Zhao
6 months
🚀 Excited to share the most inspiring work I’ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
1
1
3
@xuandongzhao
Xuandong Zhao
6 months
🚀 Excited to share the most inspiring work I’ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
84
512
3K
@TheZvi
Zvi Mowshowitz
6 months
The more I look into the system card, the more I see over and over 'oh Anthropic is actually noticing things and telling us where everyone else wouldn't even know this was happening or if they did they wouldn't tell us.'
@ESYudkowsky
Eliezer Yudkowsky ⏹️
6 months
Humans can be trained just like AIs. Stop giving Anthropic shit for reporting their interesting observations unless you never want to hear any interesting observations from AI companies ever again.
9
72
2K
@NidarMMV2
AM
6 months
Even when tasks feel bottlenecked on people (e.g. in lab throughput), it’s not a manpower problem that robotics will fix. Lab automation is expensive, limited in scope, and brittle. You’re not going to run 10k lab experiments overnight, no matter how many robot arms you buy.
1
1
2
@fish_kyle3
Kyle Fish
6 months
🧵For Claude Opus 4, we ran our first pre-launch model welfare assessment. To be clear, we don’t know if Claude has welfare. Or what welfare even is, exactly? 🫠 But, we think this could be important, so we gave it a go. And things got pretty wild…
51
72
657
@rishi_d_jha
Rishi Jha
6 months
I’m stoked to share our new paper: “Harnessing the Universal Geometry of Embeddings” with @jxmnop, Collin Zhang, and @shmatikov. We present the first method to translate text embeddings across different spaces without any paired data or encoders. Here's why we're excited: 🧵👇🏾
36
269
2K
@j0nathanj
Jonathan Jacobi
6 months
Introducing Multiverse: the first AI-generated multiplayer game. Multiplayer was the missing piece in AI-generated worlds — now it’s here. Players can interact and shape a shared AI-simulated world, in real-time. Training and research cost < $1.5K. Run it on your own PC. We
83
196
1K