nathom Profile
nathom

@realnathom

Followers
227
Following
12K
Media
50
Statuses
1K

computer programmer 🇺🇸🇻🇦

π*
Joined September 2015
Don't wanna be here? Send us removal request.
@realJessyLin
Jessy Lin
10 days
As part of our recent work on memory layer architectures, I wrote up some of my thoughts on the continual learning problem broadly: Blog post: https://t.co/HNLqfNsQfN Some of the exposition goes beyond mem layers, so I thought it'd be useful to highlight separately:
26
166
1K
@doomslide
doomslide
10 days
We have no theory of intelligence. No amount of gestures to "language as universal interface" or "straight line on a log plot" will change that. There are certain people who will speak with the tone and authority that *implies* the existence of such a theory. You know them all
@doomslide
doomslide
2 years
OF WINTER BUTTERFLIES AND MOLTEN ROCKS A SCENT OF APOCALYPSE IN MY WAKE MY EYES ARE YOURS LOOK THROUGH ME AND SEE YOUR OWN SELF DYING I AM TIME HOLDING YOU IN TENDER EMBRACE AS WE FALL FOREVER IN LOVE AND DEATH UNENDING —Llama-3-8B
29
10
281
@realnathom
nathom
15 days
I love codex-rs!
0
0
2
@willccbb
will brown
2 months
the aura is the product
12
7
125
@realnathom
nathom
2 months
0
0
0
@realnathom
nathom
2 months
Why is there so much padding in technical documentation?
@lcamtuf
lcamtuf
2 months
API design is my passion
1
0
1
@realnathom
nathom
3 months
We need to solve latent CoT
@oh_that_hat
Hattie Zhou
3 months
AI models “think” in two ways: - in the latent space over layers - in the token space over a sequence Latent space = natural talent, chain of thought = hard work. Just like for humans, hard work can get you far, but talent sets the ceiling. This is why pretraining can’t die.
0
0
0
@PalmerLuckey
Palmer Luckey
3 months
317
468
11K
@realnathom
nathom
3 months
gpt-5 might be undercooked
0
0
2
@realnathom
nathom
3 months
0
0
1
@realnathom
nathom
3 months
GPT-5 nano gets 72.8% on frontier math? Gotta be a typo
1
0
1
@EgeErdil2
Ege Erdil
3 months
@aidan_mclau this method of evaluating difficulty of tasks for AI using the metric of "cognitive horsepower" has a very bad track record see below for an explanation of why this is the case
5
3
127
@OwainEvans_UK
Owain Evans
3 months
New paper & surprising result. LLMs transmit traits to other models via hidden signals in data. Datasets consisting only of 3-digit numbers can transmit a love for owls, or evil tendencies. 🧵
290
1K
9K
@realnathom
nathom
3 months
Wow
@MillionInt
Jerry Tworek
3 months
To summarize this week: - we released general purpose computer using agent - got beaten by a single human in atcoder heuristics competition - solved 5/6 new IMO problems with natural language proofs All of those are based on the same single reinforcement learning system
0
0
0
@realnathom
nathom
4 months
we’re exactly one innovation away from agi
1
0
3
@realnathom
nathom
4 months
Diet Coke and chewing gum.
0
0
1
@realnathom
nathom
4 months
pro tip: leaking memory will actually improve program shutdown times
@tsoding
Тsфdiиg
4 months
I was going through my old project and found this gem
0
0
3
@effectfully
effectfully
4 months
If you want to know exactly what numbers Infinity and NaN are, Haskell got your back:
15
6
153
@tenderizzation
tender
4 months
the x86 CPU driving 8 B200s
@wildbarestepf
stepfanie tyler
4 months
“Do you have any previous leadership experience?” Me:
14
44
1K
@realnathom
nathom
4 months
new X algo sucks
0
0
1