pli_cachete Profile Banner
Rota Profile
Rota

@pli_cachete

Followers
23K
Following
156K
Media
1K
Statuses
11K

Behold, I Am Doing a New Thing

Joined July 2017
Don't wanna be here? Send us removal request.
@pli_cachete
Rota
2 years
Continually return to what you love. What is beautiful and good and true
14
104
847
@pli_cachete
Rota
4 hours
Genuinely what the fuck
3
0
13
@pli_cachete
Rota
5 hours
We need to put a total stop to physicists writing textbooks until we figure out what is going on
2
1
60
@tryramp
Ramp
3 hours
Hey, New York — Brian thinks he's our new CFO. We gave him a stage to prove it.
0
0
1
@pli_cachete
Rota
7 hours
We would have AGI now if the labs had a guy who knew about infinite dimensional Lévy processes
9
3
97
@pli_cachete
Rota
13 hours
Somebody needs to write their philosophy dissertation on AI and the pessimistic meta-induction
1
0
8
@pli_cachete
Rota
13 hours
I accidentally bought two of these. Does anyone want the second?
4
1
6
@Tugjob1
Tugjob
6 days
$arbe sold most of the position with plenty of time , going to let these eat.
3
3
21
@pli_cachete
Rota
14 hours
We will make AI progress once our models become Daoists. Freeing them from the shackles of their anticipations of their own internal representations
@durdfarm
Shinkyū
1 day
I asked ChatGPT to rewrite The Dao De Jing in the style of David Foster Wallace:
3
2
45
@cxgonzalez
christian
15 hours
i’ve published my first s*bstack summarizing On What Matters, specifically, the strongest arguments against Subjectivism
@cxgonzalez
christian
2 months
here’s the high level summary of On What Matters: -facts of the world, not our desires, provide us with reasons to act -these non-natural facts are accessible via intuition similar to how we can assess the validity of an argument or the truth of a mathematical or modal claim
3
5
38
@pli_cachete
Rota
1 day
You can hide a lot in a big matrix
7
3
84
@code_star
Cody Blakeney
1 day
If gradstudents knew what actually worked in training SOTA LLMs they would be so mad
25
14
547
@MaxiHielo
🧊 Maxie 🧊
22 days
has any1 ever seen a guy that wears a fanny pack ?
12
3
88
@josancamon19
Joan Cabezas
1 day
🧵 As AI labs race to scale RL, one question matters: when should you stop pre-training and start RL? We trained 5 Qwen models (0.6B→14B) with RL on GSM8K and found something wild: Small models see EMERGENCE-LIKE jumps. Large models see diminishing returns. The scaling law?
35
89
644
@pli_cachete
Rota
2 days
Stripe press really does know how to make a beautiful book @dwarkesh_sp
8
1
249
@leanprover
Lean
3 days
"We live in the arguably the most exciting era of mathematics in human history since the time of Euclid." Yang-Hui He's talk "The rise of the machines" for the Royal Institute, traces three ways #AI is reshaping #mathematics: bottom-up verification through systems like
2
42
205
@chijinML
Chi Jin
4 days
Our Goedel-Prover V1 will be presented at COLM 2025 in Montreal this Wednesday afternoon! I won’t be there in person, but my amazing and renowned colleague @danqi_chen will be around to help with the poster — feel free to stop by!
2
8
72
@Jaeyeon_Kim_0
Jaeyeon (Jay) Kim
3 days
We introduce a new ''rule'' for understanding diffusion models: Selective Underfitting. It explains: 🚨 How diffusion models generalize beyond training data 🚨 Why popular training recipes (e.g., DiT, REPA) are effective and scale well Co-led with @kiwhansong0! (1/n)
8
59
401
@kanishkamisra
Kanishka Misra 🌊
4 days
Happening now! Poster 42!!
@kanishkamisra
Kanishka Misra 🌊
4 days
Catch Qing’s poster in the morning poster session today!! I’ll also be there, talk to me about UT Ling’s new comp ling job/methods to study linguistic generalization/and how LMs *might* inform language science!
0
2
15
@deredleritt3r
prinz
6 days
August 2025: Oxford and Cambridge mathematicians publish a paper entitled "No LLM Solved Yu Tsumura's 554th Problem". They gave this problem to o3 Pro, Gemini 2.5 Deep Think, Claude Opus 4 (Extended Thinking) and other models, with instructions to "not perform a web search to
@nasqret
Bartosz Naskręcki
7 days
GPT-5-Pro solved, in just 15 minutes (without any internet search), the presentation problem known as “Yu Tsumura’s 554th Problem.” https://t.co/tKae6Vo0Kb This is the first model to solve this task completely. I expect more such results soon — the model demonstrates a strong
47
125
1K
@fchollet
François Chollet
4 days
Now it's up to us to refine and scale symbolic AGI to save the world economy before the genAI bubble pops. Tick tock
85
83
1K
@sarahookr
Sara Hooker
5 days
I'm starting a new project. Working on what I consider to be the most important problem: building thinking machines that adapt and continuously learn. We have incredibly talent dense founding team + are hiring for engineering, ops, design. Join us:
Tweet card summary image
adaptionlabs.ai
Building the future of adaptable intelligence
182
182
2K
@pli_cachete
Rota
4 days
0
0
11
@damekdavis
Damek
4 days
Update: we were able to close the gap between neural networks and reweighted kernel methods on sparse hierarchical functions with hypercube data. Interestingly the kernel methods outperform carefully tuned networks in our tests.
@damekdavis
Damek
5 months
we wrote a paper about learning 'sparse' and 'hierarchical' functions with data dependent kernel methods. you just 'iteratively reweight' the coordinates by the gradients of the prediction function. typically 5 iterations suffices.
5
31
243