LauraRuis Profile Banner
Laura Ruis Profile
Laura Ruis

@LauraRuis

Followers
5K
Following
4K
Media
61
Statuses
1K

PhD with @_rockt and @egrefen. Inc. postdoc with @jacobandreas @MIT_CSAIL. Prev. FAIR, Google, NYU. Anon feedback: https://t.co/sbebAl53tU

London
Joined October 2019
Don't wanna be here? Send us removal request.
@LauraRuis
Laura Ruis
7 months
How do LLMs learn to reason from data? Are they ~retrieving the answers from parametric knowledge🦜? In our new preprint, we look at the pretraining data and find evidence against this:. Procedural knowledge in pretraining drives LLM reasoning ⚙️🔢. 🧵⬇️
Tweet media one
24
209
969
@LauraRuis
Laura Ruis
13 hours
RT @alexUnder_sky: @LauraRuis yeah, Akbir might be the goat of AI safety . And I like his Shaper (multi-agent opponent modelling as a form….
0
1
0
@LauraRuis
Laura Ruis
13 hours
Highly recommend reading this, or at least the intro and conclusion. Some gems about the future of safety research.
@akbirkhan
akbir.
15 hours
here is my thesis “Safe Automated Research”. i worked on 3 approaches to make sure we can trust the output of automated researchers as we reach this new era of science. it was a very fun PhD
Tweet media one
1
2
16
@LauraRuis
Laura Ruis
2 days
The goal isn't perfect logic. It's reasoning that contributes to knowledge, builds understanding, and makes human life better. Given the fact that AI doesn't share many of human's inherent limitations, achieving this could be very powerful.
0
0
7
@LauraRuis
Laura Ruis
2 days
Instead of forcing AI to mimic formal logic, we should see reasoning as fundamentally social and goal-oriented. The question becomes: How do we align AI reasoning with human values when we can't logically verify every step?.
1
0
4
@LauraRuis
Laura Ruis
2 days
Reasoning then works best not in isolation, but embedded in communities that coordinate towards shared goals. It's a social technology, not just a purely cognitive one.
1
0
4
@LauraRuis
Laura Ruis
2 days
Through this lens, reasoning works just fine. It helps us convince others, defend our positions, and build shared understanding.
1
0
2
@LauraRuis
Laura Ruis
2 days
Mercier & Sperber's interactionist theory offers a reframe: reasoning didn't evolve for individual cognition. It evolved for social cooperation: to justify our beliefs, build trust, and coordinate with others.
3
1
8
@LauraRuis
Laura Ruis
2 days
Human reasoning is famously flawed: biased toward existing beliefs and vulnerable to misinformation. If reasoning evolved to enhance cognition, why is it so broken?.
2
0
3
@LauraRuis
Laura Ruis
2 days
Logical reasoning isn't how humans naturally think, it's our attempt to reverse-engineer intuition into rules. Perhaps not coincidentally, AI only became useful when we abandoned rule-based systems.
4
0
57
@LauraRuis
Laura Ruis
2 days
RT @EkdeepL: 🚨New paper! We know models learn distinct in-context learning strategies, but *why*? Why generalize instead of memorize to low….
0
64
0
@LauraRuis
Laura Ruis
3 days
RT @MinqiJiang: Recently, there has been a lot of talk of LLM agents automating ML research itself. If Llama 5 can create Llama 6, then sur….
0
175
0
@LauraRuis
Laura Ruis
7 days
RT @ammar__khairi: 🚀 Want better LLM performance without extra training or special reward models?.Happy to share my work with @Cohere_labs….
0
20
0
@LauraRuis
Laura Ruis
8 days
RT @_rockt: Fantastic work by @JonnyCoook and @silviasapora on "Programming by Backprop: LLMs Acquire Reusable Algorithmic Abstractions Dur….
0
13
0
@LauraRuis
Laura Ruis
8 days
RT @silviasapora: 🧵 Check out our latest preprint: "Programming by Backprop". What if LLMs could internalize algorithms just by reading cod….
0
20
0
@LauraRuis
Laura Ruis
9 days
Many more interesting findings in the preprint: Awesome work by first authors @JonnyCoook and @silviasapora , and collaborators @aahmadian_ , @akbirkhan , @_rockt , and @j_foerst.
4
0
31
@LauraRuis
Laura Ruis
9 days
We just discovered this and models show limited performance, but we have many follow-up questions. We find big improvements with data and model scaling, so how far can we take this? How many programs can we store in the LLM? Can models internalise things like recursion or search?.
1
0
15
@LauraRuis
Laura Ruis
9 days
We hypothesise that PBB may be an additional reason why code training helps downstream reasoning; models internalise reusable algorithmic abstractions.
1
0
14
@LauraRuis
Laura Ruis
9 days
Strikingly, using RL in the second stage allows models to evaluate programs seen in an earlier SFT stage (without I/O pairs). By contrast, SFT is much worse at this backward generalisation. Our ablations indicate it's the on-policy nature of RL enabling this.
Tweet media one
Tweet media two
1
2
17
@LauraRuis
Laura Ruis
9 days
We also find LLMs can sometimes compose programs only trained on in isolation. Training on foo() and bar() as standalone code snippets allows models to eval baz(x) = foo(bar(x)). GPT-4o can do this w/o CoT, evaluating compositions in-weights (a form of out-of-context reasoning).
Tweet media one
1
0
16
@LauraRuis
Laura Ruis
9 days
We find models can learn an input-general understanding of algorithms from a *single* piece of code. This indicates training LLMs with next-token prediction on source code can overcome (part of) the embers of autoregression (@RTomMcCoy).
Tweet media one
2
1
21