Apoorv Khandelwal
@apoorvkh
Followers
639
Following
3K
Media
19
Statuses
286
cs phd student at brown
Providence, RI
Joined April 2019
In our new paper, we ask whether language models solve compositional tasks using compositional mechanisms. ๐งต
4
27
184
i hate ML conference reviewers. i take back everything bad i ever said about ACL. every ACL reviewer i ever got was at least literate
15
20
479
Would highly recommend working with Noah and Hadar!
This is a great program for folks interested in postdocs @Cornell. For vision/graphics folks interested in NYC, @ElorHadar, @andrewhowens, and I are potentially recruiting a joint postdoc. Please apply!
0
0
7
How is memorized data stored in a model? We disentangle MLP weights in LMs and ViTs into rank-1 components based on their curvature in the loss, and find representational signatures of both generalizing structure and memorized training data
9
62
492
๐จ Reasoning models can โself-jailbreakโ: they recognize a request is harmful, invent a reason why itโs fine, then help with it. We found that after training on benign math/code reasoning, models emergently start to reason themselves out of safety alignment. ๐งต๐
1
6
17
We discovered that language models leave a natural "signature" on their API outputs that's extremely hard to fake. Here's how it works ๐ ๐ https://t.co/Yc7mnhZS96 1/
arxiv.org
The ubiquity of closed-weight language models with public-facing APIs has generated interest in forensic methods, both for extracting hidden model details (e.g., parameters) and for identifying...
3
22
116
Can LLMs reason like a student? ๐ฉ๐ปโ๐๐โ๏ธ For educational tools like AI tutors, modeling how students make mistakes is crucial. But current LLMs are much worse at simulating student errors โ than performing correct โ
reasoning. We try to fix that with our method MISTAKE ๐คญ๐
11
55
337
What does your favorite language model know about the real world? ๐ Can it distinguish between possible and impossible events? We find that LM representations not only encode these distinctions, but that they predict human judgments of event plausibility!
1
7
22
Excited to present our new paper as a spotlight talk ๐ at the Pragmatic Reasoning in LMs workshop at #COLM2025 this Friday! ๐ Come by room 520B @ 11:30am tomorrow to learn more about how LLMs' pluralistic values evolve over reasoning budgets and alignment ๐งต
1
5
28
My thesis, ๐ ๐ต๐ฉ๐ฆ๐ฐ๐ณ๐บ ๐ฐ๐ง ๐ต๐ฉ๐ฆ ๐ค๐ฐ๐ฎ๐ฑ๐ถ๐ต๐ข๐ต๐ช๐ฐ๐ฏ๐ข๐ญ ๐ฑ๐ฐ๐ธ๐ฆ๐ณ ๐ข๐ฏ๐ฅ ๐ญ๐ช๐ฎ๐ช๐ต๐ข๐ต๐ช๐ฐ๐ฏ๐ด ๐ฐ๐ง ๐ญ๐ข๐ฏ๐จ๐ถ๐ข๐จ๐ฆ ๐ฎ๐ฐ๐ฅ๐ฆ๐ญ๐ช๐ฏ๐จ ๐ข๐ณ๐ค๐ฉ๐ช๐ต๐ฆ๐ค๐ต๐ถ๐ณ๐ฆ๐ด, is now online:
8
46
386
Which mechanism is used seems tied to the geometry of embedding space. Models tend to employ the direct mechanism if there's a linear mapping from x to y in the embedding spaces. (Each dot represents a compositional task, see paper for details.) [End of ๐งต]
0
0
9
We use logit lens to identify two processing mechanisms in Llama 3 (3B) as it solves compositions. Left: a compositional mechanism, which computes z = f(x) along the way to y = g(f(x)). Right: a direct mechanism, which doesn't.
1
1
7
Paper: https://t.co/rVtBeqMb7r Code: https://t.co/u0XjYzqPca Brief hightlights below!
1
0
7
Our โacademic pre-trainingโ paper was accepted to COLM! Iโll be presenting at the Tuesday (11 AM) poster session!
Wondering how long it takes to train a 1B-param LM from scratch on your GPUs? ๐งต See our paper to learn about the current state of academic compute and how to efficiently train models! Use our code to test your own models/GPUs! https://t.co/hvrjwlApN8
https://t.co/1JnEe2CCLr
0
3
19
1/ While the world focuses on cognitive intelligence and digital agents, AI deployments in the real world have been stagnant. We are excited to announce our first milestone in building a Physical AI system - Isaac 0.1 - establishing the efficient frontier for perception.
1/ Introducing Isaac 0.1 โ our first perceptive-language model. 2B params, open weights. Matches or beats models significantly larger on core perception. We are pushing the efficient frontier for physical AI. https://t.co/dJ1Wjh2ARK
7
7
31
Super excited that the Computer Use survey I've been working on w/ @anmarasovic for a while now is ready! Originally we were planning on a more traditional survey paper but as more surveys came out we decided on an interactive website survey.
2
9
24
13/ If we fix it, we can scale good science as quickly as we scale bigger models. Thatโs the future worth betting on. https://t.co/yqOdUjAAyM
amplifypartners.com
What if the biggest bottleneck in AI research isn't more ideas โ it's actually testing them?
3
6
23
Thanks to all for making NEMI 2025 a wonderful event. Fascinating talks, inspiring posters, important discussions. You surfaced the questions animating our growing field. I learned many things and hope you did too! Looking forward to what the next year will bring.
2
15
99
Pulling this opportunity on research agent evaluation up one more time! The official title of the position will be "Senior research technician". Feel free to email either @sebschu or me directly if you have any questions. Link for more detailed info and where to apply in ๐งต
2
7
18