Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹ Profile
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹

@isabelpapad

Followers
1K
Following
1K
Media
16
Statuses
77

(jolly good) Fellow at @KempnerInst, incoming assistant professor at @UBCLinguistics (Sept 2025). PhD @stanfordnlp with the lovely @jurafsky.

Joined November 2020
Don't wanna be here? Send us removal request.
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
9 months
Do you want to understand how language models work, and how they can change language science? I'm recruiting PhD students at UBC Linguistics! The research will be fun, and Vancouver is lovely. So much cool NLP happening at UBC across both Ling and CS!
Tweet media one
9
72
328
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
21 hours
RT @jurafsky: Now that school is starting for lots of folks, it's time for a new release of Speech and Language Processing! Jim and I adde….
0
60
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
26 days
RT @aclmeeting: πŸ“… 25-Year ToT Award (2000). Daniel Gildea & Daniel Jurafsky:. β€œAutomatic Labeling of Semantic Roles”. ACL 2000. πŸ”— https://t….
Tweet card summary image
aclanthology.org
Daniel Gildea, Daniel Jurafsky. Proceedings of the 38th Annual Meeting of the Association for Computational Linguistics. 2000.
0
9
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
@antara_raaghavi is applying for PhDs this fall! She’s super impressive and awesome to work with, and conceived of this project independently and carried it out very successfully! Keep an eye out πŸ™‚.
0
0
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
So is it really this implicit operators thing that’s tripping them up? We try many other ablations, looking at the effect of giving extra context in the prompt, using numbers vs words, left-to-right ordering, and subtractive systems, and none of them seem to have that much effect.
1
0
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
Our experiments are based on Linguistics Olympiad problems that deal with number systems, like the one here. We created additional hand-standardized versions of each puzzle in order to be able to do all of the operator ablations.
Tweet media one
1
0
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
This shows the types of reasoning and variable binding jumps that are hard for LMs. It’s hard to go one level up, and bind a variable to have the meaning of an operator, or to understand that an operator is implicit.
1
0
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
If we alter the problems to make the operators explicit, the models can solve these problems pretty easily. But it’s still harder to bind a random symbol or word to mean an operator like +. It’s much easier when we use the familiar symbols for the operators, like + and x.
1
0
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
Our main finding: LMs find it hard when *operators* are implicit. We don’t say β€œ5 times 100 plus 20 plus 3”, we say β€œfive hundred and twenty-three”. The Linguistics Olympiad puzzles are pretty simple systems of equations that an LM should solve – but the operators aren’t explicit
Tweet media one
1
1
1
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
Why can’t LMs solve puzzles about the number systems of languages, when they can solve really complex math problems? Our paper led by @antara_raaghavi looks at why this intersection of language and math is difficult, and what this means for LM reasoning!
Tweet media one
2
1
18
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
2 months
Check out our ACL paper! We use shapley interactions to see which words (and phones) interact non-linearly -- what we lose when we assume linear relationships between features. Chat to Diganta in Vienna!.
@nsaphra
Naomi Saphra
2 months
ACL paper alert! What structure is lost when using linearizing attribution like Shapley? We show the nonlinear interactions between features reflect structures described by the sciences of syntax, semantics, and phonology.
Tweet media one
0
7
36
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
3 months
RT @SashaBoguraev: A key hypothesis in the history of linguistics is that different constructions share underlying structure. We take advan….
0
22
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
6 months
Man, those sparse autoencoder concepts sure can be unstable: you see one thing if you train it once and something totally different if you train it again. unless you use Archetypal SAEs!.
@Napoolar
Thomas Fel
6 months
Train your vision SAE on Monday, then again on Tuesday, and you'll find only about 30% of the learned concepts match. βš“ We propose Archetypal SAE which anchors concepts in the real data’s convex hull, delivering stable and consistent dictionaries.
Tweet media one
0
1
19
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
6 months
RT @iseeaswell: 😼SMOL DATA ALERT! 😼Anouncing SMOL, a professionally-translated dataset for 115 very low-resource languages! Paper: https://….
0
11
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
7 months
RT @QuantaMagazine: Computational linguists Julie Kallini (left) and Christopher Potts think that large language models might have somethin….
0
24
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
9 months
I will be at NeurIPS starting tomorrow! Would love to chat about interpretability, linguistics, language structure, meaning in LLMs. Reach out!. Aaaand if you love Vancouver, apply to do a PhD at UBC and work with me!
2
8
54
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
9 months
RT @jcrwhittington: πŸ”ŠPanel Announcement! Very excited to ask β€œIs this the end of Scale?” at the Global Summit in Boston on Dec 4: https://t….
0
3
0
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
11 months
I'm fascinated by the lack of determiner on (the?) Labour Conference, and why this sounds so strange to me! We do name-ize things like "at ICLR" with no determiner, but all the "at Labor Conference" kind of breaks my brain. Would someone who has nothing to do with Labour use it?.
0
1
9
@isabelpapad
Isabel Papadimitriou @ ICML πŸŒ„πŸŒ²πŸ‹
1 year
Isaac does some of the most impactful NLP work that I know! This is not just 'Google LM magic', it's the result of extremely hard-nosed and outside-the-box data work and linguistic work, as well as working with speakers. And the whole combo that makes Isaac Isaac!.
@iseeaswell
iseeaswell꩜bΚ‚ky
1 year
Excited to announce that 110 languages got added to Google Translate today! Time for context on these languages, especially the communities who helped a lot over the past few years, including Cantonese, NKo, and Faroese volunteers. Also, a 110-language youtube playlist. 🧡.
2
4
60