Andrei Mircea Profile
Andrei Mircea

@mirandrom

Followers
165
Following
189
Media
18
Statuses
100

PhD student @Mila_Quebec ⊗ mechanistic interpretability + systematic generalization + LLMs for science ⊗ https://t.co/xg8aE8CWM3

Montreal, QC
Joined December 2017
Don't wanna be here? Send us removal request.
@mirandrom
Andrei Mircea
25 days
Interested in LLM training dynamics and scaling laws? Come to our #ACL2025 oral tomorrow!.⏰ Tuesday 2:55pm.📍 Hall C (Language Modeling 1).🌐 . If you're in Vienna and want to chat, let me know!.@Mila_Quebec.
@mirandrom
Andrei Mircea
1 month
Step 1: Understand how scaling improves LLMs. Step 2: Directly target underlying mechanism. Step 3: Improve LLMs independent of scale. Profit. In our ACL 2025 paper we look at Step 1 in terms of training dynamics. Project: .Paper:
Tweet media one
0
6
12
@mirandrom
Andrei Mircea
17 hours
RT @ziling_cheng: Our paper on reasoning × interpretability × evaluation has been accepted to EMNLP main! . Excited because this marks the….
0
13
0
@mirandrom
Andrei Mircea
4 days
RT @darshilhdoshi1: Big news! Congratulations to the brilliant members of the collaboration!. Very excited to participate in the research t….
0
1
0
@mirandrom
Andrei Mircea
10 days
Outcome over process.
@hardeep_gambhir
Hardeep
10 days
correlation i have observed: . the most talented people i know (who also happen to be high agency) have the shittiest workflow organization. - no notion organization hell, usually messy apple notes and google docs.- regular users of pen and paper .- no superhuman or fancy email.
0
0
0
@mirandrom
Andrei Mircea
11 days
So yes, record all this data of scientists doing science, because it is 100% useful and important and under reported. But also try and replicate the science with the AI you want to train, even in an imperfect digital twin of a lab.
0
0
0
@mirandrom
Andrei Mircea
11 days
But at the end of the day, recording all that by itself would not be enough to train a language model that generates discourse, I think. The language model is trained on generating text. And in the same way I think AI scientists will need to be trained on doing science.
1
0
0
@mirandrom
Andrei Mircea
11 days
Just like science, that tacit knowledge is not recorded, and arguably would be useful in replicating the process with AI. What were the revisions and rounds of feedback leading to the final manuscript? What was someone hoping to achieve arguing online? Etc etc.
1
0
1
@mirandrom
Andrei Mircea
11 days
There's a lot of tacit knowledge that goes into discourse; whether it's writing an academic paper or arguing with someone online or whatever.
1
0
0
@mirandrom
Andrei Mircea
11 days
I love this idea; it's ambitious but grounded, and it breaks the mould of typical discourse on AI for science I've seen. But I'm not sure that recording without replicating is enough. Here's a rough analogy with language modeling I think clarifies my point 👇.
@Ben_Reinhardt
Ben Reinhardt
11 days
I do think AI could increase scientific productivity. But so much of science isn't easily mechanized tasks like pipetting -- it's adjusting optics, troubleshooting custom equipment, and keeping critters from dying. Here's a plan to capture and use this "tacit knowledge.".
2
0
1
@mirandrom
Andrei Mircea
11 days
RT @KateLobacheva: Join us at our lab’s symposium Aug 19–20! 🚀. I’ll present our recent ACL oral Training Dynamics Underlying Language Mode….
0
3
0
@mirandrom
Andrei Mircea
15 days
a lot of wisdom in *maybe* house cleaning.
@NC_Renic
Neil Renic
16 days
Ursula K. Le Guin had a dream writing schedule
Tweet media one
0
0
3
@mirandrom
Andrei Mircea
22 days
and now with LLMs we have the same problem for different reasons. I haven't seen LLMs that can consistently dive in and out of sources the way a human would.
0
0
0
@mirandrom
Andrei Mircea
22 days
I'm actually surprised how often I've seen this reviewing papers in ML. I guess the expectation has become that many reviewers will only do a surface level reading and accept/reject based on vibes.
@IsaacKing314
Isaac King 🔍
22 days
I've started checking sources for fun when I see a claim that seems dubious in something I'm reading, and probably more than half the time the source doesn't support the claim being made. A thread of some I saved:.
1
0
5
@mirandrom
Andrei Mircea
25 days
RT @nsaphra: If you’re in Vienna for ACL go check out our interpretability poster on how feature interactions reflect linguistic structure!….
0
6
0
@mirandrom
Andrei Mircea
25 days
RT @ziling_cheng: What do systematic hallucinations in LLMs tell us about their generalization abilities?. Come to our poster at #ACL2025 o….
0
7
0
@mirandrom
Andrei Mircea
25 days
RT @cesare_spinoso: How can we use models of cognition to help LLMs interpret figurative language (irony, hyperbole) in a more human-like m….
0
11
0
@mirandrom
Andrei Mircea
26 days
RT @ljyflores38: ⏰ Sharing our work on calibrated confidence scores at #ACL2025NLP, July 29 – 4PM Vienna time (Virtual)!. 📰 Improving the C….
0
9
0
@mirandrom
Andrei Mircea
27 days
RT @tongshuangwu: We all agree that AI models/agents should augment humans instead of replace us in many cases. But how do we pick when to….
0
21
0
@mirandrom
Andrei Mircea
29 days
RT @lasha_nlp: Life update: I’m excited to share that I’ll be starting as faculty at the Max Planck Institute for Software Systems(@mpi_sws….
0
43
0