scychan_brains Profile Banner
Stephanie Chan Profile
Stephanie Chan

@scychan_brains

Followers
5K
Following
5K
Media
33
Statuses
721

Staff Research Scientist at Google DeepMind. Artificial & biological brains 🤖 🧠 Views are my own.

San Francisco, CA
Joined November 2018
Don't wanna be here? Send us removal request.
@scychan_brains
Stephanie Chan
2 months
Check out our new work: Generalization from context often outperforms generalization from finetuning. And you might get the best of both worlds by spending extra compute at train-time.
@AndrewLampinen
Andrew Lampinen
2 months
How do language models generalize from information they learn in-context vs. via finetuning? We show that in-context learning can generalize more flexibly, illustrating key differences in the inductive biases of these modes of learning — and ways to improve finetuning. Thread: 1/
Tweet media one
4
20
206
@scychan_brains
Stephanie Chan
2 days
New ideas for our information ecosystem.
@bakkermichiel
Michiel Bakker
3 days
🚨🚨 Excited to share a new paper led by @Li_Haiwen_ with the @CommunityNotes team!. LLMs will reshape the information ecosystem. Community Notes offers a promising model for keeping human judgment central but it's an open question how to best integrate LLMs. Thread👇
Tweet media one
0
0
9
@scychan_brains
Stephanie Chan
2 days
An important line of research -- understanding complementarity between humans and AIs.
@shubadubadub
Rishub Jain
6 months
How do we ensure humans can still effectively oversee increasingly powerful AI systems? In our blog, we argue that achieving Human-AI complementarity is an underexplored yet vital piece of this puzzle! And, it’s hard, but we achieved it. 🧵(1/10)
Tweet media one
0
0
12
@scychan_brains
Stephanie Chan
11 days
RT @GoogleDeepMind: We’re bringing powerful AI directly onto robots with Gemini Robotics On-Device. 🤖. It’s our first vision-language-actio….
0
554
0
@scychan_brains
Stephanie Chan
15 days
I wonder if LLM sycophancy is actually also second-order. We know the responses are a reflection of us (writing style etc), and we like what we see, and that makes us feel good about ourselves too.
4
1
17
@scychan_brains
Stephanie Chan
17 days
RT @oswaldjoh: Super happy and proud to share our novel scalable RNN model - the MesaNet! . This work builds upon beautiful ideas of 𝗹𝗼𝗰𝗮𝗹𝗹….
0
64
0
@scychan_brains
Stephanie Chan
17 days
RT @coolboi95: I’m really excited to announce @GeneralistAI_ ! Our mission is to make general-purpose robots a reality. Getting to this “Ch….
0
13
0
@scychan_brains
Stephanie Chan
17 days
RT @emollick: McKinsey's new report on AI agents shows the same mindset I see in many firms: a focus on making small, obsolete models do ba….
0
231
0
@scychan_brains
Stephanie Chan
19 days
RT @cogscikid: Excited to share project specifying a research direction I think will be particularly fruitful for theory-driven cognitive s….
0
44
0
@scychan_brains
Stephanie Chan
24 days
RT @mpshanahan: Does It Make Sense to Speak of Introspection in Large Language Models? New paper with Iulia M. Comsa (@astronomind). https:….
0
14
0
@scychan_brains
Stephanie Chan
25 days
Very proud of @Aaditya6284 for garnering an ICML Oral Award for this work, even while moving countries and starting a new job!. The paper shows: When there are two different circuits to solve the same problem, the circuits can compete *and* cooperate at the same time. This.
@scychan_brains
Stephanie Chan
4 months
Dropping a few high-level takeaways in this thread. For more details please see Aaditya's thread,.or the paper itself.
0
2
31
@scychan_brains
Stephanie Chan
26 days
RT @METR_Evals: At METR, we’ve seen increasingly sophisticated examples of “reward hacking” on our tasks: models trying to subvert or explo….
0
46
0
@scychan_brains
Stephanie Chan
27 days
RT @EkdeepL: Paper alert—accepted as a NeurIPS *Spotlight*!🧵👇. We build on our past work relating emergence to task compositionality and an….
0
92
0
@scychan_brains
Stephanie Chan
29 days
RT @chengmyra1: Do people actually like human-like LLMs? In our #ACL2025 paper HumT DumT, we find a kind of uncanny valley effect: users di….
0
19
0
@scychan_brains
Stephanie Chan
29 days
I think it's important to remember that emergence is real, because it reminds us that model behaviors are not as predictable as we might like. NB: we can refer to emergence with model scale, or emergence with longer training. These are somewhat conflated, because larger models.
2
2
27
@scychan_brains
Stephanie Chan
29 days
shows theoretically how emergence occurs when learning sparse attention, and that these dynamics match real transformers.(@NicolasZucchet @dngfra).4/5.
1
2
30
@scychan_brains
Stephanie Chan
29 days
These results build on top of this analysis of S-curves in the context of multi-component tasks:.(@AdamSJermyn @bshlgrs).3/5.
1
1
22
@scychan_brains
Stephanie Chan
29 days
Emergence can occur due to multiple interacting subcircuits, even if each subcircuit is learned in a smooth gradual way. showed this in a mathematical model and with extensive experiments in small transformers (@Aaditya6284). 2/5.
1
3
44
@scychan_brains
Stephanie Chan
29 days
Emergence in transformers is a real phenomenon!. Behaviors and capabilities can appear in models in sudden ways. Emergence is not always just a "mirage". Compiling some examples here (please share any I missed): đź§µ.
12
42
358
@scychan_brains
Stephanie Chan
30 days
RT @Aaditya6284: Was super fun to be a part of this work! Felt very satisfying to bring the theory work on ICL with linear attention a bit….
0
5
0
@scychan_brains
Stephanie Chan
1 month
RT @jxmnop: new paper from our work at Meta!. **GPT-style language models memorize 3.6 bits per param**. we compute capacity by measuring t….
0
385
0