Griffiths Computational Cognitive Science Lab Profile
Griffiths Computational Cognitive Science Lab

@cocosci_lab

Followers
6K
Following
98
Media
9
Statuses
181

Tom Griffiths' Computational Cognitive Science Lab. Studying the computational problems human minds have to solve.

Princeton, NJ
Joined August 2020
Don't wanna be here? Send us removal request.
@cocosci_lab
Griffiths Computational Cognitive Science Lab
11 months
(1/5) Very excited to announce the publication of Bayesian Models of Cognition: Reverse Engineering the Mind. More than a decade in the making, it's a big (600+ pages) beautiful book covering both the basics and recent work: https://t.co/5dnLpcMQzu
21
458
2K
@cocosci_lab
Griffiths Computational Cognitive Science Lab
4 days
Princeton also has a postdoc position associated with this multi-institution project focused on modeling human decision-making and human-AI trust. The link to apply for the Princeton position is:
@mark_ho_
Mark Ho
6 days
We’re hiring! @sucholutsky and I are seeking a postdoc and RA for a project on trust in AI systems with folks at NYU, Princeton, BU, and Cornell Positions open until filled. Apply soon! Please share 🔁 postdoc: https://t.co/iARVtYrMLN RA:
0
5
28
@cocosci_lab
Griffiths Computational Cognitive Science Lab
2 months
Paper now out in @CognitionJourn shows that people make suboptimal decisions about how to make investments to mitigate existential risks (as well as what the optimal decision rule looks like!)
@JQ_Zhu
Zhu Jian-Qiao
2 months
🚨🚨New paper in Cognition (with Adam Elga and @cocosci_lab) on how people assess existential risk. 🚨🚨 A thread🧵 (1/10)
0
4
29
@cocosci_lab
Griffiths Computational Cognitive Science Lab
2 months
Our new preprint explores how advances in AI change how we think about the role of symbols in human cognition. As neural networks show capabilities once used to argue for symbolic processes, we need to revisit how we can identify the level of analysis at which symbols are useful.
@RTomMcCoy
Tom McCoy
2 months
🤖🧠 NEW PAPER ON COGSCI & AI 🧠🤖 Recent neural networks capture properties long thought to require symbols: compositionality, productivity, rapid learning So what role should symbols play in theories of the mind? For our answer...read on! Paper: https://t.co/VsCLpsiFuU 1/n
3
3
57
@theryanliu
Ryan Liu @ COLM
3 months
A short 📹 explainer video on how LLMs can overthink in humanlike ways 😲! had a blast presenting this at #icml2025 🥳
6
15
64
@LanceYing42
Lance Ying
3 months
A hallmark of human intelligence is the capacity for rapid adaptation, solving new problems quickly under novel and unfamiliar conditions. How can we build machines to do so? In our new preprint, we propose that any general intelligence system must have an adaptive world model,
14
103
505
@kaiqu_liang
Kaiqu Liang
3 months
🤔 Feel like your AI is bullshitting you? It’s not just you. 🚨 We quantified machine bullshit 💩 Turns out, aligning LLMs to be "helpful" via human feedback actually teaches them to bullshit—and Chain-of-Thought reasoning just makes it worse! 🔥 Time to rethink AI alignment.
24
103
673
@Ikuperwajs
Ionatan Kuperwajs
3 months
New review on computational approaches to studying human planning out now in @TrendsCognSci! Really enjoyed having the opportunity to write something broader about the field with the help of @evanrussek @marcelomattar @weijima01 and @cocosci_lab
2
31
120
@JQ_Zhu
Zhu Jian-Qiao
4 months
Our paper is out today at @NatureHumBehav. We used machine learning to uncover what makes economic games complex for people.
Tweet card summary image
nature.com
Nature Human Behaviour - Zhu et al. use machine learning to reveal complex insights into human strategic decision-making.
1
13
41
@cocosci_lab
Griffiths Computational Cognitive Science Lab
5 months
Video games are a powerful tool for assessing the inductive biases of AI systems, as they are engineered based on how humans perceive the world and pursue their goals. This new benchmark evaluates the ability of vision language models using some challenging classic video games.
@a1zhang
Alex L Zhang
5 months
Can GPT, Claude, and Gemini play video games like Zelda, Civ, and Doom II? 𝗩𝗶𝗱𝗲𝗼𝗚𝗮𝗺𝗲𝗕𝗲𝗻𝗰𝗵 evaluates VLMs on Game Boy & MS-DOS games given only raw screen input, just like how a human would play. The best model (Gemini) completes just 0.48% of the benchmark! 🧵👇
0
4
24
@cocosci_lab
Griffiths Computational Cognitive Science Lab
5 months
In this new preprint we use methods from cognitive science to explore how well large language models make inferences from observations and construct interventions for understanding complex black-box systems that are analogous to those that scientists seek to understand
@JiayiiGeng
Jiayi Geng
5 months
Using LLMs to build AI scientists is all the rage now (e.g., Google’s AI co-scientist [1] and Sakana’s Fully Automated Scientist [2]), but how much do we understand about their core scientific abilities? We know how LLMs can be vastly useful (solving complex math problems) yet
0
7
58
@cocosci_lab
Griffiths Computational Cognitive Science Lab
5 months
New preprint shows that training large language models to produce better chains of thought for predicting human decisions also results in them producing better psychological explanations.
@JQ_Zhu
Zhu Jian-Qiao
5 months
1/14 Can we build an AI that thinks like psychologists or economists? 🤔Our new preprint shows how reinforcement learning (RL) can train LLMs to explain human decisions—not just predict them! That is, we're pushing LLMs beyond mere prediction into explainable cognitive models.
0
2
28
@cocosci_lab
Griffiths Computational Cognitive Science Lab
5 months
This paper uses metalearning to distill a Bayesian prior into a set of initial weights for a neural network, providing a way to create networks with interpretable soft inductive biases. The resulting networks can learn just as quickly as a Bayesian model when applied to new data.
@RTomMcCoy
Tom McCoy
5 months
🤖🧠Paper out in Nature Communications! 🧠🤖 Bayesian models can learn rapidly. Neural networks can handle messy, naturalistic data. How can we combine these strengths? Our answer: Use meta-learning to distill Bayesian priors into a neural network! https://t.co/vmOkilhMxJ 1/n
0
18
124
@alex_y_ku
Alexander Ku
5 months
(1/11) Evolutionary biology offers powerful lens into Transformers learning dynamics! Two learning modes in Transformers (in-weights & in-context) mirror adaptive strategies in evolution. Crucially, environmental predictability shapes both systems similarly.
12
26
167
@harootonian
Sev Harootonian
5 months
🚨 New preprint alert! 🚨 Thrilled to share new research on teaching! Work supervised by @cocosci_lab, @yael_niv, and @mark_ho_. This project asks: When do people teach by mentalizing vs with heuristics? 1/3 https://t.co/EnFTJrlOSz
1
4
17
@cocosci_lab
Griffiths Computational Cognitive Science Lab
5 months
New preprint! In-context and in-weights learning are two interacting forms of plasticity, like genetic evolution and phenotypic plasticity. We use ideas from evolutionary biology to predict when neural networks will use each kind of learning.
@alex_y_ku
Alexander Ku
5 months
(1/11) Evolutionary biology offers powerful lens into Transformers learning dynamics! Two learning modes in Transformers (in-weights & in-context) mirror adaptive strategies in evolution. Crucially, environmental predictability shapes both systems similarly.
0
8
75
@gianlucabencomo
Gianluca Bencomo
6 months
Every ChatGPT query costs more energy than the entire life of a fruit fly.
1
5
21
@VminVsky
Veniamin Veselovsky
6 months
New paper: Language models have “universal” concept representation – but can they capture cultural nuance? 🌏 If someone from Japan asks an LLM what color a pumpkin is, will it correctly say green (as they are in Japan)? Or does cultural nuance require more than just language?
6
34
132
@cocosci_lab
Griffiths Computational Cognitive Science Lab
7 months
We are looking for a new lab manager, shared with the Concepts and Cognition Lab of @TaniaLombrozo. Apply here:
1
21
27
@MaxDavidGupta1
Max David Gupta
7 months
Happy to share my first first-authored work at @cocosci_lab. Determining sameness or difference between objects is utterly trivial to humans, but surprisingly inaccessible to AI. Meta-learning can help neural networks overcome this barrier. Link: https://t.co/ID8DfXOImj (1/5)
2
15
102
@LanceYing42
Lance Ying
8 months
Many studies suggest AI has achieved human-like performance on various cognitive tasks. But what is “human-like” performance? Our new paper conducted a human re-labeling of several popular AI benchmarks and found widespread biases and flaws in task and label designs. We make 5
7
46
184