Charlotte Caucheteux @ICML24 Profile
Charlotte Caucheteux @ICML24

@c_caucheteux

Followers
2K
Following
227
Media
18
Statuses
47

Research Scientist @GoogleDeepMind | Deep Learning | Large Language Modelling | Cognitive Neuroscience

Paris, France
Joined June 2020
Don't wanna be here? Send us removal request.
@honualx
Alexandre Défossez
2 years
We release stereo models for all MusicGen variants (+ a new large melody both mono and stereo): 6 new models available on HuggingFace (thanks @reach_vb). We show how a simple fine tuning procedure with codebook interleaving takes us from boring mono to immersive stereo🎧👇
17
79
391
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
1
0
13
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Overall, these results strengthen the importance of distant and hierarchical predictions in natural language processing, and thus pave the way towards better algorithms inspired by the human brain.
4
1
21
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Similarly, we assess whether brain responses are best modeled by proximal or distant predictions. The results reveal a hierarchy of predictions in the 🧠: the fronto-parietal areas predict deeper & more distant representations than the superior temporal areas.
1
0
16
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
To clarify how these hierarchical predictions are organized in the brain, we manually enhanced language models with different types of predictions. For each brain region, we assess whether brain activity is best accounted for by shallow or deep predictions.
1
0
14
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Then, we further trained GPT-2 with two different objectives: 1. The classic next-word prediction loss 2. A hierarchical loss to predict latent and distant representations of the future. Our results show that the hierarchical model is more similar to the brain than the other.
1
2
18
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
To test this hypothesis, we first confirm that language models like GPT-2 build language representations partly similar to those of the brain, with the fMRI brain recordings of 345 subjects listening to stories.
1
3
17
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Deep Language Models are getting increasingly better by learning to predict the next word from its context. Is this really what the human brain does? Here, we hypothesize that our brain 🧠 rather makes distant and hierarchical predictions.
1
4
31
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Our paper is out in Nature Human Behaviour🔥🔥 ‘Evidence of a predictive coding hierarchy in the human brain listening to speech’ 📄 https://t.co/bkZ3AYMqDi 💡Unlike language models, our brain makes distant & hierarchical predictions with @agramfort and @JeanRemiKing Thread👇
11
195
739
@gizacard
Gautier Izacard
3 years
Happy to release a collection of LLaMA 🦙, large language models ranging from 7B to 65B parameters and trained on publicly available datasets. LLaMA-65B is competitive with Chinchilla and PaLM. Paper:
@GuillaumeLample
Guillaume Lample @ NeurIPS 2024
3 years
Today we release LLaMA, 4 foundation models ranging from 7B to 65B parameters. LLaMA-13B outperforms OPT and GPT-3 175B on most benchmarks. LLaMA-65B is competitive with Chinchilla 70B and PaLM 540B. The weights for all models are open and available at https://t.co/q51f2oPZlE 1/n
3
16
118
@GuillaumeLample
Guillaume Lample @ NeurIPS 2024
3 years
Today we release LLaMA, 4 foundation models ranging from 7B to 65B parameters. LLaMA-13B outperforms OPT and GPT-3 175B on most benchmarks. LLaMA-65B is competitive with Chinchilla 70B and PaLM 540B. The weights for all models are open and available at https://t.co/q51f2oPZlE 1/n
158
1K
7K
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
🙏 Thanks to @samnastase, @HassonUri, John Hale, @nilearn, @pyvista and the open-source and open-science communities for making this possible! 7/7
2
1
24
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
This is a joint work with our great team 🤩🤩 Juliette Millet, @PierreOrhan, Y Boubenec, @agramfort, E Dunbar, @chrplr and @JeanRemiKing, at @MetaAI, @ENS_ULM, @Inria & @Neurospin 6/n
1
2
17
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Conclusion: Modeling human-level intelligence is a far-off goal. Still, the emergence of brain-like functions in self-supervised algorithms suggests that we may be on the right path. 5/n
4
2
34
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Result 3: With an additional 386 subjects, we show that wav2vec 2.0 learns both the speech-specific and the language-specific representations of the prefrontal and temporal cortices, respectively. 4/n
1
2
19
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Result 2: The hierarchy learnt by the algorithm maps onto the brain's: The auditory cortex is best aligned with the first layer of the transformer (blue), whereas the prefrontal cortex is best aligned with its deepest layers (red). 3/n
1
3
27
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Result 1: self-supervised learning suffices to make this algorithm learn brain-like representations (i.e. most brain areas significantly correlate with its activations in response to the same speech input). 2/n
1
3
26
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Question: can a model trained on a *plausible* amount of *raw* speech explain both intelligent behavior and its brain bases? Here, we train wav2vec 2.0 w/ 600h of audio and map its activations onto the brains of 417 volunteers recorded with fMRI while listening to audio books.
1
1
27
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
🔥Our work has now been accepted to NeurIPS 2022 !! `Toward a realistic model of speech processing in the brain with self-supervised learning’: https://t.co/weiGlaiD65 Let’s meet in New Orleans on Tue 29 Nov 2:30pm PST (Hall J #524). A recap of the 3 main results below 👇
21
261
1K
@c_caucheteux
Charlotte Caucheteux @ICML24
3 years
Thanks to @samnastase and @HassonUri's lab for having publicly released their dataset, as well as @nilearn, @huggingface, @Inria, @ENS_ULM and @MetaAI for making this possible 🙏! 9/9
0
4
16