FlorentinGuth Profile Banner
Florentin Guth Profile
Florentin Guth

@FlorentinGuth

Followers
502
Following
174
Media
13
Statuses
96

Postdoc at @NYUDataScience and @FlatironCCN. Wants to understand why deep learning works.

New York
Joined March 2014
Don't wanna be here? Send us removal request.
@FlorentinGuth
Florentin Guth
1 month
RT @fedichev: As you know I'm obsessed with power laws in biology, which is a biological consequence of fundamental principles, like energy….
0
129
0
@FlorentinGuth
Florentin Guth
1 month
RT @unireps: 🔥 Mark your calendars for the next session of the @ELLISforEurope x UniReps Speaker Series! . 🗓️ When: 31th July – 16:00 CEST….
0
8
0
@FlorentinGuth
Florentin Guth
2 months
RT @rdMorel: For evolving unknown PDEs, ML models are trained on next-state prediction. But do they actually learn the time dynamics: the "….
0
50
0
@FlorentinGuth
Florentin Guth
2 months
RT @unireps: 🎥 The recording of the third ELLISxUniReps Speaker Series session with @victorveitch and @luigigres .is now available at: http….
0
11
0
@FlorentinGuth
Florentin Guth
2 months
RT @beenwrekt: The NeurIPS paper checklist corroborates the bureaucratic theory of statistics.
Tweet card summary image
argmin.net
The NeurIPS checklist corroborates the bureaucratic theory of statistics.
0
28
0
@FlorentinGuth
Florentin Guth
3 months
For a more in-depth discussion of the approach and results (and more!):
0
1
19
@FlorentinGuth
Florentin Guth
3 months
Finally, we test the manifold hypothesis: what is the local dimensionality around an image? We find that this depends both on the image and the size of the local neighborhood, and there exists images with both large full-dimensional and small low-dimensional neighborhoods.
Tweet media one
3
1
11
@FlorentinGuth
Florentin Guth
3 months
High probability ≠ typicality: very high-probability images are rare. This is not a contradiction: frequency = probability density *multiplied by volume*, and volume is weird in high dimensions! Also, the log probabilities are Gumbel-distributed, and we don't know why!
Tweet media one
2
1
17
@FlorentinGuth
Florentin Guth
3 months
These are the highest and lowest probability images in ImageNet64. An interpretation is that -log2 p(x) is the size in bits of the optimal compression of x: higher probability images are more compressible. Also, the probability ratio between these is 10^14,000! 🤯
Tweet media one
2
2
18
@FlorentinGuth
Florentin Guth
3 months
But how do we know our probability model is accurate on real data?.In addition to computing cross-entropy/NLL, we show *strong* generalization: models trained on *disjoint* subsets of the data predict the *same* probabilities if the training set is large enough!
Tweet media one
1
1
7
@FlorentinGuth
Florentin Guth
3 months
We call this approach "dual score matching". The time derivative constrains the learned energy to satisfy the diffusion equation, which enables recovery of accurate and *normalized* log probability values, even in high-dimensional multimodal distributions.
Tweet media one
1
1
8
@FlorentinGuth
Florentin Guth
3 months
We also propose a simple procedure to obtain good network architectures for the energy U: choose any pre-existing score network s and simply take the inner product with the input image y! We show that this preserves the inductive biases of the base score network: grad_y U ≈ s.
Tweet media one
1
1
11
@FlorentinGuth
Florentin Guth
3 months
How do we train an energy model?.Inspired by diffusion models, we learn the energy of both clean and noisy images along a diffusion. It is optimized via a sum of two score matching objectives, which constrain its derivatives with both the image (space) and the noise level (time).
Tweet media one
1
1
13
@FlorentinGuth
Florentin Guth
3 months
What is the probability of an image? What do the highest and lowest probability images look like? Do natural images lie on a low-dimensional manifold?.In a new preprint with @ZKadkhodaie @EeroSimoncelli, we develop a novel energy-based model in order to answer these questions: 🧵
Tweet media one
11
73
366
@FlorentinGuth
Florentin Guth
3 months
RT @EeroSimoncelli: @jpillowtime Tweedie’s formula wasn’t published by Tweedie (AFAIK). It was published by Miyasawa in 1961:. Miyasawa, K….
0
2
0
@FlorentinGuth
Florentin Guth
4 months
🌈 I'll be presenting our JMLR paper "A rainbow in deep network black boxes" today at 3pm at #ICLR25! Come to poster #334 if you're interested, I'll be happy to chat.More details in the quoted thread (two levels deep).
@FlorentinGuth
Florentin Guth
10 months
I'm delighted to announce that the rainbow 🌈 paper has been accepted at JMLR!.➡️ updated paper with brand new intro: We released code, along with a self-contained tutorial to reproduce our results in a simple setting: More below ⬇️
Tweet media one
0
2
13
@FlorentinGuth
Florentin Guth
7 months
RT @docmilanfar: When x and y are independent random variables, their joint cumulative dist function (CDF) is the product of the individual….
0
23
0
@FlorentinGuth
Florentin Guth
8 months
RT @EeroSimoncelli: Graduate students and advanced undergraduates: .Interested in a 3-month summer research internship in Computational Neu….
0
43
0
@FlorentinGuth
Florentin Guth
9 months
RT @scifordl: Poster session still going on in West meeting rooms 205-207! Come say hi and check your favorite posters (also there is deca….
0
1
0
@FlorentinGuth
Florentin Guth
9 months
RT @scifordl: Can LLMs plan? The answer might be in @HanieSedghi's talk. Only one way to know, join us in West meeting rooms 205-207 https:….
0
3
0