matt_niedoba Profile Banner
Matt Niedoba Profile
Matt Niedoba

@matt_niedoba

Followers
21
Following
1K
Media
7
Statuses
14

Computer Science PhD Student at UBC

Joined January 2013
Don't wanna be here? Send us removal request.
@matt_niedoba
Matt Niedoba
2 months
Proud to share that my paper w/ @frankdonaldwood & @sirbayes is a #ICML2025 spotlight!. TLDR: We find a simple patch-based mechanism that closely approximates how image diffusion models generalize, regardless of architecture.
Tweet media one
5
3
33
@matt_niedoba
Matt Niedoba
1 month
I'll be presenting this today in the evening poster session from 4:30pm to 7. Come say hi if you want to discuss diffusion generalization!.
@matt_niedoba
Matt Niedoba
2 months
Proud to share that my paper w/ @frankdonaldwood & @sirbayes is a #ICML2025 spotlight!. TLDR: We find a simple patch-based mechanism that closely approximates how image diffusion models generalize, regardless of architecture.
Tweet media one
0
0
0
@matt_niedoba
Matt Niedoba
2 months
We’re excited to keep working to improve PSPC and close the gap to network denoisers! We hope PSPC will lay the groundwork for a better understanding of how diffusion models generalize, and allow us to attribute which training samples are responsible for each generated image 11/.
0
0
3
@matt_niedoba
Matt Niedoba
2 months
However, there is still work to do! Although PSPC samples have some similarity to network samples, error accumulation over the diffusion process results in artifacts in the final samples. 10/
Tweet media one
1
0
3
@matt_niedoba
Matt Niedoba
2 months
Comparing PSPC to network denoiser outputs, we find that it closely approximates network outputs across all diffusion times and datasets - suggesting that local denoising comprises a key component of diffusion generalization 9/
Tweet media one
1
0
3
@matt_niedoba
Matt Niedoba
2 months
Can we combine patch-based denoisers to reproduce diffusion model generalization? We propose compositing patch posterior means with a simple mechanism which we call Patch Set Posterior Composite (PSPC) 8/
Tweet media one
1
0
1
@matt_niedoba
Matt Niedoba
2 months
If network denoisers use local operations during training, we would expect the same mechanism to apply for sampling. Indeed, we find that patch posterior means are good estimators of network output patches when run on intermediate noisy images drawn from the sampling process. 7/.
1
0
2
@matt_niedoba
Matt Niedoba
2 months
Why would diffusion models use local operations? We find that for most of the training distribution, patch posterior means are equivalent to optimal denoiser patches. For much of training, a local estimator is sufficient to estimate the minima of the training objective! (left) 6/
Tweet media one
1
0
1
@matt_niedoba
Matt Niedoba
2 months
We hypothesize that these gradients are the result of network denoisers employing local denoising operations. We approximate these operations with a patch posterior means - a weighted average of the same patch of each training set image 5/.
1
0
3
@matt_niedoba
Matt Niedoba
2 months
Looking at the gradients of neural-network denoisers suggests a local inductive bias. We find networks are more sensitive to pixels within a local patch around the output pixel. The size of this patch increases with diffusion time. 4/
Tweet media one
1
1
2
@matt_niedoba
Matt Niedoba
2 months
Our work analyzes these approximation errors. We find that in both U-Net and DiTs, neural network denoisers make similar errors in both magnitude and quality. This suggests a shared inductive bias common to all image diffusion models 3/
Tweet media one
1
0
2
@matt_niedoba
Matt Niedoba
2 months
However, this optimal denoiser has a problem - it can’t generalize. Sampling with the optimal denoiser generates training set copies. The generalization behaviour of diffusion models is therefore the result of neural network approximation errors 2/.
1
0
1
@matt_niedoba
Matt Niedoba
2 months
Diffusion models produce samples through repeated denoising operations - they estimate clean images from ones which has been corrupted by Gaussian noise. Notably, the optimal denoising operation has a simple closed form. It is a weighted average of training set images. 1/.
1
0
2