Shreyas Padhy Profile
Shreyas Padhy

@shreyaspadhy

Followers
374
Following
1K
Media
14
Statuses
188

PhD student at the University of Cambridge. Ex @GoogleAI Resident, @jhubme and @iitdelhi. I like the math of machine learning & neuroscience. Also DnD.

Cambridge, England
Joined July 2011
Don't wanna be here? Send us removal request.
@shreyaspadhy
Shreyas Padhy
7 months
I will be at my first in-person NeurIPS, presenting 3 posters at the main conference (🧵)! Please get in touch to chat about - . - Diffusion models, sampling and conditional generation.- BayesOPT, GPs and BNNs. P.S. I'll be on the job market early next year, please reach out!.
2
0
15
@shreyaspadhy
Shreyas Padhy
5 months
Checkout this paper with some really interesting insights led by the excellent @JiajunHe614 and @YuanqiD - . TLDR: Neural density samplers really need guidance imposed through Langevin annealing to make them work well.
@YuanqiD
Yuanqi Du
5 months
Working on sampling and seeking neural network ansatz? Longing for simulation-free* training approaches? – we review neural samplers and present a “failed” attempt towards it with pitfalls and promises!. Joint work with @JiajunHe614 (co-lead), Francisco Vargas …. 🧵1/n.
0
2
12
@shreyaspadhy
Shreyas Padhy
5 months
Thanks for the kind words @ArnaudDoucet1 ! I wanted to shout-out some other great work in the same vein as us - . (@julberner, @lorenz_richter, @MarcinSendera et al).(@msalbergo et al).(@junhua_c et al).
@ArnaudDoucet1
Arnaud Doucet
5 months
Tweeting again about sampling, my favourite 2024 Monte Carlo paper is by F. Vargas, @shreyaspadhy, D. Blessing & N. Nüsken: . Propose a "simple" loss to learn the drift you need to add to Langevin to follow a fixed probability path.
2
5
44
@shreyaspadhy
Shreyas Padhy
5 months
RT @ArnaudDoucet1: Tweeting again about sampling, my favourite 2024 Monte Carlo paper is by F. Vargas, @shreyaspadh….
0
38
0
@shreyaspadhy
Shreyas Padhy
7 months
Come chat with folks from our group!.
@CambridgeMLG
Cambridge MLG
7 months
We're excited to be at #NeurIPS Vancouver!. See the papers we'll be presenting at the main conference below:
Tweet media one
0
0
3
@shreyaspadhy
Shreyas Padhy
7 months
RT @CambridgeMLG: We're excited to be at #NeurIPS Vancouver!. See the papers we'll be presenting at the main conference below: https://t.co….
0
2
0
@shreyaspadhy
Shreyas Padhy
7 months
RT @AtinaryTech: Atinary @ #NeurIPS in Vancouver this week🍁 .Connect with our #AI #ML researchers @VictorSabanza & @shreyaspadhy. Our rese….
0
3
0
@shreyaspadhy
Shreyas Padhy
7 months
Finally, we're presenting a new symmetry-aware generative model that discovers which (approximate) symmetries exist in data for improved data efficiency. Catch co-author @JamesAllingham on Friday, Dec 13, 06:00 at Poster Session 6 East, #2500.
@JamesAllingham
James Allingham
7 months
I'll be at NeurIPS next week, presenting our work "A Generative Model of Symmetry Transformations." In it, we propose a symmetry-aware generative model that discovers which (approximate) symmetries are present in a dataset, and can be leveraged to improve data efficiency. 🧵⬇️
Tweet media one
0
0
3
@shreyaspadhy
Shreyas Padhy
7 months
Second, we're presenting our work on speeding up marginal likelihood estimation in GPs by up to 72x without sacrificing performance! Catch co-author @JihaoAndreasLin on Thursday, Dec 12, 16:30 at Poster Session 4 East, #3910.
@JihaoAndreasLin
Jihao Andreas Lin
1 year
"Improving Linear System Solvers for Hyperparameter Optimisation in Iterative Gaussian Processes". Three techniques to accelerate marginal likelihood training in GPs by up to 72x without sacrificing performance!. Check out our paper here: . (1/6).
1
0
3
@shreyaspadhy
Shreyas Padhy
7 months
First, we're presenting our paper on efficient fine-tuning of diffusion models for conditional generation. Find us (w/ @DenkerAlexander, Francisco Vargas) on Thursday, Dec 12, 11:00 at Poster Session 3 East, #2500.
@shreyaspadhy
Shreyas Padhy
7 months
I’m really excited to be attending NeurIPS and presenting our work on efficient fine-tuning of pre-trained diffusion models for SOTA conditional generation. Come chat with us on 12th Dec (Thursday) at 11am! . Thread below (🧵) -
Tweet media one
1
1
3
@shreyaspadhy
Shreyas Padhy
7 months
RT @JamesAllingham: I'll be at NeurIPS next week, presenting our work "A Generative Model of Symmetry Transformations." In it, we propose a….
0
27
0
@shreyaspadhy
Shreyas Padhy
7 months
We’re looking forward to chatting with folks at NeurIPS this year about this work! Work done with excellent collaborators @DenkerAlexander, Francisco Vargas, @DidiKieran, @SimMat20, @vdutor, @BarbanoRiccardo, @MathieuEmile, @julia_tweeting_, and @pl219_Cambridge!.
0
2
5
@shreyaspadhy
Shreyas Padhy
7 months
In conclusion, DEFT is a fine-tuning framework which.(1) leverages pre-trained models, reducing training time and computational resources, and .(2) is applicable to various conditional generation tasks, both linear and non-linear. (11/11).
1
0
4
@shreyaspadhy
Shreyas Padhy
7 months
Further, we apply DEFT to the protein design task of motif scaffolding. Here, we obtain good results with a h-transform model only 4% to 9% of the size of the original model, drastically reducing the training time. (10/11)
Tweet media one
1
0
3
@shreyaspadhy
Shreyas Padhy
7 months
We show promising results both on linear imaging tasks (inpainting, deblurring, super-resolution and computed tomography) as well as non-linear imaging, including non-linear deblurring and phase retrieval. (9/11)
Tweet media one
1
0
3
@shreyaspadhy
Shreyas Padhy
7 months
Naively conditioning SDEs can lead to a bias due to choosing the incorrect initial distribution. We show how to address this for the VP-SDE and connect it to recent work (Domingo-Enrich et al., 2024) that links this bias to the value function from stochastic control. (8/11).
1
0
4
@shreyaspadhy
Shreyas Padhy
7 months
Further, the connection to prior work on stochastic optimal control motivates us to use similar network architectures and encode the likelihood term as an inductive bias. This parameterization allows for a fast training of the model! (7/11)
Tweet media one
1
0
3
@shreyaspadhy
Shreyas Padhy
7 months
We demonstrate that DEFT can also be interpreted as a stochastic optimal control problem, enabling the use of variational inference techniques for learning the h-transform with only a single noisy observation. (6/11).
1
0
3
@shreyaspadhy
Shreyas Padhy
7 months
The Offline Score-Matching loss requires a small dataset for fine-tuning. For Image Inpainting on ImageNet we obtain SOTA results with only 200 additional training images. (5/11)
Tweet media one
1
0
2
@shreyaspadhy
Shreyas Padhy
7 months
In particular, we propose two loss functions: Offline Score-Matching and Online Stochastic Control. Both loss functions only require access to the forward pass of the pre-trained model and are agnostic to the underlying network architecture. (4/11)
Tweet media one
1
0
3