malkin1729 Profile
malkin1729

@FelineAutomaton

Followers
112
Following
17
Media
6
Statuses
27

Mathematician/informatician thinking probabilistically, expecting the same from you. ‘Tis categories in the mind and guns in their hands which keep us enslaved.

Edinburgh, Scotland
Joined September 2024
Don't wanna be here? Send us removal request.
@FelineAutomaton
malkin1729
5 hours
RT @nvimorozov: (1/n) The usual assumption in GFlowNet environments is acyclicity. Have you ever wondered if it can be relaxed? Does the ex….
0
4
0
@FelineAutomaton
malkin1729
5 days
RT @QueerinAI: 1/ 💻 Queer in AI is hosting a social at #ICML2025 in Vancouver on 📅 July 16, and you’re invited! Let’s network, enjoy food a….
0
2
0
@FelineAutomaton
malkin1729
8 days
Relative clause attachment ambiguity not intentional ;).
0
0
1
@FelineAutomaton
malkin1729
8 days
An oasis of inclusive science and solidarity amid the monotonically increasing NeurIPS madness that I'm proud to be supporting in a small role this year.
@QueerinAI
QueerInAI
8 days
🏳️‍🌈 Queer in AI is thrilled to announce another season of our affinity workshop at #NeurIPS2025! We announce a Call for Contributions to the workshop, with visa-friendly submissions due by 📅 July 31, 2025, all other submissions due by 📅 August 14, 2025. #QueerInAI #CallForPapers
Tweet media one
1
1
5
@FelineAutomaton
malkin1729
13 days
A great pleasure to crash two Bayesian statistics conferences with a dose of diffusion wisdom — last week in Singapore (, now in Cambridge ( — with the two authors of this very nice paper.
@Branchini_Nic
Nicola Branchini
2 months
🚨 New paper: “Towards Adaptive Self-Normalized IS”. TLDR; To estimate µ = E_p[f(θ)] when p(θ) has intractable partition, instead of doing MCMC on p(θ) or learning a parametric q(θ), we try MCMC directly on p(θ)| f(θ)-µ | - variance-minimizing proposal.
0
3
10
@FelineAutomaton
malkin1729
1 month
@siddarthv66 @mh_steps Paper here:
0
0
4
@FelineAutomaton
malkin1729
1 month
Great paper by @siddarthv66, @mh_steps, et al. on amortised inference in latent spaces of generative models, generalising our past work (. Useful for alignment, planning in latent space, inference in probabilistic programs?.
@siddarthv66
Siddarth Venkatraman
1 month
Is there a universal strategy to turn any generative model—GANs, VAEs, diffusion models, or flows—into a conditional sampler, or finetuned to optimize a reward function?.Yes! Outsourced Diffusion Sampling (ODS) accepted to @icmlconf , does exactly that!
Tweet media one
1
1
24
@FelineAutomaton
malkin1729
2 months
RT @josephdviviano: Ecstatic to show off some work my brilliant colleagues and I did at @iclr_conf this year! 🚀. We address the credit assi….
0
15
0
@FelineAutomaton
malkin1729
4 months
RT @sarthmit: 🚀 New Preprint! 🚀. In-Context Parametric Inference: Point or Distribution Estimators?. Thrilled to share our work on inferrin….
0
6
0
@FelineAutomaton
malkin1729
6 months
RT @TristanDeleu: My PhD thesis entitled "Generative Flow Networks: Theory and Applications to Structure Learning" is now available on Arxi….
0
77
0
@FelineAutomaton
malkin1729
6 months
RT @MarcinSendera: Happy to share one of my last works! If you are interested in diffusion samplers, please take a look🙃! Many thanks for a….
0
6
0
@FelineAutomaton
malkin1729
6 months
This delightful collaboration built upon my past work with @MarcinSendera @jarridrb ( and that of the brilliant @julberner and @lorenz_richter ( . Thanks to all! 9/9.
0
0
5
@FelineAutomaton
malkin1729
6 months
Have a look at our code here. Multiple objectives, exploration strategies, and time discretisation strategies are implemented in a common framework. 8/9.
1
0
5
@FelineAutomaton
malkin1729
6 months
We are eager to see extensions of this work to non-Markovian sequential generation, discrete state spaces, and posterior sampling under diffusion priors, as well as discretisation error and generalisation bounds. Numerical analysis and stochastic calculus are key tools here! 7/9.
1
0
5
@FelineAutomaton
malkin1729
6 months
The fact that these objectives are well-behaved asymptotically justifies the use of coarser (perhaps non-uniform) time discretisations during training than during sampling. This leads to greatly improved sample efficiency and even allows the use of time-local objectives. 6/9
Tweet media one
1
0
5
@FelineAutomaton
malkin1729
6 months
But all of these objectives run in a time discretisation. What happens when we take the step size to zero? For each objective, we prove that the limit is a well-understood continuous-time object. For example, the detailed balance condition approaches the Fokker-Planck PDE. 5/9
Tweet media one
Tweet media two
1
0
5
@FelineAutomaton
malkin1729
6 months
Such time-reversal can be enforced using a zoo of objectives computed through differentiable simulation (connected to stochastic control) or off-policy divergences (connected to entropy-regularised RL). We show connections among these objectives. 4/9
Tweet media one
1
0
5
@FelineAutomaton
malkin1729
6 months
Diffusion models are usually trained to maximise a variational bound on data likelihood, but they can also be fit without any data, by enforcing time-reversal of the generative and noising processes, where the latter has a target density as its boundary condition. 3/9
Tweet media one
Tweet media two
1
0
5
@FelineAutomaton
malkin1729
6 months
Inference, or sampling, in high-dimensional spaces is important in Bayesian statistics and in scientific applications (inverse problems, molecular dynamics, . ). Problem: How can we train diffusion models to amortise the cost of sampling a black-box target density? 2/9.
1
0
5
@FelineAutomaton
malkin1729
6 months
Happy to share our latest work on #diffusion models without data: building theoretical bridges between existing methods, analysing their continuous-time asymptotics, and showing some cool practical implications. #MachineLearning 1/9
Tweet media one
1
26
100