
Lorenz Richter
@lorenz_richter
Followers
345
Following
326
Media
12
Statuses
30
Postdoc at @ZuseInstitute, founder of @dida_ML
Joined February 2015
I'm especially looking forward to meeting my fantastic collaborators @julberner, @DenBless94, @junhua_c and of course many other friends.
0
1
3
On my way to #ICLR2025 in Singapore. We will present.β’ Sequential controlled Langevin diffusions, Thu, 10am,.β’ Underdamped diffusion bridges with applications to sampling, Sat, 10am + oral at FPI workshop, Mon, 10am. Let's meet!
3
22
145
Our new work extends the theory of diffusion bridges to degenerate noise settings, including underdamped Langevin dynamics (with @DenBless94, @julberner). This enables more efficient diffusion-based sampling with substantially fewer discretization steps.
1
31
142
My friend @daspantom is looking for interns working on generative modeling for proteins. Let him know in case you are interested.
We are hiring an intern on generative models for proteins for 6-12 months and typically results in publication. Find out what science is like in a bio-ml industry research lab. Bonus: float along the Rhine to work (perfectly normal commute here):
0
0
9
RT @papercopilot: Also on open/close peer review, I analyzed the AI/ML community's interest by analyzing user activity on open statistics aβ¦.
0
1
0
Our new work studies connections between discrete and continuous time diffusion samplers - after all, most things are very much related (GFlowNets, optimal control, path measures, PDEs). This allows us to reach faster convergence by randomized time steps.
Happy to share our latest work on #diffusion models without data: building theoretical bridges between existing methods, analysing their continuous-time asymptotics, and showing some cool practical implications. #MachineLearning 1/9
2
19
100
Check out our new paper on diffusion-based sampling, combining diffusion models with Sequential Monte Carlo in a principled way, It improves sampling quality and leads to more robust training. Thanks @junhua_c, @julberner, @DenBless94 for the great work!
3
5
100
Our work from last year on solving high-dimensional PDEs with tensor trains via robust BSDE-based methods has just appeared in @JmlrOrg. It's a follow-up of and suggests a new loss that is robust and explicit, i.e. fast.
'From continuous-time formulations to discretization schemes: tensor trains and robust regression for BSDEs and parabolic PDEs', by Lorenz Richter, Leon Sallandt, Nikolas NΓΌsken. . #iterative #tensor #numerically.
1
6
77
And back to Vienna again. I will present our work on time-continuous discrete-space diffusion models at @icmlconf (with @ludiXIVwinkler, Thu, 11:30). For the first time, we can connect those models to score-based generative modeling, see Ping me!
0
9
84
I gave a talk on our latest work on the connections between dynamical systems, PDEs, control and path space measures for sampling from densities at the @FieldsInstitute in Toronto last week (with @julberner, @Jeff_J_Sun). You can find the recording here:
0
26
108
I will give at talk about our latest work on PINN-based sampling from densities via SDEs and ODEs (with @julberner and @Jeff_J_Sun) at the Fields Institute in Toronto today. Check out the paper at
0
9
84
I gave a presentation on our latest diffusion-based sampling work virtually at @turinginst yesterday. The work has been mostly done jointly with @julberner and Nik NΓΌsken. You can find the recording here:
0
6
39
For the first time, we can directly link state-discrete continuous-time diffusion models to their time- and space-continuous (SDE-based) counterparts, i.e. score-based generative modeling. Credits go to the Ehrenfest process (+ @ludiXIVwinkler & Manfred):
0
2
18
Especially looking forward to meet my collaborators @julberner, @daspantom, @ludiXIVwinkler, @LVaitl, @guanhorng_liu and proud to represent @ZuseInstitute and @dida_ML.
0
0
9