bayesgroup Profile Banner
Bayesian Methods Research Group Profile
Bayesian Methods Research Group

@bayesgroup

Followers
3K
Following
83
Media
78
Statuses
244

Research in Bayesian Deep Learning, Reinforcement Learning, Optimization, Structured Prediction, Drug Discovery and more

Joined July 2017
Don't wanna be here? Send us removal request.
@bayesgroup
Bayesian Methods Research Group
9 days
Not a mirage but our new paper!
@MNakhodnov
Maksim Nakhodnov
10 days
๐Ÿšจ New paper alert! ๐Ÿšจ Our new paper on ArXiv: "MiAD: Mirage Atom Diffusion for De Novo Crystal Generation". We unlock an ability to dynamically add or remove atoms during generation in diffusion models for better materials discovery. ๐Ÿ‘‰ https://t.co/9uWpxqcS7b 1/5
0
0
3
@bayesgroup
Bayesian Methods Research Group
21 days
New paper! #NeurIPS2025
@Viacheslav91112
Viacheslav Meshchaninov
22 days
๐Ÿš€COSMOS is OUT! @NeurIPSconf 2025! ๐Ÿ“ˆCOSMOSย achieves up toย 2xย faster text generation compared to other diffusion models, utilizing up toย 8x compression in text representations for superior efficiency. ๐Ÿ“„ย Paper:ย  https://t.co/UF4TNZpq4t ๐Ÿ’ปย Code:ย  https://t.co/zonTgWVXNF (1/6)
0
0
1
@dtiapkin
Daniil Tiapkin
30 days
While frontier labs are announcing their new models, we also want to be part of this parade. So, weโ€™re happy to announce gfnx โ€“ a JAX-first library with environments and a single-file baseline implementation for GFlowNet research.
3
10
20
@gritsaev
Timofei Gritsaev
2 months
1/ Can we efficiently learn the destruction process of diffusion samplers? Can we learn not just the drift, but also the variance for all transition kernels? โ€“ We answer YES in our recent paper โ€œAdaptive Destruction Processes for Diffusion Samplersโ€ (Oral at NeurIPS 2025 FPI
1
9
17
@bayesgroup
Bayesian Methods Research Group
6 months
Check out our new paper!
@nvimorozov
Nikita Morozov
6 months
(1/n) The usual assumption in GFlowNet environments is acyclicity. Have you ever wondered if it can be relaxed? Does the existing GFlowNet theory translate to the non-acyclic case? Is efficient training possible? We shed new light on these questions in our latest work! @icmlconf
0
0
3
@bayesgroup
Bayesian Methods Research Group
6 months
Check out our new work!
@MNakhodnov
Maksim Nakhodnov
6 months
๐Ÿšจ New paper alert! ๐Ÿšจ Our new paper on ArXiv: "DreamBooth DPO: Controlled Optimization of Personalized Diffusion Models" It addresses the core trade-off in personalized T2I: concept fidelity vs. prompt alignment, without any human-curated data ๐Ÿ‘‰ https://t.co/nT7nC7RtiM 1/5
0
0
2
@bayesgroup
Bayesian Methods Research Group
6 months
Check out our new work!
@MNakhodnov
Maksim Nakhodnov
6 months
๐Ÿšจ New paper alert! ๐Ÿšจ Our new paper on ArXiv: "ImageReFL: Balancing Quality and Diversity in Human-Aligned Diffusion Models". It tackles a key challenge in diffusion models: aligning with human preferences without collapsing diversity ๐Ÿ‘‰ https://t.co/gJaqMAi2b8 1/5
0
0
3
@bayesgroup
Bayesian Methods Research Group
9 months
Check our new paper! Catch the presentation at #ICLR2025 by @gritsaev!
@gritsaev
Timofei Gritsaev
10 months
1/ GFlowNets are known for training a forward policy to generate complex objects step by step. However, an equally important piece specific to the GFlowNet paradigm is a backward policy, which undoes these steps and plays a crucial role in training.
0
0
2
@GrigoryBartosh
Grigory Bartosh
1 year
๐Ÿ‘จโ€๐Ÿ’ผNeural Flow Diffusion Models at #NeurIPS2024 tomorrow! Discover how to build learnable noising processes for straight-line generative trajectories end-to-end and without simulations!๐Ÿคฏ ๐Ÿ“West Ballroom A-D #6809 โฐFri 13 Dec 4:30 pm โ€” 7:30 pm ๐Ÿ”— https://t.co/v0wW6VLKCL
@GrigoryBartosh
Grigory Bartosh
2 years
๐Ÿ”ฅ Excited to share our new work on Neural Flow Diffusion Models โ€” a general, end-to-end, simulation-free framework that works with an arbitrary noising processes and even enables learning them! ๐Ÿ“œ: https://t.co/nGbIIuEqzs ๐Ÿงต 1/11
1
12
73
@bayesgroup
Bayesian Methods Research Group
1 year
Check out our new paper! To be presented at #NeurIPS2024 by @KateLobacheva this Friday (poster #2408 / Poster Session 5 East / 13 Dec 11 am โ€“ 2 pm PST)
@KateLobacheva
Ekaterina Lobacheva
1 year
Starting training with a large learning rate benefits generalizationโ€”but why? In our new #NeurIPS2024 paper, we investigate its role in navigating the loss landscape and its effect on feature learning! 1/7 Paper: https://t.co/nzkkMk22hA Poster: https://t.co/aL7rqBjyq4
0
0
3
@bayesgroup
Bayesian Methods Research Group
1 year
๐—›๐—ฎ๐—ถ๐—ฟ๐—™๐—ฎ๐˜€๐˜๐—š๐—”๐—ก: ๐—ฅ๐—ฒ๐—ฎ๐—น๐—ถ๐˜€๐˜๐—ถ๐—ฐ ๐—ฎ๐—ป๐—ฑ ๐—ฅ๐—ผ๐—ฏ๐˜‚๐˜€๐˜ ๐—›๐—ฎ๐—ถ๐—ฟ ๐—ง๐—ฟ๐—ฎ๐—ป๐˜€๐—ณ๐—ฒ๐—ฟ ๐˜„๐—ถ๐˜๐—ต ๐—ฎ ๐—™๐—ฎ๐˜€๐˜ ๐—˜๐—ป๐—ฐ๐—ผ๐—ฑ๐—ฒ๐—ฟ-๐—•๐—ฎ๐˜€๐—ฒ๐—ฑ ๐—”๐—ฝ๐—ฝ๐—ฟ๐—ผ๐—ฎ๐—ฐ๐—ต by Maxim Nikolaev, Mikhail Kuznetsov, Dmitry Vetrov, @ai_alanov https://t.co/Uz4Jhb1hdL
0
0
2
@bayesgroup
Bayesian Methods Research Group
1 year
๐—ก๐—ฒ๐˜‚๐—ฟ๐—ฎ๐—น ๐—™๐—น๐—ผ๐˜„ ๐——๐—ถ๐—ณ๐—ณ๐˜‚๐˜€๐—ถ๐—ผ๐—ป ๐— ๐—ผ๐—ฑ๐—ฒ๐—น๐˜€: ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ฎ๐—ฏ๐—น๐—ฒ ๐—™๐—ผ๐—ฟ๐˜„๐—ฎ๐—ฟ๐—ฑ ๐—ฃ๐—ฟ๐—ผ๐—ฐ๐—ฒ๐˜€๐˜€ ๐—ณ๐—ผ๐—ฟ ๐—œ๐—บ๐—ฝ๐—ฟ๐—ผ๐˜ƒ๐—ฒ๐—ฑ ๐——๐—ถ๐—ณ๐—ณ๐˜‚๐˜€๐—ถ๐—ผ๐—ป ๐— ๐—ผ๐—ฑ๐—ฒ๐—น๐—น๐—ถ๐—ป๐—ด by Grigory Bartosh, Dmitry Vetrov, Christian A. Naesseth https://t.co/8J99VLhrdV
Tweet card summary image
arxiv.org
Conventional diffusion models typically relies on a fixed forward process, which implicitly defines complex marginal distributions over latent variables. This can often complicate the reverse...
1
1
1
@bayesgroup
Bayesian Methods Research Group
1 year
๐—š๐—ฟ๐—ผ๐˜‚๐—ฝ ๐—ฎ๐—ป๐—ฑ ๐—ฆ๐—ต๐˜‚๐—ณ๐—ณ๐—น๐—ฒ: ๐—˜๐—ณ๐—ณ๐—ถ๐—ฐ๐—ถ๐—ฒ๐—ป๐˜ ๐—ฆ๐˜๐—ฟ๐˜‚๐—ฐ๐˜๐˜‚๐—ฟ๐—ฒ๐—ฑ ๐—ข๐—ฟ๐˜๐—ต๐—ผ๐—ด๐—ผ๐—ป๐—ฎ๐—น ๐—ฃ๐—ฎ๐—ฟ๐—ฎ๐—บ๐—ฒ๐˜๐—ฟ๐—ถ๐˜‡๐—ฎ๐˜๐—ถ๐—ผ๐—ป by Mikhail Gorbunov, Nikolay Yudin, Vera Soboleva, @ai_alanov, Alexey Naumov, Maxim Rakhuba https://t.co/Mr4PHYIFf0
Tweet card summary image
arxiv.org
The increasing size of neural networks has led to a growing demand for methods of efficient fine-tuning. Recently, an orthogonal fine-tuning paradigm was introduced that uses orthogonal matrices...
1
0
1
@bayesgroup
Bayesian Methods Research Group
1 year
๐—ช๐—ต๐—ฒ๐—ฟ๐—ฒ ๐——๐—ผ ๐—Ÿ๐—ฎ๐—ฟ๐—ด๐—ฒ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐—ฅ๐—ฎ๐˜๐—ฒ๐˜€ ๐—Ÿ๐—ฒ๐—ฎ๐—ฑ ๐—จ๐˜€? ๐—” ๐—™๐—ฒ๐—ฎ๐˜๐˜‚๐—ฟ๐—ฒ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐—ฃ๐—ฒ๐—ฟ๐˜€๐—ฝ๐—ฒ๐—ฐ๐˜๐—ถ๐˜ƒ๐—ฒ by @irsadrtdinov, Maxim Kodryan, Eduard Pokonechny, @KateLobacheva, Dmitry Vetrov (stay tuned for the full paper, previously: https://t.co/Iu3EsHJT8z)
openreview.net
It is a conventional wisdom that using large learning rates (LRs) early in training improves generalization. Following a line of research devoted to understanding this effect mechanistically, we...
1
0
0
@bayesgroup
Bayesian Methods Research Group
1 year
We got 4 papers accepted to #NeurIPS2024!
1
0
4
@KateLobacheva
Ekaterina Lobacheva
1 year
Did you know that networks trained with different learning rates extract different features (and a different number of them!) from the data? Come by our poster at HiLD Workshop #ICML2024 tomorrow to discuss it with @irsadrtdinov! Paper: https://t.co/AHWFaK5wog 1/6
openreview.net
It is a conventional wisdom that using large learning rates (LRs) early in training improves generalization. Following a line of research devoted to understanding this effect mechanistically, we...
3
9
45
@irsadrtdinov
Ildus Sadrtdinov
2 years
I will be presenting our NeurIPS-2023 paper https://t.co/rTUdEZcin3 at @ml_collective this Friday, March 8, 10am PT / 7pm CET! If you haven't decided yet whether to stay in the pre-train basin or not, you definitely need to see this talk!
1
1
11
@bayesgroup
Bayesian Methods Research Group
2 years
Check out our new paper!
@dtiapkin
Daniil Tiapkin
2 years
๐ŸŒŸ News from the GFlowNet world: our paper โ€œGenerative Flow Networks as Entropy-Regularized RLโ€ was honored with oral presentation at #AISTATS2024! Long story short, our result can be described by this picture.
0
0
3
@bayesgroup
Bayesian Methods Research Group
2 years
At #NeurIPS2023? Come check out our latest work!
@KateLobacheva
Ekaterina Lobacheva
2 years
Large learning rates improve generalization, but are they all beneficial? The short answer is No, for more details check out our paper at the #NeurIPS2023 Mathematics of Modern Machine Learning (M3L) Workshop! Paper: https://t.co/517xCsfrWA 1/4
0
0
1
@KateLobacheva
Ekaterina Lobacheva
2 years
Can we improve ensembles in the transfer learning setup by exploring the target task loss landscape? Find out in our new #NeurIPS2023 paper! Joint work with Ildus Sadrtdinov, Dmitrii Pozdeev, and Dmitry Vetrov. Paper: https://t.co/y14hXqdIaa 1/7
1
12
57