Sergey Pozdnyakov Profile
Sergey Pozdnyakov

@spozdn

Followers
358
Following
318
Media
8
Statuses
138

Postdoc @SchwallerGroup @EFFL | Geometric deep learning on 3D point clouds for atomistic modeling

Lausanne
Joined May 2023
Don't wanna be here? Send us removal request.
@JunwuChen25
Junwu Chen
9 days
Thrilled to share our new work MatInvent, a general and efficient reinforcement learning workflow that optimizes diffusion models for goal-directed crystal generation. Thanks to @JeffGuo__ , @efssh , @pschwllr , @SchwallerGroup , @NCCR_Catalysis . https://t.co/S7ibrRNdxp
Tweet card summary image
arxiv.org
Diffusion models promise to accelerate material design by directly generating novel structures with desired properties, but existing approaches typically require expensive and substantial labeled...
2
9
18
@spozdn
Sergey Pozdnyakov
1 month
See you there! 🤗
@HannesStaerk
Hannes Stärk
1 month
Reading group session on Monday: "Lookup multivariate Kolmogorov-Arnold Networks" https://t.co/W11eNHMlR9 with Sergey Pozdnyakov On zoom at 9am PT / 12pm ET / 6pm CE(S)T: https://t.co/Kew3F4DXag
0
0
4
@kabylda_
Adil Kabylda
1 month
Our new work, “QCell: Comprehensive Quantum-Mechanical Dataset Spanning Diverse Biomolecular Fragments,” is now out on arXiv! 🌱 1/7
1
1
6
@pschwllr
Philippe Schwaller (he/him)
1 month
Exciting postdoc opportunity in the @SchwallerGroup at EPFL! We're hiring a postdoc to advance ML-driven synthesis planning after Zlatko Joncev’s successful exit to co-found B-12 (YC '25) 🚀 Work on: - LLMs for strategic synthesis planning - Chemical reasoning at scale -
Tweet card summary image
arxiv.org
While automated chemical tools excel at specific tasks, they have struggled to capture the strategic thinking that characterizes expert chemical reasoning. Here we demonstrate that large language...
0
17
38
@XuanVuNguyen18
Xuan-Vu Nguyen
2 months
You don’t like molecular dynamics? We get it. That’s why at this year’s LLM hackathon for Chemistry and Materials Science, we built not one, but ✨two✨ AI agents for molecular dynamics 👇
1
7
17
@spozdn
Sergey Pozdnyakov
2 months
Shout-out to my supervisor @pschwllr, and huge thanks to all the members of the LIAC group @SchwallerGroup ♥️♥️♥️
1
0
6
@spozdn
Sergey Pozdnyakov
2 months
We did it for Convolutional Neural Networks, and it works too. lmKAN-based CNNs are Pareto optimal on both the CIFAR-10 and ImageNet-1k datasets, achieving 1.6-2.1Ă— reduction in inference FLOPs at matched accuracy.
1
0
4
@spozdn
Sergey Pozdnyakov
2 months
High-dimensional linear mappings are a primary component of most architectures, not just fully connected neural networks. So, other architectures can also be made lmKAN-based.
1
0
4
@spozdn
Sergey Pozdnyakov
2 months
It works! Our lookup multivariate Kolmogorov-Arnold Networks (lmKANs) are consistently inference FLOPs Pareto optimal. For general function approximation, the inference FLOPs reduction is up to 6.0Ă—. For the methane dataset, the H100 wall-clock time reduction is up to 12.9Ă—.
1
0
3
@spozdn
Sergey Pozdnyakov
2 months
Furthermore, we introduce a multivariate extension to KANs, which is built upon multivariate functions as building blocks. Our layers can have hundreds of times as many trainable parameters as a linear layer with the same shape. Yet the inference FLOPs are only twice as large.
1
0
4
@spozdn
Sergey Pozdnyakov
2 months
Kolmogorov-Arnold Networks are natural hosts for this idea as they express general high-dimensional mappings in terms of multiple optimizable 1D functions that can be splined. We implement CUDA kernels for this approach.
1
0
3
@spozdn
Sergey Pozdnyakov
2 months
Spline lookup tables allow us to do better than that. Let's take an optimizable piecewise linear function defined on a regular grid with G intervals. It contains G + 1 independent trainable parameters. While G can be arbitrarily large, the computational cost to evaluate it at any
2
0
5
@spozdn
Sergey Pozdnyakov
2 months
The number of trainable parameters is a primary predictor of a model's capabilities. For a linear layer with dimensions N_in and N_out, the number of parameters is ~N_in N_out, and so is the number of multiplications. One trainable parameter = one multiplication at inference.
1
0
3
@spozdn
Sergey Pozdnyakov
2 months
You can have a fully connected layer with hundreds of times as many parameters as a linear layer of the same shape, but only 2Ă— the inference FLOPs. CUDA kernels included. Paper: https://t.co/qrhuyCUg0n Code: https://t.co/Cn1GcWAJOR #lmKAN Main idea below.
Tweet card summary image
github.com
Contribute to schwallergroup/lmkan development by creating an account on GitHub.
1
0
4
@spozdn
Sergey Pozdnyakov
2 months
High-dimensional linear mappings, or linear layers, dominate both the parameter count and inference cost in most deep learning models. We propose a general-purpose drop-in replacement with a substantially better capacity - inference cost ratio. Check it out!đź§µ
1
10
34
@kabylda_
Adil Kabylda
2 months
Our recent work on SO3LR, a general-purpose machine learned force field for molecular simulations, has been published in @J_A_C_S! 🌞
6
24
167
@ylysogorskiy
Yury Lysogorskiy
3 months
Introducing GRACE foundational machine learning interatomic potentials – a new SOTA Pareto front for accuracy vs. performance, built for efficient MD simulations, versatile fine-tuning & distillation. Learn more: https://t.co/8xC2KKYq1q & https://t.co/S5e7OXZbJx #MaterialsScience
1
3
17
@martoskreto
Marta Skreta
4 months
AI4Mat is back for NeurIPS! time to crystallize those ideas and make a solid-state submission by august 22, 2025 đź’Ş new this year: opt-in your work for our Research Learning from Speaker Feedback program -- a new structured discussion format where spotlight presenters receive
0
20
83
@ycombinator
Y Combinator
4 months
b-12 ( https://t.co/sUMDUAK8TT) is building AI agents that help chemists design experiments and automatically run them on lab robots. They turn months of manual chemistry work into minutes of automated execution. Congrats on the launch @drecmb & @ZJoncev! https://t.co/waQzlnrS7z
20
27
129
@bwood_m
Brandon Wood
5 months
🚀Exciting news! We are releasing new UMA-1.1 models (Small and Medium) today and the UMA paper is now on arxiv! UMA represents a step-change in what’s possible with a single machine learning interatomic potential (short overview in the post below). The goal was to make a model
2
27
125