Stanislav Frolov Profile
Stanislav Frolov

@stfrolov

Followers
211
Following
983
Media
26
Statuses
367

Researcher @DFKI Generative Image Modeling | Intern @MetaAI '22 & @AdobeResearch '21

Kaiserslautern, Germany
Joined November 2012
Don't wanna be here? Send us removal request.
@stfrolov
Stanislav Frolov
1 year
I am happy to share that SpotDiffusion was accepted to WACV 2025. Page: https://t.co/YkCPNa3Px0 Code: https://t.co/AJYj82peIO Paper: https://t.co/9yd2iq1AzQ SpotDiffusion is an efficient method for seamless panorama generation from text. 🧵
Tweet card summary image
github.com
Official Pytorch Implementation for "SpotDiffusion: A Fast Approach For Seamless Panorama Generation Over Time" (WACV 2025) https://spotdiffusion.github.io/ - stanifrolov/spotdiffusion
1
1
3
@BaldassarreFe
Federico Baldassarre
5 months
Say hello to DINOv3 🦖🦖🦖 A major release that raises the bar of self-supervised vision foundation models. With stunning high-resolution dense features, it’s a game-changer for vision tasks! We scaled model size and training data, but here's what makes it special 👇
40
260
2K
@stfrolov
Stanislav Frolov
9 months
Happy to share that TKG-DM, a training-free chroma key content generation diffusion model was accepted to CVPR 25. Project led by @Oguryu417 Paper: https://t.co/BpV0xFmxrP Code: https://t.co/EohvQ4qzzO
0
5
5
@stfrolov
Stanislav Frolov
9 months
Checkout PromptMap, presented at IUI'25, a new interaction style with text-to-image models/data that allows users to freely explore a vast collection of synthetic prompts through a map-like view with semantic zoom. Paper: https://t.co/eYJ6UBHOGK Code: https://t.co/BRHQxjwjTV
0
1
1
@Michael_J_Black
Michael Black
1 year
I received feedback that my post about reviews not being "random" caused stress for some students. I'm sorry for that. It was meant to be empowering. Personally, I find the idea that I don't have some control over the destiny of my papers to be disheartening. If the process is
Tweet card summary image
perceiving-systems.blog
5
11
180
@stfrolov
Stanislav Frolov
1 year
We propose a time-dependent, attention-guided masking approach that prioritizes high-attention regions first, gradually refining the entire image. This improves quality across various models. Paper: https://t.co/XkTqtW3pkh Thanks to @LuckyOwl95 @rave78 @spalaciob @DFKI
0
1
1
@stfrolov
Stanislav Frolov
1 year
We find that important image pixels, as measured by the attention values of DINO, are more challenging to learn (higher reconstruction error).
1
0
1
@stfrolov
Stanislav Frolov
1 year
Dynamic attention-guided diffusion accepted to #WACV2025 🎉 We challenge the common SR diffusion approach: must the entire image be updated at each step? Some regions, like faces, may need more focus than plain backgrounds. 🧵
1
0
0
@stfrolov
Stanislav Frolov
1 year
We can produce seamless panoramas much faster by leveraging the iterative nature of diffusion models and shifting non-overlapping denoising windows over time.
1
0
1
@stfrolov
Stanislav Frolov
1 year
To generate images beyond the training resolution, MultiDiffusion averages overlapping denoising windows. While this works, it can be slow because large overlap between the windows is required.
1
0
1
@stanislavfort
Stanislav Fort
1 year
✨🎨🏰Super excited to share our new paper Ensemble everything everywhere: Multi-scale aggregation for adversarial robustness Inspired by biology we 1) get adversarial robustness + interpretability for free, 2) turn classifiers into generators & 3) design attacks on vLLMs 1/12
24
199
1K
@stfrolov
Stanislav Frolov
2 years
I can’t find a recent paper (and tweet) that had emojis all over an image. I think it was a method about interpreting (possibly segmenting) images with/from diffusion models. Can somebody help?
0
0
0
@stfrolov
Stanislav Frolov
2 years
Wow that’s cool! LoRA but for training.
@tydsh
Yuandong Tian
2 years
Thanks @_akhaliq for promoting our work! With GaLore, now it is possible to pre-train a 7B model in NVidia RTX 4090s with 24G memory! 🤔How? Instead of assuming low-rank weight structure like LoRA, we show that the weight gradient is naturally low-rank and thus can be
0
0
0
@tydsh
Yuandong Tian
2 years
Thanks @_akhaliq for promoting our work! With GaLore, now it is possible to pre-train a 7B model in NVidia RTX 4090s with 24G memory! 🤔How? Instead of assuming low-rank weight structure like LoRA, we show that the weight gradient is naturally low-rank and thus can be
@_akhaliq
AK
2 years
GaLore Memory-Efficient LLM Training by Gradient Low-Rank Projection Training Large Language Models (LLMs) presents significant memory challenges, predominantly due to the growing size of weights and optimizer states. Common memory-reduction approaches, such as low-rank
21
80
426
@CSProfKGD
Kosta Derpanis
2 years
“Write me a scientific review in the voice of Dr. Seuss and as reviewer 2, the negative reviewer who clearly doesn’t understand the paper and has probably not read the paper. Mention that there is no novelty and that the contribution is limited.” GPT-4: Oh, I've read your work,
2
36
174
@TimDarcet
TimDarcet
2 years
Vision transformers need registers! Or at least, it seems they 𝘸𝘢𝘯𝘵 some… ViTs have artifacts in attention maps. It’s due to the model using these patches as “registers”. Just add new tokens (“[reg]”): - no artifacts - interpretable attention maps 🦖 - improved performances!
40
310
2K
@borisdayma
Boris Dayma 🖍️
2 years
When training a model, you need to make sure data loading is not your bottleneck or you're just wasting precious compute 😱 Here are some simple tests you can use to verify it 🤓
2
25
179
@hardmaru
hardmaru
2 years
Excellent article by @sedielem about diffusion models! My favorite part is about the link to RNNs: “Diffusion models present a way to train deep RNNs without backpropagating through the recurrence at all, yielding a much more scalable training procedure.” https://t.co/IOEQGXKcuL
Tweet card summary image
sander.ai
Perspectives on diffusion, or how diffusion models are autoencoders, deep latent variable models, score function predictors, reverse SDE solvers, flow-based models, RNNs, and autoregressive models,...
1
56
337