Stanislav Frolov
@stfrolov
Followers
211
Following
983
Media
26
Statuses
367
Researcher @DFKI Generative Image Modeling | Intern @MetaAI '22 & @AdobeResearch '21
Kaiserslautern, Germany
Joined November 2012
I am happy to share that SpotDiffusion was accepted to WACV 2025. Page: https://t.co/YkCPNa3Px0 Code: https://t.co/AJYj82peIO Paper: https://t.co/9yd2iq1AzQ SpotDiffusion is an efficient method for seamless panorama generation from text. 🧵
github.com
Official Pytorch Implementation for "SpotDiffusion: A Fast Approach For Seamless Panorama Generation Over Time" (WACV 2025) https://spotdiffusion.github.io/ - stanifrolov/spotdiffusion
1
1
3
Say hello to DINOv3 🦖🦖🦖 A major release that raises the bar of self-supervised vision foundation models. With stunning high-resolution dense features, it’s a game-changer for vision tasks! We scaled model size and training data, but here's what makes it special 👇
40
260
2K
Happy to share that TKG-DM, a training-free chroma key content generation diffusion model was accepted to CVPR 25. Project led by @Oguryu417 Paper: https://t.co/BpV0xFmxrP Code: https://t.co/EohvQ4qzzO
0
5
5
Checkout PromptMap, presented at IUI'25, a new interaction style with text-to-image models/data that allows users to freely explore a vast collection of synthetic prompts through a map-like view with semantic zoom. Paper: https://t.co/eYJ6UBHOGK Code: https://t.co/BRHQxjwjTV
0
1
1
I received feedback that my post about reviews not being "random" caused stress for some students. I'm sorry for that. It was meant to be empowering. Personally, I find the idea that I don't have some control over the destiny of my papers to be disheartening. If the process is
perceiving-systems.blog
5
11
180
I am very happy to share that our open-access survey about diffusion models in the field of image super-resolution got accepted by #IEEE TNNLS: https://t.co/OEz3JJGnHk
#images #diffusion #models #survey
@stfrolov @rave78 @spalaciob
ieeexplore.ieee.org
Diffusion models (DMs) have disrupted the image super-resolution (SR) field and further closed the gap between image quality and human perceptual preferences. They are easy to train and can produce...
0
3
3
We propose a time-dependent, attention-guided masking approach that prioritizes high-attention regions first, gradually refining the entire image. This improves quality across various models. Paper: https://t.co/XkTqtW3pkh Thanks to @LuckyOwl95 @rave78 @spalaciob @DFKI
0
1
1
We find that important image pixels, as measured by the attention values of DINO, are more challenging to learn (higher reconstruction error).
1
0
1
Check out our project page and paper for more visual results. Page: https://t.co/YkCPNa3Px0 Code: https://t.co/AJYj82peIO Paper: https://t.co/9yd2iq1AzQ Thanks to my collaborators @bmoser95 and Andreas Dengel. @DFKI @rptu_kl_ld @wacv_official
github.com
Official Pytorch Implementation for "SpotDiffusion: A Fast Approach For Seamless Panorama Generation Over Time" (WACV 2025) https://spotdiffusion.github.io/ - stanifrolov/spotdiffusion
0
0
1
We can produce seamless panoramas much faster by leveraging the iterative nature of diffusion models and shifting non-overlapping denoising windows over time.
1
0
1
To generate images beyond the training resolution, MultiDiffusion averages overlapping denoising windows. While this works, it can be slow because large overlap between the windows is required.
1
0
1
✨🎨🏰Super excited to share our new paper Ensemble everything everywhere: Multi-scale aggregation for adversarial robustness Inspired by biology we 1) get adversarial robustness + interpretability for free, 2) turn classifiers into generators & 3) design attacks on vLLMs 1/12
24
199
1K
I can’t find a recent paper (and tweet) that had emojis all over an image. I think it was a method about interpreting (possibly segmenting) images with/from diffusion models. Can somebody help?
0
0
0
Wow that’s cool! LoRA but for training.
Thanks @_akhaliq for promoting our work! With GaLore, now it is possible to pre-train a 7B model in NVidia RTX 4090s with 24G memory! 🤔How? Instead of assuming low-rank weight structure like LoRA, we show that the weight gradient is naturally low-rank and thus can be
0
0
0
Thanks @_akhaliq for promoting our work! With GaLore, now it is possible to pre-train a 7B model in NVidia RTX 4090s with 24G memory! 🤔How? Instead of assuming low-rank weight structure like LoRA, we show that the weight gradient is naturally low-rank and thus can be
GaLore Memory-Efficient LLM Training by Gradient Low-Rank Projection Training Large Language Models (LLMs) presents significant memory challenges, predominantly due to the growing size of weights and optimizer states. Common memory-reduction approaches, such as low-rank
21
80
426
“Write me a scientific review in the voice of Dr. Seuss and as reviewer 2, the negative reviewer who clearly doesn’t understand the paper and has probably not read the paper. Mention that there is no novelty and that the contribution is limited.” GPT-4: Oh, I've read your work,
2
36
174
Vision transformers need registers! Or at least, it seems they 𝘸𝘢𝘯𝘵 some… ViTs have artifacts in attention maps. It’s due to the model using these patches as “registers”. Just add new tokens (“[reg]”): - no artifacts - interpretable attention maps 🦖 - improved performances!
40
310
2K
When training a model, you need to make sure data loading is not your bottleneck or you're just wasting precious compute 😱 Here are some simple tests you can use to verify it 🤓
2
25
179
Excellent article by @sedielem about diffusion models! My favorite part is about the link to RNNs: “Diffusion models present a way to train deep RNNs without backpropagating through the recurrence at all, yielding a much more scalable training procedure.” https://t.co/IOEQGXKcuL
sander.ai
Perspectives on diffusion, or how diffusion models are autoencoders, deep latent variable models, score function predictors, reverse SDE solvers, flow-based models, RNNs, and autoregressive models,...
1
56
337