PKiasari Profile Banner
Peymàn M. Kiasari Profile
Peymàn M. Kiasari

@PKiasari

Followers
23
Following
450
Media
27
Statuses
108

ML researcher and engineer | currently focusing on computer vision, especially DS-CNNs

Vienna, Austria
Joined August 2021
Don't wanna be here? Send us removal request.
@PKiasari
Peymàn M. Kiasari
28 days
It's fixed.
0
0
1
@PKiasari
Peymàn M. Kiasari
2 months
Our paper "Visual Graph Arena: Evaluating AI's Visual Conceptualization" has been accepted at #ICML2025! 🎉🥳🎉. We introduce a dataset testing AI systems' ability to conceptualize graph representations. Available at: Camera ready version coming soon!
Tweet media one
2
2
11
@PKiasari
Peymàn M. Kiasari
3 months
Make peer review non-anonymous. Just like in other fields of science.
@xwang_lk
Xin Eric Wang
3 months
All reviews should be made publicly available, like ICLR.
0
0
2
@PKiasari
Peymàn M. Kiasari
4 months
I noticed the image quality is not great! It should be due to Twitters image compression. Here is a link to the PDF file of the poster:.
0
0
1
@PKiasari
Peymàn M. Kiasari
4 months
If you are interested in generalization in deep learning, you can find our poster in the poster session right now at #AAAI2025 : )
Tweet media one
@PKiasari
Peymàn M. Kiasari
5 months
Camera-ready version is out! (. TL;DR: Deep CNN filters may be generalized, not specialized as previously believed. Major update(Fig4): We froze layers from end-to-start instead of start-to-end. The result ironically suggests early layers are specialized!
Tweet media one
1
1
5
@PKiasari
Peymàn M. Kiasari
5 months
Camera-ready version is out! (. TL;DR: Deep CNN filters may be generalized, not specialized as previously believed. Major update(Fig4): We froze layers from end-to-start instead of start-to-end. The result ironically suggests early layers are specialized!
Tweet media one
@PKiasari
Peymàn M. Kiasari
5 months
🧵1/13 Happy to announce our AAAI paper! We challenge the notion that CNN filters should specialize in deeper layers- mostly established in @jasonyo's influential work. Our findings suggest something quite different!.This is a continuation of our CNN understanding paper series:.
0
0
3
@PKiasari
Peymàn M. Kiasari
5 months
RT @giffmana: I took a brief look at the Harmonic Loss paper. tl;dr: instead of dot-product with softmax, do euclid dist with normalized 1/….
0
147
0
@PKiasari
Peymàn M. Kiasari
5 months
A company created a huge open-source model along with a full paper. They publicly shared everything they had—unlike closed companies. Now, people are debating whether they are a "threat" or not. I wonder if the reaction would be the same if they weren't Chinese.
@DarioAmodei
Dario Amodei
5 months
My thoughts on China, export controls and two possible futures
0
0
9
@PKiasari
Peymàn M. Kiasari
5 months
RT @PaulGavrikov: Our discoveries? VLMs naturally lean more towards shape, mimicking human perception more closely than we thought! And we….
0
2
0
@PKiasari
Peymàn M. Kiasari
5 months
13/13 If you prefer video content, you can check out the video I made for AAAI:. Thanks for reading! I wish a great day for you. #DeepLearning #ComputerVision #AAAI2025.
0
0
1
@PKiasari
Peymàn M. Kiasari
5 months
12/13 Thank you for reading this far! Curious to hear your thoughts on this. You can find our complete paper "The Master Key Filters Hypothesis" here:
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
11/13 We went even further, transferring FROZEN❄️ filters between two different architectures (Hornet → ConvNeXt) trained on unrelated datasets (Food → Pets) improved accuracy by 3.1%. We suggest these "master key" filters are truly architecture-independent!
Tweet media one
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
10/13 To test this, we conducted cross-dataset FROZEN❄️ transfer experiments. If true, filters trained on larger datasets should help models perform better on smaller, even unrelated datasets - as they'd better converged to these "master key" filters. Results were confirming.
Tweet media one
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
9/13 We propose that there exist universal "master key" filter sets optimal for visual data processing. The depth-wise filters in DS-CNN naturally converge toward these general-purpose filters, regardless of the specific dataset, task, or architecture!.
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
8/13 This contradicts the intuitive notion that models should develop more specialized filters as get deeper. Enter our "Master Key Filters" hypothesis. 🔑.
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
7/13 With top model reaching ~94% retention, the trend suggests near~100% retention may be possible with better models. Conversely, as accuracy drops, we approach Yosinski's findings. But why do deeper, better-performing, models become MORE general, and not more specialized?
Tweet media one
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
6/13 Our investigation says no! When we transfer all of the layers of ResNets and draw a chart of "retained accuracy %" vs "original accuracy", a clear trend emerged: the better a model performs on its original task, the more "general" (not specialized) its deep filters become.
Tweet media one
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
5/13 We had to test with CNNs too. ResNet50 (a classical CNN) maintained 92.5% of its original accuracy when all layers were transferred! This number reaches to 93-94% for ResNet152. But why? is this because of residual connections?
Tweet media one
Tweet media two
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
4/13 So we replicated Yosinski's experiment with modern DS-CNNs: we split ImageNet into man-made and natural images, transferred & froze filters between models up to layer 'n'. Surprisingly, there was NO significant accuracy drop!. But why? is this because DS-CNNs architecture?
Tweet media one
Tweet media two
1
0
1
@PKiasari
Peymàn M. Kiasari
5 months
3/13 Our previous paper, "Unveiling the Unseen", revealed something intriguing: DS-CNN filters showed remarkably similar patterns across ALL layers! This made us question: how can early layers be "general" and deeper ones "specialized"? Time to revisit Yosinski's experiment. 🤔
Tweet media one
1
0
1