
Michal Golovanevsky
@MichalGolov
Followers
47
Following
20
Media
7
Statuses
28
CS PhD student @BrownCSDept | Multimodal Learning | Mechanistic Interpretability | Clinical Deep Learning.
Providence, RI
Joined September 2022
RT @Michael_Lepori: How do VLMs balance visual information presented in-context with linguistic priors encoded in-weights? In this project,….
0
2
0
RT @WilliamRudmanjr: Models rely on memorized priors early in their processing but shift toward visual evidence in mid-to-late layers. This….
0
1
0
RT @WilliamRudmanjr: We create Visual CounterFact: a dataset of realistic images that contrast pixel evidence against memorized knowledge.….
0
1
0
RT @WilliamRudmanjr: With PvP, we can shift 92.5% of color predictions and 74.6% of size predictions from memorized priors to counterfactua….
0
1
0
RT @WilliamRudmanjr: When vision-language models answer questions, are they truly analyzing the image or relying on memorized facts? We int….
0
4
0
RT @CSVisionPapers: Pixels Versus Priors: Controlling Knowledge Priors in Vision-Language Models through Visual Counterfacts. https://t.co/….
0
1
0
If SOTA models fail to recognize simple shapes, should we be evaluating them on complex geometric tasks? Most MLLMs struggle with counting the number of sides of regular polygons and all MLLMs receive 0% on novel shapes. @WilliamRudmanjr.@_amirbar @vedantpalit1008 [1/6]
1
6
12
RT @WilliamRudmanjr: NOTICE uses Symmetric Token Replacement for text corruption and Semantic Image Pairs (SIP) for image corruption. SIP r….
0
1
0
RT @WilliamRudmanjr: We extend the generalizability of NOTICE by using Stable-Diffusion to generate semantic image pairs and find results a….
0
1
0
RT @WilliamRudmanjr: The finding that important attention heads implement one of a small set of interpretable functions boosts transparency….
0
2
0
RT @WilliamRudmanjr: How do VLMs like BLIP and LLaVA differ in how they process visual information? Using our mech-interp pipeline for VLMs….
0
2
0
RT @WilliamRudmanjr: Instead, LLaVA relies on self-attention heads to manage “outlier” attention patterns in the image, focusing on regulat….
0
1
0
RT @WilliamRudmanjr: The finding that important cross-attention heads implement one of a small set of interpretable functions helps boost V….
0
1
0
RT @WilliamRudmanjr: By visualizing cross-attention patterns, we've discovered that these universal heads fall into three functional catego….
0
1
0