Sa_9810 Profile Banner
Shaden Profile
Shaden

@Sa_9810

Followers
137
Following
200
Media
14
Statuses
36

graduate student @MIT | doing representation learning and math

Cambridge, MA
Joined April 2021
Don't wanna be here? Send us removal request.
@Sa_9810
Shaden
3 months
Excited to share our ICLR 2025 paper, I-Con, a unifying framework that ties together 23 methods across representation learning, from self-supervised learning to dimensionality reduction and clustering. Website: A thread ๐Ÿงต 1/n
1
24
88
@Sa_9810
Shaden
16 days
RT @MasonKamb: I'm at ICML presenting this work! Come by on Tuesday to hear about/chat about combinatorial generalization and creativity inโ€ฆ.
0
28
0
@Sa_9810
Shaden
2 months
RT @ema_marconato: ๐ŸงตWhy are linear properties so ubiquitous in LLM representations?. We explore this question through the lens of ๐—ถ๐—ฑ๐—ฒ๐—ป๐˜๐—ถ๐—ณ๐—ถ๐—ฎโ€ฆ.
0
61
0
@Sa_9810
Shaden
2 months
RT @_AmilDravid: Artifacts in your attention maps? Forgot to train with registers? Use ๐™ฉ๐™š๐™จ๐™ฉ-๐™ฉ๐™ž๐™ข๐™š ๐™ง๐™š๐™œ๐™ž๐™จ๐™ฉ๐™š๐™ง๐™จ! We find a sparse set of activatโ€ฆ.
0
62
0
@Sa_9810
Shaden
2 months
RT @KumailAlhamoud: We've seen hilarious fails from generative models struggling with "NO" (e.g., asking for "a clear sky with no planes",โ€ฆ.
0
1
0
@Sa_9810
Shaden
3 months
RT @ShivamDuggal4: Drop by our poster at Hall 3 + Hall 2B, #99 at 10 AM SGT!.Unfortunately none of us could travel, but our amazing friendsโ€ฆ.
0
5
0
@Sa_9810
Shaden
3 months
RT @juliachae_: My first first-authored (w/ @shobsund) paper of my phd is finally out! ๐Ÿš€ . Check out our thread to see how general-purposeโ€ฆ.
0
7
0
@Sa_9810
Shaden
3 months
n/n. Huge thanks to my amazing collaborators and advisors: @mhamilton723, John Hershey, Axel Feldmann, and William T. Freeman! . โ€ข Website: โ€ข Full Paper:
Tweet card summary image
arxiv.org
As the field of representation learning grows, there has been a proliferation of different loss functions to solve different classes of problems. We introduce a single information-theoretic...
3
0
4
@Sa_9810
Shaden
3 months
7/n . There are still many open questions from these insights:. (1) What new methods emerge by filling more gaps or even adding new rows or columns?.(2) How does using divergences beyond KL reshape things?.(3) What makes methods outside the framework fundamentally different?.
1
0
4
@Sa_9810
Shaden
3 months
6/n . Our use of neighborhood expansion was inspired by the heavy-tailed distributions such as Student-T in t-SNE for dimensionality reduction, so we adapted the idea to discrete settings like neighbor propagation.
Tweet media one
1
0
3
@Sa_9810
Shaden
3 months
RT @_akhaliq: I-Con. A Unifying Framework for Representation Learning
Tweet media one
0
24
0
@Sa_9810
Shaden
3 months
5/n . How does I-Con yield practical gains?.(1) Enables principled design over heuristics.(2) Transfers ideas across domains. We applied neighborhood expansion for debiasing, achieving:.โœ… +8% & SOTA on clustering ImageNet1K.โœ… Better probabilities calibration
Tweet media one
1
0
4
@Sa_9810
Shaden
3 months
4/n. We provide a codebase that enables one-line implementation and training for many of these methods to make it easy to explore and compare different choices of P and Q
1
1
4
@Sa_9810
Shaden
3 months
3/n. Varying the definitions of P and Q recovers a wide range of existing methods! Here's an illustration using specific choices of P and Q that recover t-SNE, SimCLR, K-Means, and supervised cross-entropy.
1
0
4
@Sa_9810
Shaden
3 months
2/n. I-Con is built on a simple but general principle: define two conditional distributions โ€” a supervisory signal P and a learned signal Q, where Q is parameterized by the learned representations. Learning proceeds by minimizing the KL divergence.
1
0
3
@Sa_9810
Shaden
3 months
RT @mhamilton723: Excited to share our new discovery of an equation that generalizes over 23 different machine learning algorithms. We useโ€ฆ.
0
4
0
@Sa_9810
Shaden
9 months
RT @ShivamDuggal4: Current vision systems use fixed-length representations for all images. In contrast, human intelligence or LLMs (eg: Opeโ€ฆ.
0
67
0