JainRohan16 Profile Banner
Rohan Jain Profile
Rohan Jain

@JainRohan16

Followers
182
Following
4K
Media
9
Statuses
60

• MSc in ML @UCalgary, ML Researcher @ml_collective • Prev: Math + CS @UofT

Calgary, Alberta
Joined October 2021
Don't wanna be here? Send us removal request.
@JainRohan16
Rohan Jain
3 months
🎉 Excited to share our latest work has been accepted at #ICML2025. We explain how Lottery Ticket masks can be extended to arbitrary random initializations by leveraging permutations of weight symmetries achieving strong generalization. Check out the details:.
@adnan_ahmad1306
Mohammed Adnan @ ICML 2025
3 months
1/10 🧵.🔍Can weight symmetry provide insights into sparse training and the Lottery Ticket Hypothesis?. 🧐We dive deep into this question in our latest paper, "Sparse Training from Random Initialization: Aligning Lottery Ticket Masks using Weight Symmetry", accepted at #ICML2025
Tweet media one
2
11
32
@JainRohan16
Rohan Jain
2 days
RT @jxbz: I just wrote my first blog post in four years! It is called "Deriving Muon". It covers the theory that led to Muon and how, for m….
0
132
0
@JainRohan16
Rohan Jain
4 days
RT @unireps: Ready to present your latest work? The Call for Papers for #UniReps2025 @NeurIPSConf is open!. 👉Check the CFP: .
0
11
0
@JainRohan16
Rohan Jain
9 days
Interesting insights presented in this work. Would definitely check it out!.
@adnan_ahmad1306
Mohammed Adnan @ ICML 2025
10 days
[1/2] 🧵.Presenting our work on Understanding Normalization Layers for Sparse Training today at the HiLD Workshop, ICML 2025!. If you're curious about how BatchNorm & LayerNorm impact sparse networks, drop by our poster!.
0
0
4
@JainRohan16
Rohan Jain
12 days
🚨 Curious about neural network loss landscapes and sparse neural networks? Come visit our poster happening NOW! #ICML2025 . 📍East Exhibition Hall A-B, #E-2106. 🕚 11:00-1:30
Tweet media one
1
1
23
@JainRohan16
Rohan Jain
14 days
Super cool!.
@Theus__A
Alexander Theus
18 days
1/ 🚨 New paper alert! 🚨.We explore a key question in deep learning:.Can independently trained Transformers be linearly connected in weight space — without a loss barrier?.Yes — if you uncover their rich symmetries. 📄 arXiv:
Tweet media one
0
1
5
@JainRohan16
Rohan Jain
17 days
RT @joyce_xxz: I am excited to share that I will present two of my works at #ICML2025 workshops!. If you are interested in AI security and….
0
8
0
@JainRohan16
Rohan Jain
18 days
RT @yanii: @UCalgaryML will be at #ICML2025 in Vancouver🇨🇦 next week: our lab has 6 different works being presented by 5 students across bo….
0
9
0
@JainRohan16
Rohan Jain
18 days
RT @Anthony_Bonato: Everyone on Earth would fail this test
Tweet media one
0
101
0
@JainRohan16
Rohan Jain
1 month
RT @BoZhao__: When and why are neural network solutions connected by low-loss paths?. In our #ICML2025 paper, we show that mode connectivit….
0
30
0
@JainRohan16
Rohan Jain
2 months
RT @mikelasby: Want faster #LLM inference without sacrificing accuracy? 🚀. Introducing SD², a novel method using Self-Distilled Sparse Draf….
0
9
0
@JainRohan16
Rohan Jain
2 months
RT @roydanroy: More work at the intersection of weight symmetry, linear mode connectivity, and sparsity in deep networks. There really nee….
0
3
0
@JainRohan16
Rohan Jain
4 months
RT @yanii: Attending @CPALconf this week on the beautiful Stanford campus! . Will be presenting our work with @adnan_ahmad1306 @JainRohan1….
0
2
0
@JainRohan16
Rohan Jain
5 months
RT @manifest__ai: Why gradient descent minimizes training loss:
Tweet card summary image
manifestai.com
Convexity is unnecessary to guarantee convergence to low loss
0
5
0
@JainRohan16
Rohan Jain
7 months
Excited to share our work with the community. Thanks @CohereForAI for hosting!.
@Cohere_Labs
Cohere Labs
7 months
Mark your calendars for January 17th when @JainRohan16 presents "Winning Tickets from Random Initialization: Aligning Masks for Sparse Training" . Special thanks to @Sree_Harsha_N and @aniervs.for organizing this event 🥳 . Learn more:
Tweet media one
1
6
23
@JainRohan16
Rohan Jain
8 months
📢 Interested in weight symmetries in NN loss landscapes, the Lottery Ticket Hypothesis & sparse neural networks . Come visit our poster at @unireps today!. 📆 Dec 14, 3:45-5:00 PM . 📍West Exhibition Hall C.
@JainRohan16
Rohan Jain
8 months
✨Our new @unireps paper tries to answer why the Lottery Ticket Hypothesis (LTH) fails to work for different random inits through the lens of weight-space symmetry. We improve the transferability of LTH masks to new random inits leveraging weight symmetries. 🧵(1/6)
Tweet media one
Tweet media two
0
4
8
@JainRohan16
Rohan Jain
8 months
RT @unireps: The UniReps Workshop is happening THIS SATURDAY at #NeurIPS! 🤖🧠. Join us for a day of insightful talks and engaging discussion….
0
14
0
@JainRohan16
Rohan Jain
8 months
RT @yanii: I'm proud that the @UCalgaryML lab will have 6 different works being presented by 6 students across #NeurIPS2024, in workshops (….
0
10
0
@JainRohan16
Rohan Jain
8 months
For more details, check out our full paper … 👇🏼. 🔗: Joint work w/ @adnan_ahmad1306, @EkanshSh, and @yanii from @VectorInst and @UCalgary. Visit our poster session on Dec 14 at #NeurIPS2024! 👋🏻. Supported by: @UCalgaryML. (6/6).
0
2
15
@JainRohan16
Rohan Jain
8 months
We also observed that larger width exhibits better linear mode connectivity. As the width of the model increases, the permutation matching algorithm gets more accurate, thereby reducing the loss barrier & improving our permuted solution . 👇🏼. (5/6)
Tweet media one
1
0
11