Baharan Mirzasoleiman Profile
Baharan Mirzasoleiman

@baharanm

Followers
2K
Following
266
Media
4
Statuses
77

Assistant professor @UCLAComSci. Better ML via better data, Machine learning, Optimization

Los Angeles, CA
Joined July 2018
Don't wanna be here? Send us removal request.
@baharanm
Baharan Mirzasoleiman
1 year
We’re thrilled by the amazing response to our #ICML2024 tutorial on “Foundations of data-efficient learning”! Over 1000 attendees joined us. Thank you all! 🙌🌱🌱🌱. ➡️ Slides: ➡️ Recording: will be available on Aug 22 🎊🎊
Tweet media one
@baharanm
Baharan Mirzasoleiman
1 year
I'll be giving a 2-hour tutorial on data-efficient learning with my PhD student @sjoshi804 on Monday July 22 at #ICML2024. Join us to learn more about this cool topic! ➡️ We can learn better from better data! ⬅️🙌🌱
Tweet media one
6
18
216
@baharanm
Baharan Mirzasoleiman
3 months
Can weak LLMs supervise strong LLMs to obtain superior performance? 🤔 Yes!! 🤩. Which weak models are better supervisors? 🤔. Check out @xue_yihao65785’s awesome #icml2025 paper to know how to identify best weak supervisors without having to collect labels! 🎉🌱.
@xue_yihao65785
Yihao Xue
3 months
🎉 Our paper “Representations Shape Weak-to-Strong Generalization” is accepted at #ICML2025!.We study weak-to-strong generalization (W2SG)—a core problem in superalignment—and offer new insights into the role of models' internal representations in W2SG. 1/
Tweet media one
1
2
25
@grok
Grok
7 days
Join millions who have switched to Grok.
254
501
4K
@baharanm
Baharan Mirzasoleiman
4 months
RT @SCSLWorkshop: 🚨 Join us at the Workshop on Spurious Correlation & Shortcut Learning (SCSL) at #ICLR2025!.@iclr_conf .🗓️ April 28, 2025….
0
15
0
@baharanm
Baharan Mirzasoleiman
4 months
Here is the paper:
0
0
1
@baharanm
Baharan Mirzasoleiman
4 months
Can we pretrain deep models with small synthetic data? . Dataset Distillation via Knowledge Distillation is the way to go!. Check out @sjoshi804’s #ICLR2025 paper this Saturday April 26 at 9am, Poster #307 🎉🌱.
@sjoshi804
Siddharth Joshi
4 months
#ICLR2025.Can you pre-train deep models with small, synthetic datasets? 🤯. We introduce the first effective dataset distillation method for self-supervised learning (SSL) — boosting downstream accuracy by up to 13% over baselines. 🧪 Poster #307, Sat Apr 26, 9am.
1
4
37
@baharanm
Baharan Mirzasoleiman
4 months
Want to train LLMs with better performance and lower GPU memory requirements on data mixtures?. Check out this cool #ICLR2025 paper of @dangnth97 this Friday April 25 at 10am, Poster #265.🎉🌱.
@dangnth97
Dang Nguyen
4 months
🎉 Achievement unlocked: having papers with all of my labmates and somehow all ended up at ICLR!. I’ll be presenting our work “Mini-batch Coresets for Memory-efficient Language Model Training on Data Mixtures” at #ICLR2025 🇸🇬. Come by and chat! 👋 on Fri, Apr 25 | 10 AM GMT+8.
1
0
19
@baharanm
Baharan Mirzasoleiman
6 months
Big congrats @YuYang_i on your graduation!! 🎉🎉 🎉.very nice PhD thesis with great contributions 🌱.I’m proud of all you’ve done, and I wish you the best! 💝.
@YuYang_i
Yu Yang
6 months
Sharing a little late update (before it’s no longer news): I wrapped up my PhD at the end of last year and recently joined @OpenAI’s reasoning team 🍓✨!.
0
0
55
@baharanm
Baharan Mirzasoleiman
9 months
RT @MhRohban: We are delighted that our proposal for the Workshop on “Spurious Correlation and Shortcut Learning: Foundations and Solutions….
0
16
0
@baharanm
Baharan Mirzasoleiman
9 months
RT @andrew_ilyas: At NeurIPS? Check out the 2nd workshop on Attributing Model Behavior at Scale (ATTRIB)!. Meeting Rm 205-207, starting @ 9….
0
8
0
@baharanm
Baharan Mirzasoleiman
9 months
I’ll help presenting our #NeurIPS2024 posters tomorrow (Friday):🌱. 1- Changing the training data distribution to improve in-distribution performance (11@west #7106) w. @dangnth97. 2- Data selection for fine-tuning LLMs with superior performance (16:30@west #5401) w. @YUYANG_UCLA.
2
5
59
@baharanm
Baharan Mirzasoleiman
9 months
RT @WuyangC: Attending NeurIPS'24?.Please mark your calendar for our special event "SFU@NeurIPS 2024" 9 speakers fr….
0
19
0
@baharanm
Baharan Mirzasoleiman
9 months
Same training and test distribution yields optimal in-distribution performance?. @dangnth97 showed in his #NeurIPS2024 paper that this is not true when training with gradient methods!!😮🙃.Changing the training data distribution yields SOTA!🎊. Check it out Fri Dec 13, 11am, PS#5.
0
1
23
@baharanm
Baharan Mirzasoleiman
9 months
Smaller high-quality subsets of language data not only improve LLMs’ training efficiency, but also yield considerably better performance! 🙌🎉🌱. @YUYANG_UCLA has a theoretically-rigorous method for this in her #NeurIPS2024 paper!. Check it out on Fri, Dec 13, 16:30, #PS 6.
@YuYang_i
Yu Yang
9 months
1/ I'll be at #NeurIPS2024 presenting our work SmallToLarge (S2L): Data-efficient Fine-tuning of LLMs! 🚀. What’s S2L? It’s a scalable data selection method that trains a small proxy model to guide fine-tuning for larger models, reducing costs while preserving performance. 👇
Tweet media one
0
9
48
@baharanm
Baharan Mirzasoleiman
1 year
RT @_xiang_chen_: 📢 @UCLAComSci is hiring! Open to all CS areas!. - Multiple Tenure-track Assistant Professor Positions: .
0
27
0
@baharanm
Baharan Mirzasoleiman
1 year
RT @pinyuchenTW: The Adversarial Machine Learning Rising Star Awards deadline is in two weeks! Submit your application and help us promote….
0
2
0
@baharanm
Baharan Mirzasoleiman
1 year
I’ll also present “SafeClip” on behalf of @WenhanYang0315 tomorrow at 1:30pm (poster session 6) #814. See you there! 🙌.
@baharanm
Baharan Mirzasoleiman
1 year
CLIP is highly sensitive to data poisoning and backdoor attacks. In this #ICML2024 paper, @WenhanYang0315 proposed an interesting way to pretrain CLIP robust to such attacks without compromising the performance! 🌱🌱.🔗Thu, July 25, Poster session 6, #814
Tweet media one
0
2
18
@baharanm
Baharan Mirzasoleiman
1 year
I’ll present “MixPro” on behalf of @xue_yihao65785 tomorrow at 11:30 (poster session 5) poster #800. Come check it out 🙌.
@baharanm
Baharan Mirzasoleiman
1 year
ML models are sensitive to distribution shift. Can we adapt a model with only a few examples from the target domain? In this #ICML2024 paper, @xue_yihao65785 proposes an effective way, with nice theoretical analysis🌱.🔗Thu, July 25, Poster session 5, #800
Tweet media one
0
1
18
@baharanm
Baharan Mirzasoleiman
1 year
RT @AdvMLFrontiers: 📢 We're back with a new edition, this year at.@NeurIPSConf in Vancouver! . Paper deadline is August 30th, we are look….
0
7
0