Youssef Allouah
@ys_alh
Followers
208
Following
140
Media
11
Statuses
31
PhD in CS at @EPFL_en. Working on trustworthy AI/ML. Previously: @Polytechnique, @AmazonScience, @StanfordAILab.
Joined January 2022
I have officially graduated with a Ph.D. from @EPFL! It was an honor to defend my thesis before an exceptional jury: Francis Bach @BachFrancis, Samy Bengio, Gautam Kamath @thegautamkamath, Adam Smith, Rachid Guerraoui (advisor), and Emre Telatar (chair).
16
11
211
Stanford Trustworthy AI Research (@stai_research) has exceptional researchers on the faculty market this year! ๐งต ๐น Anka Reuel - Technical AI governance ๐น Dr. Olawale Salaudeen - AI measurement & robustness ๐น Dr. Andreas Haupt - AI, Economics, and Policy
2
18
123
My thesis on the "cost of trust in machine learning" is available online (link below). I could not have done this research without my collaborators, mentors, friends, and family. I am deeply grateful to all of them. Thesis link:
infoscience.epfl.ch
As machine learning systems move from statistical tools to core societal infrastructure, their trustworthiness has become a primary scientific challenge. This requires a foundational shift from...
0
0
6
@sanmikoyejo @Ana_koloskova @sanmi ๐ฃ๐ผ๐๐๐ฒ๐ฟ: ๐ง๐ฟ๐๐๐๐๐ผ๐ฟ๐๐ต๐ ๐๐ฒ๐ฑ๐ฒ๐ฟ๐ฎ๐๐ฒ๐ฑ ๐๐ฒ๐ฎ๐ฟ๐ป๐ถ๐ป๐ด ๐๐ถ๐๐ต ๐จ๐ป๐๐ฟ๐๐๐๐ฒ๐ฑ ๐ฃ๐ฎ๐ฟ๐๐ถ๐ฐ๐ถ๐ฝ๐ฎ๐ป๐๐ (๐ง๐ต๐๐ฟ๐๐ฑ๐ฎ๐, ๐๐๐น๐ ๐ญ๐ณ). A culmination of several works on privacy vs. robustness trade-offs. Paper: https://t.co/JUhpY5wehp w/ John, Rachid
0
1
2
@sanmikoyejo ๐ฃ๐ผ๐๐๐ฒ๐ฟ: ๐๐ฒ๐ฟ๐๐ถ๐ณ๐ถ๐ฒ๐ฑ ๐จ๐ป๐น๐ฒ๐ฎ๐ฟ๐ป๐ถ๐ป๐ด ๐ณ๐ผ๐ฟ ๐ก๐ฒ๐๐ฟ๐ฎ๐น ๐ก๐ฒ๐๐๐ผ๐ฟ๐ธ๐ (๐ง๐๐ฒ๐๐ฑ๐ฎ๐, ๐๐๐น๐ ๐ญ๐ฑ). An exciting and important step towards making certified unlearning practical. Paper: https://t.co/F7og8rjDtE (w/ @Ana_koloskova, Animesh, Rachid, @sanmi )
1
1
3
๐ข๐ฟ๐ฎ๐น ๐ฎ๐ ๐๐ต๐ฒ ๐ ๐ฎ๐ฐ๐ต๐ถ๐ป๐ฒ ๐จ๐ป๐น๐ฒ๐ฎ๐ฟ๐ป๐ถ๐ป๐ด ๐ณ๐ผ๐ฟ ๐๐ฒ๐ป๐ฒ๐ฟ๐ฎ๐๐ถ๐๐ฒ ๐๐ ๐๐ผ๐ฟ๐ธ๐๐ต๐ผ๐ฝ (๐๐ฟ๐ถ๐ฑ๐ฎ๐, ๐๐๐น๐ ๐ญ๐ด). On our latest research introducing a new distributional framework for unlearning. Paper: https://t.co/PWbnmu7Y41 (w/ @sanmikoyejo, Rachid)
1
1
4
I will be attending #icml2025 in Vancouver next week! This is a special one for me, as it may be my last as a PhD student. I'll be presenting works on unlearning, privacy, and robustness. I'm happy to connect and exchange ideas! You can find me at these presentations:
1
0
10
Joint work with @Ana_koloskova (co-first author), Animesh Jha, Rachid Guerraoui, and @sanmikoyejo. @stai_research @StanfordAILab @ICepfl
0
2
5
Excited our paper "Certified Unlearning for Neural Networks" is accepted at ICML 2025! We introduce a method for provable machine unlearning-- truly "forgetting" data without restrictive assumptions like convexity. Paper: https://t.co/F7og8rjDtE Code: https://t.co/LLQD8HvcbY
4
20
107
I am presenting the poster for our unlearning paper (below) at @iclr_conf in Hall 2, now! Come by to have a chat.
0
0
2
This is joint work with @JoshuaK92829, R. Guerraoui, @sanmikoyejo. Paper: https://t.co/VJv6P2nq33 n/n
arxiv.org
Machine unlearning, the process of selectively removing data from trained models, is increasingly crucial for addressing privacy concerns and knowledge gaps post-deployment. Despite this...
0
2
4
Solution: Use robust mean estimation during pre-training. Trimmed mean strategies make OOD unlearning faster & more reliable. See controlled experiment below comparing vanilla (Alg. 1) vs. robust fine-tuning (Alg. 2) for various number f (out of n) of forget samples. 10/n
1
2
3
Finding 3. For Out-of-Distribution Forget Data, Robust Pre-Training is Crucial If the forget data is very different from the retained data, fine-tuning alone can failโit might take more time to unlearn a single sample, than to fully retrain the model. ๐จ 9/n
1
2
3
The latter capacity decreases with the model dimension for DP (Sekhari et al., 2021, Huang & Canonne, 2023). We show that the same quantity is independent of model dimension for noisy fine-tuning. See experiment on linear regression below (Alg. 1 = Noisy fine-tuning) 8/n
1
2
3
Finding 2. Differential Privacy (DP) implies UnlearningโBut is an Overkill DP implies unlearning, but it severely limits model performance. We prove a tight separation between DP and noisy fine-tuning, in terms of how many samples can be deleted at a fixed test loss. 7/n
1
2
2
For our analysis, we require access to an approximate global risk minimizer. For convex tasks, this can simply be SGD, with near-linear time complexity (Neel et al., 2021). Such a minimizer also exists for structured non-convex tasks, e.g., PCA, matrix completion. 6/n
1
2
2
Finding 1: Noisy Fine-Tuning Works for In-Distribution Forget Data If the forget data comes from the same distribution as the retained data, noisy fine-tuning (see below) is a highly effective and practical unlearning method. 5/n
1
2
2
Problem: What Does It Mean to "Unlearn"? We define certified unlearning as making the post-unlearning model statistically near-indistinguishable from a model that was trained without the forget data. This is inspired from differential privacy (Dwork & Roth, 2014). 4/n
1
3
5
Our main insight: it depends on the nature of the forget data. If it's in-distribution โ Noisy fine-tuning works great. Differential privacy may be an overkill. If it's out-of-distribution โ (Noisy) Fine-tuning struggles. A robust pre-training strategy is needed. 3/n
1
2
4
Key Question: Can we truly remove data from a trained AI modelโwithout hurting performance or requiring a full retrain? 2/n
1
2
3