Youssef Allouah Profile
Youssef Allouah

@ys_alh

Followers
208
Following
140
Media
11
Statuses
31

PhD in CS at @EPFL_en. Working on trustworthy AI/ML. Previously: @Polytechnique, @AmazonScience, @StanfordAILab.

Joined January 2022
Don't wanna be here? Send us removal request.
@ys_alh
Youssef Allouah
19 days
I have officially graduated with a Ph.D. from @EPFL! It was an honor to defend my thesis before an exceptional jury: Francis Bach @BachFrancis, Samy Bengio, Gautam Kamath @thegautamkamath, Adam Smith, Rachid Guerraoui (advisor), and Emre Telatar (chair).
16
11
211
@sanmikoyejo
Sanmi Koyejo
10 days
Stanford Trustworthy AI Research (@stai_research) has exceptional researchers on the faculty market this year! ๐Ÿงต ๐Ÿ”น Anka Reuel - Technical AI governance ๐Ÿ”น Dr. Olawale Salaudeen - AI measurement & robustness ๐Ÿ”น Dr. Andreas Haupt - AI, Economics, and Policy
2
18
123
@ys_alh
Youssef Allouah
19 days
My thesis on the "cost of trust in machine learning" is available online (link below). I could not have done this research without my collaborators, mentors, friends, and family. I am deeply grateful to all of them. Thesis link:
Tweet card summary image
infoscience.epfl.ch
As machine learning systems move from statistical tools to core societal infrastructure, their trustworthiness has become a primary scientific challenge. This requires a foundational shift from...
0
0
6
@ys_alh
Youssef Allouah
5 months
@sanmikoyejo @Ana_koloskova @sanmi ๐—ฃ๐—ผ๐˜€๐˜๐—ฒ๐—ฟ: ๐—ง๐—ฟ๐˜‚๐˜€๐˜๐˜„๐—ผ๐—ฟ๐˜๐—ต๐˜† ๐—™๐—ฒ๐—ฑ๐—ฒ๐—ฟ๐—ฎ๐˜๐—ฒ๐—ฑ ๐—Ÿ๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐˜„๐—ถ๐˜๐—ต ๐—จ๐—ป๐˜๐—ฟ๐˜‚๐˜€๐˜๐—ฒ๐—ฑ ๐—ฃ๐—ฎ๐—ฟ๐˜๐—ถ๐—ฐ๐—ถ๐—ฝ๐—ฎ๐—ป๐˜๐˜€ (๐—ง๐—ต๐˜‚๐—ฟ๐˜€๐—ฑ๐—ฎ๐˜†, ๐—๐˜‚๐—น๐˜† ๐Ÿญ๐Ÿณ). A culmination of several works on privacy vs. robustness trade-offs. Paper: https://t.co/JUhpY5wehp w/ John, Rachid
0
1
2
@ys_alh
Youssef Allouah
5 months
@sanmikoyejo ๐—ฃ๐—ผ๐˜€๐˜๐—ฒ๐—ฟ: ๐—–๐—ฒ๐—ฟ๐˜๐—ถ๐—ณ๐—ถ๐—ฒ๐—ฑ ๐—จ๐—ป๐—น๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐—ณ๐—ผ๐—ฟ ๐—ก๐—ฒ๐˜‚๐—ฟ๐—ฎ๐—น ๐—ก๐—ฒ๐˜๐˜„๐—ผ๐—ฟ๐—ธ๐˜€ (๐—ง๐˜‚๐—ฒ๐˜€๐—ฑ๐—ฎ๐˜†, ๐—๐˜‚๐—น๐˜† ๐Ÿญ๐Ÿฑ). An exciting and important step towards making certified unlearning practical. Paper: https://t.co/F7og8rjDtE (w/ @Ana_koloskova, Animesh, Rachid, @sanmi )
1
1
3
@ys_alh
Youssef Allouah
5 months
๐—ข๐—ฟ๐—ฎ๐—น ๐—ฎ๐˜ ๐˜๐—ต๐—ฒ ๐— ๐—ฎ๐—ฐ๐—ต๐—ถ๐—ป๐—ฒ ๐—จ๐—ป๐—น๐—ฒ๐—ฎ๐—ฟ๐—ป๐—ถ๐—ป๐—ด ๐—ณ๐—ผ๐—ฟ ๐—š๐—ฒ๐—ป๐—ฒ๐—ฟ๐—ฎ๐˜๐—ถ๐˜ƒ๐—ฒ ๐—”๐—œ ๐˜„๐—ผ๐—ฟ๐—ธ๐˜€๐—ต๐—ผ๐—ฝ (๐—™๐—ฟ๐—ถ๐—ฑ๐—ฎ๐˜†, ๐—๐˜‚๐—น๐˜† ๐Ÿญ๐Ÿด). On our latest research introducing a new distributional framework for unlearning. Paper: https://t.co/PWbnmu7Y41 (w/ @sanmikoyejo, Rachid)
1
1
4
@ys_alh
Youssef Allouah
5 months
I will be attending #icml2025 in Vancouver next week! This is a special one for me, as it may be my last as a PhD student. I'll be presenting works on unlearning, privacy, and robustness. I'm happy to connect and exchange ideas! You can find me at these presentations:
1
0
10
@ys_alh
Youssef Allouah
5 months
Joint work with @Ana_koloskova (co-first author), Animesh Jha, Rachid Guerraoui, and @sanmikoyejo. @stai_research @StanfordAILab @ICepfl
0
2
5
@ys_alh
Youssef Allouah
5 months
Excited our paper "Certified Unlearning for Neural Networks" is accepted at ICML 2025! We introduce a method for provable machine unlearning-- truly "forgetting" data without restrictive assumptions like convexity. Paper: https://t.co/F7og8rjDtE Code: https://t.co/LLQD8HvcbY
4
20
107
@ys_alh
Youssef Allouah
7 months
I am presenting the poster for our unlearning paper (below) at @iclr_conf in Hall 2, now! Come by to have a chat.
0
0
2
@ys_alh
Youssef Allouah
10 months
Solution: Use robust mean estimation during pre-training. Trimmed mean strategies make OOD unlearning faster & more reliable. See controlled experiment below comparing vanilla (Alg. 1) vs. robust fine-tuning (Alg. 2) for various number f (out of n) of forget samples. 10/n
1
2
3
@ys_alh
Youssef Allouah
10 months
Finding 3. For Out-of-Distribution Forget Data, Robust Pre-Training is Crucial If the forget data is very different from the retained data, fine-tuning alone can failโ€”it might take more time to unlearn a single sample, than to fully retrain the model. ๐Ÿ˜จ 9/n
1
2
3
@ys_alh
Youssef Allouah
10 months
The latter capacity decreases with the model dimension for DP (Sekhari et al., 2021, Huang & Canonne, 2023). We show that the same quantity is independent of model dimension for noisy fine-tuning. See experiment on linear regression below (Alg. 1 = Noisy fine-tuning) 8/n
1
2
3
@ys_alh
Youssef Allouah
10 months
Finding 2. Differential Privacy (DP) implies Unlearningโ€”But is an Overkill DP implies unlearning, but it severely limits model performance. We prove a tight separation between DP and noisy fine-tuning, in terms of how many samples can be deleted at a fixed test loss. 7/n
1
2
2
@ys_alh
Youssef Allouah
10 months
For our analysis, we require access to an approximate global risk minimizer. For convex tasks, this can simply be SGD, with near-linear time complexity (Neel et al., 2021). Such a minimizer also exists for structured non-convex tasks, e.g., PCA, matrix completion. 6/n
1
2
2
@ys_alh
Youssef Allouah
10 months
Finding 1: Noisy Fine-Tuning Works for In-Distribution Forget Data If the forget data comes from the same distribution as the retained data, noisy fine-tuning (see below) is a highly effective and practical unlearning method. 5/n
1
2
2
@ys_alh
Youssef Allouah
10 months
Problem: What Does It Mean to "Unlearn"? We define certified unlearning as making the post-unlearning model statistically near-indistinguishable from a model that was trained without the forget data. This is inspired from differential privacy (Dwork & Roth, 2014). 4/n
1
3
5
@ys_alh
Youssef Allouah
10 months
Our main insight: it depends on the nature of the forget data. If it's in-distribution โ†’ Noisy fine-tuning works great. Differential privacy may be an overkill. If it's out-of-distribution โ†’ (Noisy) Fine-tuning struggles. A robust pre-training strategy is needed. 3/n
1
2
4
@ys_alh
Youssef Allouah
10 months
Key Question: Can we truly remove data from a trained AI modelโ€”without hurting performance or requiring a full retrain? 2/n
1
2
3