Somnath Basu Roy Chowdhury Profile
Somnath Basu Roy Chowdhury

@SomnathBrc

Followers
98
Following
192
Media
27
Statuses
52

Research Scientist at Google Research

Joined January 2024
Don't wanna be here? Send us removal request.
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
๐‡๐จ๐ฐ ๐œ๐š๐ง ๐ฐ๐ž ๐ฉ๐ž๐ซ๐Ÿ๐ž๐œ๐ญ๐ฅ๐ฒ ๐ž๐ซ๐š๐ฌ๐ž ๐œ๐จ๐ง๐œ๐ž๐ฉ๐ญ๐ฌ ๐Ÿ๐ซ๐จ๐ฆ ๐‹๐‹๐Œ๐ฌ?. Our method, Perfect Erasure Functions (PEF), erases concepts from LLM representations w/o parameter estimation, achieving pareto optimal erasure-utility tradeoff w/ guarantees. #AISTATS2025 ๐Ÿงต
Tweet media one
2
35
153
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
RT @snigdhac25: @SomnathBrc is at #ICLR2025 presenting our work on perfect unlearning. Find him!.
0
2
0
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
@snigdhac25 (9/n) Iโ€™m attending ICLR in person and presenting our poster on 25th April in Poster session 3 between 10AM-1230PM. Please feel free to stop by our poster if youโ€™re interested. Iโ€™m also happy to chat about unlearning or AI safety in general. cc: @uncnlp @unccs.
0
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(8/n) Finally, I would like to thank all my amazing co-authors: Krzysztof, Arijit, Avinava, and @snigdhac25. Code: Paper link:
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(7/n) Empirically, we observe that SยณT can handle around 2.5x more deletion requests while achieving superior task performance compared to baselines
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(6/n) Theoretically, we show that SยณT achieves a better deletion rate than the existing SOTA exact unlearning technique, SISA. More interestingly, we observe that training the model only on B=L different permutations can achieve the best deletion performance.
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(5/n) When we have access to deletion prior, we generate diverse permutations using the bipartite matching algorithm that considers the prior probabilities.
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(4/n) To prevent this, we propose training the model on multiple sequences (or permutations) of the same data. While functioning within a budget, we generate diverse permutations using cyclic rotation when no prior deletion information is available.
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(3/n) We can easily unlearn data using SยณT by switching off LoRA layers below the affected module. However, when the topmost module is affected in the rare scenario, we may need to retrain from scratch.
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
(2/n) We propose a sequential fine-tuning strategy that trains individual PEFT layers using different data subsets. This helps convert LLMs into a modular system that is helpful in executing exact unlearning.
Tweet media one
1
0
1
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
๐‡๐จ๐ฐ ๐œ๐š๐ง ๐ฐ๐ž ๐ฉ๐ž๐ซ๐Ÿ๐ž๐œ๐ญ๐ฅ๐ฒ ๐ฎ๐ง๐ฅ๐ž๐š๐ซ๐ง ๐๐š๐ญ๐š ๐Ÿ๐ซ๐จ๐ฆ ๐‹๐‹๐Œ๐ฌ ๐ฐ๐ก๐ข๐ฅ๐ž ๐ฉ๐ซ๐จ๐ฏ๐ข๐๐ข๐ง๐  ๐ ๐ฎ๐š๐ซ๐š๐ง๐ญ๐ž๐ž๐ฌ?. We present SยณT, a scalable unlearning framework that guarantees data deletion from LLMs by leveraging parameter-efficient fine-tuning. #ICLR2025 ๐Ÿงต
Tweet media one
1
9
32
@SomnathBrc
Somnath Basu Roy Chowdhury
3 months
RT @abeirami: Finally, if you are also going to #AISTATS2025, @SomnathBrc will be presenting ๐ฉ๐ž๐ซ๐Ÿ๐ž๐œ๐ญ ๐œ๐จ๐ง๐œ๐ž๐ฉ๐ญ ๐ž๐ซ๐š๐ฌ๐ฎ๐ซ๐ž. Somnath will be at Iโ€ฆ.
0
1
0
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(9/n) Finally, I would like to thank all my amazing co-authors: Avinava, @abeirami, Rahul, @nicholasmonath, Amr, @snigdhac25. cc: @uncnlp @unccs.
0
0
3
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(8/n) Here is a blog post with a simplified overview of our work: Code: Paper link:
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(7/n) We would like to highlight previous great works, like LEACE, that perfectly erase concepts to protect against linear adversaries. In our work, we improve upon this method and present a technique that can protect against any adversary.
Tweet media one
@norabelrose
Nora Belrose
2 years
Ever wanted to mindwipe an LLM?. Our method, LEAst-squares Concept Erasure (LEACE), provably erases all linearly-encoded information about a concept from neural net activations. It does so surgically, inflicting minimal damage to other concepts. ๐Ÿงต.
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(6/n) We also visualize the learned representations from different erasure methods. We observe that PEF perfectly erasure group (or concept) information without losing other information (collapsing the representation space).
Tweet media one
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(5/n) Empirically, we observe that PEF reaches the theoretical limits of erasure even in challenging settings where other methods struggle, including both linear (INLP, LEACE) and non-linear techniques (FaRM, KRaM).
Tweet media one
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(4/n) When the distributions are unequal, we still achieve perfect erasure but with a slightly reduced utility. The erasure function in this setting is shown below.
Tweet media one
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(3/n) From the above limits, we show that optimally perfect concept erasure is only feasible when the underlying distributions are equal up to permutations. In such scenarios, the erasure function is shown in the diagram.
Tweet media one
1
0
2
@SomnathBrc
Somnath Basu Roy Chowdhury
4 months
(2/n) We study the fundamental limits of concept erasure. Borrowing from the work of @FlavioCalmon et al. in information theory literature, we characterize the erasure capacity and maximum utility that can be retained during concept erasure.
Tweet media one
1
0
3