
Nicolas Boizard
@N1colAIs
Followers
166
Following
156
Media
16
Statuses
86
NLP - AI | 25y PhD student @CentraleSupelec | EuroBERT 🇪🇺 & Universal Logit Distillation Loss ⚗️ & CroissantLLM 🥐
Paris, France
Joined December 2023
EuroBERT is going to @COLM_conf 2025! Can’t wait to be in Montreal with @gisship and @DuarteMRAlves to see all the great research everyone’s bringing!
0
4
22
RT @ManuelFaysse: 🚨Should We Still Pretrain Encoders with Masked Language Modeling? We have recently seen massively trained causal decoders….
0
37
0
More in the paper, go check it out! You can also check out our blog post: Huge thanks to the dream team 💥.@gisship , @ManuelFaysse , @DuarteMRAlves , Emmanuel Malherbe, @andre_t_martins , Céline Hudelot, @PierreColombo6 🙌. 7/7 🧵.
0
0
0
So nice to see new distillation libraries emerging around the ULD Loss and other cross-tokenizer distillation methods. Big kudos to @vitransformer and the whole DistillKitPlus team for their work.
its been an exciting few weeks finding my footing @lossfunk 🥳 and working on DistillKitPlus (with @AmanGokrani ), a toolkit for making smaller models learn from larger ones through efficient logits distillation. (check it out @ . some recent updates:.
0
0
2
• Optimized: Fused ops (Liger Kernel, Flash), distributed training (FSDP, DDP). • Data Processing: Tokenization, packing, subsampling. • Highly Customisable and more. For any large-scale continuous pre-training feel free to contact me or @DuarteMRAlves for extensive support 🤗.
0
0
1
RT @antoine_chaffin: PyLate 1.1.7 - The big batches update - is out! 🚀.As contrastive learning remains the go-to for large-scale pre-traini….
0
13
0
Great to see the community using EuroBERT! As hoped, it’s proving to be an excellent foundation model, especially for information retrieval tasks across multiple languages after just one epoch of finetuning. Check it out: .@Engomar_10
1
1
5