ClemDelangue Profile Banner
Clem Delangue 🤗 Profile
Clem Delangue 🤗

@ClemDelangue

Followers
1K
Following
85
Media
16
Statuses
68

Co-founder & CEO at Hugging Face 🤗. We teach computers to understand human language.

Brooklyn, NY
Joined October 2018
Don't wanna be here? Send us removal request.
@soumithchintala
Soumith Chintala
6 years
The first full paper on @pytorch after 3 years of development. It describes our goals, design principles, technical details uptil v0.4 Catch the poster at #NeurIPS2019 Authored by @apaszke , @colesbury et. al. https://t.co/XFyX0qt1RH
13
410
2K
@Thom_Wolf
Thomas Wolf
6 years
Interesting work (and a nice large and clean dataset as well, looking forward to see it released): "Compressive Transformers for Long-Range Sequence Modelling" by Jack W. Rae, Anna Potapenko, Siddhant M. Jayakumar, Timothy P. Lillicrap (at DeepMind) Paper: https://t.co/CV3ThAAweg
1
78
294
@algo_diver
chansung
6 years
Some more results. Now I made it fully supported by all kinds of model and vocabs. Good experience to use @huggingface with @SlackHQ. And it looks pretty smart
0
1
3
@Thom_Wolf
Thomas Wolf
6 years
The @SustaiNLP2020 workshop at #EMNLP2020 will try to remove a little bit of SOTA addiction from NLP research 😉 We'll promote sensible trade-offs between performances & models that are - computationally more efficient - conceptually simpler ... [1/2] https://t.co/wQQyl75lnP
1
34
255
@fchollet
François Chollet
6 years
Perhaps a great opportunity to use @huggingface's TF 2.0 Transformer implementations :)
1
9
52
@timothy_lkh_
Timothy Liu
6 years
Happy to have a small PR accepted to the HuggingFace Transformer library demonstrating substantial mixed precision speed-up with @NVIDIA Tensor Core #GPU even at small batch size in the demo script
1
3
47
@julien_c
Julien Chaumond
6 years
GPT-2 on device is blazing fast on iPhone 11 ⚡️ Core ML 3 is officially out so we can do state-of-the-art text generation on mobile (117M parameters running ~3 times par second on the neural engine!) We put together a small video benchmark ⬇️
9
152
598
@kyoun
Kyosuke Nishida
6 years
DistilBERT (huggingface) BERT baseから蒸留にて6層に小型化(40%減)。推論は60%高速化、精度はGLUEで95%程度保持。8個の16GB V100 GPUで3.5日ぐらいで学習。hidden sizeは768のままで、層数の方が高速化には効果があるとのこと。github https://t.co/QX6002E2te blog
1
23
80
@ClemDelangue
Clem Delangue 🤗
6 years
.@julien_c watching the @Apple keynote
0
0
0
@julien_c
Julien Chaumond
6 years
1,060 days ago, @Thom_Wolf and I launched a Deep learning for NLP study group: https://t.co/3vdOQGdVTg
Tweet card summary image
medium.com
A remote study group to Stanford’s CS224d “Deep learning for NLP” class
5
16
101
@huggingface
Hugging Face
6 years
💃PyTorch-Transformers 1.1.0 is live💃 It includes RoBERTa, the transformer model from @facebookai, current state-of-the-art on the SuperGLUE leaderboard! Thanks to @myleott @julien_c @LysandreJik and all the 100+ contributors!
5
191
631
@Thom_Wolf
Thomas Wolf
6 years
A question I get from time to time is how to convert a pretrained TensorFlow model in PyTorch easily and reliably. We're starting to be quite familiar with the process so I've written a short blog post summarizing our workflow and some lessons learned 👇 https://t.co/d8ZMs30nGq
Tweet card summary image
medium.com
Friends and users of our open-source tools are often surprised how fast 🚀 we reimplement the latest SOTA pretrained TensorFlow models to…
8
117
496
@Thom_Wolf
Thomas Wolf
6 years
New release of Transformers repo is shaping up & I'm very excited! Gifts for all: -SOTA Lovers: new XLNet & XLM archi + 6 new Bert/GPT trained chkpt -Research Lovers: unified model API, attention/hidden-state outputs to swap/study models -Speed Lovers: Torchscript & head pruning!
4
111
500
@julien_c
Julien Chaumond
6 years
🔥 Thrilled to release our Swift Core ML implementation of BERT for question answering.🔥🔥 Transformers models now also live on the edge. 📱📲 You now CAN do state-of-the-art NLP on mobile devices! https://t.co/YwDkxWftD4 Built w/ @LysandreJik and @Thom_Wolf at @huggingface
5
91
320
@ClemDelangue
Clem Delangue 🤗
7 years
Best Long Paper #naacl2019 BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova #NLProc
0
3
18
@julien_c
Julien Chaumond
7 years
Welcome to Minne-SOTA #NAACL2019
0
3
11
@ClemDelangue
Clem Delangue 🤗
7 years
They’re very big fans of @Thom_Wolf here at #NAACL2019
0
0
1
@soldni
Luca Soldaini 🌯 NeurIPS 2025
7 years
Absolutely PACKED room for @seb_ruder, @Thom_Wolf, @swabhz, and @mattthemathman’s tutorial on transfer learning for NLP #NAACL2019
0
11
83