Clem Delangue 🤗
@ClemDelangue
Followers
1K
Following
85
Media
16
Statuses
68
Co-founder & CEO at Hugging Face 🤗. We teach computers to understand human language.
Brooklyn, NY
Joined October 2018
The first full paper on @pytorch after 3 years of development. It describes our goals, design principles, technical details uptil v0.4 Catch the poster at #NeurIPS2019 Authored by @apaszke , @colesbury et. al. https://t.co/XFyX0qt1RH
13
410
2K
Interesting work (and a nice large and clean dataset as well, looking forward to see it released): "Compressive Transformers for Long-Range Sequence Modelling" by Jack W. Rae, Anna Potapenko, Siddhant M. Jayakumar, Timothy P. Lillicrap (at DeepMind) Paper: https://t.co/CV3ThAAweg
1
78
294
Some more results. Now I made it fully supported by all kinds of model and vocabs. Good experience to use @huggingface with @SlackHQ. And it looks pretty smart
0
1
3
The @SustaiNLP2020 workshop at #EMNLP2020 will try to remove a little bit of SOTA addiction from NLP research 😉 We'll promote sensible trade-offs between performances & models that are - computationally more efficient - conceptually simpler ... [1/2] https://t.co/wQQyl75lnP
1
34
255
Perhaps a great opportunity to use @huggingface's TF 2.0 Transformer implementations :)
1
9
52
GPT-2 on device is blazing fast on iPhone 11 ⚡️ Core ML 3 is officially out so we can do state-of-the-art text generation on mobile (117M parameters running ~3 times par second on the neural engine!) We put together a small video benchmark ⬇️
9
152
598
DistilBERT (huggingface) BERT baseから蒸留にて6層に小型化(40%減)。推論は60%高速化、精度はGLUEで95%程度保持。8個の16GB V100 GPUで3.5日ぐらいで学習。hidden sizeは768のままで、層数の方が高速化には効果があるとのこと。github https://t.co/QX6002E2te blog
1
23
80
1,060 days ago, @Thom_Wolf and I launched a Deep learning for NLP study group: https://t.co/3vdOQGdVTg
medium.com
A remote study group to Stanford’s CS224d “Deep learning for NLP” class
5
16
101
💃PyTorch-Transformers 1.1.0 is live💃 It includes RoBERTa, the transformer model from @facebookai, current state-of-the-art on the SuperGLUE leaderboard! Thanks to @myleott @julien_c @LysandreJik and all the 100+ contributors!
5
191
631
A question I get from time to time is how to convert a pretrained TensorFlow model in PyTorch easily and reliably. We're starting to be quite familiar with the process so I've written a short blog post summarizing our workflow and some lessons learned 👇 https://t.co/d8ZMs30nGq
medium.com
Friends and users of our open-source tools are often surprised how fast 🚀 we reimplement the latest SOTA pretrained TensorFlow models to…
8
117
496
New release of Transformers repo is shaping up & I'm very excited! Gifts for all: -SOTA Lovers: new XLNet & XLM archi + 6 new Bert/GPT trained chkpt -Research Lovers: unified model API, attention/hidden-state outputs to swap/study models -Speed Lovers: Torchscript & head pruning!
4
111
500
🔥 Thrilled to release our Swift Core ML implementation of BERT for question answering.🔥🔥 Transformers models now also live on the edge. 📱📲 You now CAN do state-of-the-art NLP on mobile devices! https://t.co/YwDkxWftD4 Built w/ @LysandreJik and @Thom_Wolf at @huggingface
5
91
320
Best Long Paper #naacl2019 BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. Jacob Devlin, Ming-Wei Chang, Kenton Lee and Kristina Toutanova #NLProc
0
3
18
Absolutely PACKED room for @seb_ruder, @Thom_Wolf, @swabhz, and @mattthemathman’s tutorial on transfer learning for NLP #NAACL2019
0
11
83