Thomas Dehaene Profile
Thomas Dehaene

@TDehaene

Followers
581
Following
1K
Media
92
Statuses
544

NLP engineer and meme sommelier

Belgium
Joined September 2011
Don't wanna be here? Send us removal request.
@satyanadella
Satya Nadella
2 years
We remain committed to our partnership with OpenAI and have confidence in our product roadmap, our ability to continue to innovate with everything we announced at Microsoft Ignite, and in continuing to support our customers and partners. We look forward to getting to know Emmett
4K
14K
88K
@raphaelmilliere
Raphaël Millière
3 years
I asked ChatGPT to rewrite Bohemian Rhapsody to be about the life of a postdoc, and the output was flawless:
135
2K
9K
@ai_fast_track
Farid
3 years
😍 Lama Cleaner: A free and open-source inpainting tool powered by SOTA AI model 🎁 Completely Free and Open-Source β€’ Fully self-hosted β€’ Multiple SOTA AI models β€’ Classical image inpainting algorithm β€’ CPU & GPU β€’ Various inpainting strategy β€’ Run as a Desktop APP
36
468
3K
@TDehaene
Thomas Dehaene
3 years
It's that time of the year again πŸ‘€! https://t.co/AJlWUecwNk Everything to know about the past year of happenings in #AI 🎁. My favorite graph: a showcase of how cross-modal #transformers have become πŸ‘‰
0
0
6
@ClementDelangue
clem πŸ€—
3 years
2
3
30
@TDehaene
Thomas Dehaene
3 years
Note: if the interactive version is down for some reason, you can check out the Medium version:
0
0
1
@TDehaene
Thomas Dehaene
3 years
#Transformers πŸ€– for #summarization are sometimes prone to hallucinations πŸ˜΅β€πŸ’«. Our #NLP team investigated some post-processing steps in an interactive blogpost on @huggingface πŸ‘‰ https://t.co/PSWr5WzGky
1
2
5
@suzatweet
Suzana Ilić
3 years
Machine Learning Operations (MLOps): Overview, Definition, and Architecture https://t.co/2MHAoBjfRN
55
824
4K
@TDehaene
Thomas Dehaene
4 years
@BigScienceLLM
BigScience Large Model Training
4 years
β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“β–“ 99%
0
0
0
@emilymbender
@emilymbender.bsky.social
4 years
For those playing along at home, here's a "AI is sentient!" argument bingo card.
78
506
2K
@NielsRogge
Niels Rogge
4 years
🀩 LayoutLMv3 by @Microsoft is now available @huggingface! πŸ”₯ The model replaces the CNN backbone of its predecessor with (much simpler) patch embeddings Γ  la ViT. πŸš€ SOTA performance on all document AI benchmarks, both image-only and text+image! (1/2)
8
98
541
@TDehaene
Thomas Dehaene
4 years
πŸ’‘ Adapting semantic search to a new domain with unlabeled data πŸ”? Using #GPL from @UKPLab, we demonstrate how to do this on non-English datasets πŸ‡©πŸ‡ͺ . πŸ”—:
Tweet card summary image
blog.ml6.eu
Lexical based information retrieval systems are great for quickly fetching relevant information in a large text corpus. However, these…
0
0
1
@huggingface
Hugging Face
4 years
Introducing the biggest change to https://t.co/dKZjob5Hcx since its inception πŸ”₯ The Community Tab πŸ₯³ read more and discuss: https://t.co/SI07ac47Z1
3
55
185
@TDehaene
Thomas Dehaene
4 years
πŸ“ I wrote an #NLP blogpost detailing how we optimized our Dutch #GPT2 model to reduce our Google Cloud bill by a factor of 2.4 πŸ€‘ ! πŸ‘‰ https://t.co/DCU4XJV7Bt
0
0
6
@TDehaene
Thomas Dehaene
4 years
🚨 new model alert 🚨. We're releasing: ✌️ translated Dutch summarization datasets (CNN-nl & XSUM-nl) βž• a finetuned mBART model for Dutch summarization on @huggingface ! https://t.co/WQjkNsFL2t https://t.co/RUsmTF8ck7
0
1
15
@TDehaene
Thomas Dehaene
4 years
It do be like that with #AI πŸ™ˆ
0
0
9
@TDehaene
Thomas Dehaene
4 years
When your funding round raises about as much dough as a BERT-Base model has parameters πŸ€―πŸ’΅! Big congrats to @huggingface πŸ’ͺ!
1
0
3
@TDehaene
Thomas Dehaene
4 years
Our official @huggingface HuggingMugs (or MuggingFaces πŸ€”) have arrived β˜•οΈ! And already our Rouge scores have improved, our Perplexities are down and my eyesight is better 🀯
1
4
34
@TDehaene
Thomas Dehaene
4 years
The @amazon MASSIVE dataset is... well... massive 😱! One million realistic, parallel, labeled virtual-assistant utterances in 51 languages πŸŽ™οΈπŸŒŽ.
Tweet card summary image
amazon.science
MASSIVE dataset and Massively Multilingual NLU (MMNLU-22) competition and workshop will help researchers scale natural-language-understanding technology to every language on Earth.
0
1
4
@StasBekman
Stas Bekman
4 years
πŸ€— Trainer now sports --optim adamw_bnb_8bit which activates the 8-bit Adam optimizer https://t.co/X1KdSIZZNC and uses 6 bytes less per param for training. ~1/3 of total memory savings! Huge thanks to @ManuelCiosici & @Tim_Dettmers for integrating it! Use transformers@main
Tweet card summary image
github.com
Library for 8-bit optimizers and quantization routines. - facebookresearch/bitsandbytes
1
17
102