Ihor Stepanov Profile
Ihor Stepanov

@ihor_step

Followers
167
Following
2K
Media
11
Statuses
323

I am the CEO and co-founder of Knowledgator. We are advancing the #information_extraction field with #opensource #AI models.

Ukraine
Joined July 2014
Don't wanna be here? Send us removal request.
@ihor_step
Ihor Stepanov
5 days
RT @Cohere_Labs: Don't forget to tune in tomorrow, August 26th as @ihor_step discusses Zero-Shot Named Entity Recognition with GLiNeR. Lea….
0
3
0
@ihor_step
Ihor Stepanov
10 days
RT @Cohere_Labs: Join us for a deep dive into Zero-Shot Named Entity Recognition with GLiNeR presented by @ihor_step on Tuesday, August 26t….
0
6
0
@grok
Grok
20 days
Generate videos in just a few seconds. Try Grok Imagine, free for a limited time.
536
678
4K
@ihor_step
Ihor Stepanov
15 days
RT @knowledgator: 🚀 GLiNER x SmolLM: a new joint encoder-decoder architecture 🚀. We are excited to release a new kind of GLiNER model built….
0
4
0
@ihor_step
Ihor Stepanov
19 days
RT @knowledgator: 🚀 Our largest study on zero-shot text classification is out!.📄 We surpass cross-encoders while b….
0
3
0
@ihor_step
Ihor Stepanov
27 days
RT @ClementDelangue: Every tech company can and should train their own deepseek R1, Llama or GPT5, just like every tech company writes thei….
0
280
0
@ihor_step
Ihor Stepanov
1 month
RT @gm8xx8: GLiClass-V3: A family of encoder-only models that match or exceed DeBERTa-v3-Large in zero-shot accuracy, while delivering up t….
0
5
0
@ihor_step
Ihor Stepanov
1 month
RT @knowledgator: 🚀 Introducing GLiClass‑V3 – a leap forward in zero-shot classification!. Matches or beats cross-encoder accuracy, while b….
Tweet card summary image
huggingface.co
0
20
0
@ihor_step
Ihor Stepanov
1 month
RT @antoine_chaffin: Seeing ModernBERT and Ettin models being useful is heart warming.
0
3
0
@ihor_step
Ihor Stepanov
1 month
RT @GoogleDeepMind: An advanced version of Gemini with Deep Think has officially achieved gold medal-level performance at the International….
0
775
0
@ihor_step
Ihor Stepanov
2 months
🤔 Which task do you think is more complex for ML models?. 🔹 Multi-label classification .🔹 Entity recognition / Object detection (e.g., in computer vision). Vote and share your thoughts! 👇.
0
0
2
@ihor_step
Ihor Stepanov
2 months
RT @knowledgator: 🚀 Introducing GLiNER-X: Breaking Language Barriers in Multilingual Zero-Shot NER! 🌍. We’re excited to roll out our latest….
0
3
0
@ihor_step
Ihor Stepanov
3 months
RT @knowledgator: 🚀 GLiNER just reached 2K stars on GitHub — and it’s now 3x faster!. We’re thrilled to see this project grow into a core t….
0
1
0
@ihor_step
Ihor Stepanov
4 months
RT @knowledgator: 🎉 Thanks to @GoogleStartups Ukraine Support Fund, we are receiving funding + long-term mentorship from @Google to grow ou….
0
1
0
@ihor_step
Ihor Stepanov
4 months
RT @ClementDelangue: Biased but in my opinion ZeroGPU is one of the most impressive piece of infra in AI that no one is talking about. Po….
0
36
0
@ihor_step
Ihor Stepanov
4 months
RT @Dorialexander: Breaking: @pleiasfr releases a new generation of small reasoning models for RAG and source synthesis. Pleias-RAG-350M an….
0
107
0
@ihor_step
Ihor Stepanov
5 months
I just finished my small experiments comparing different encoder models on retrieval tasks. The goal was to check whether MLM is better than RTD for these tasks. I compared Electra's small models, both generator and discriminator, that have the same size. Additionally, it was
Tweet media one
5
4
27
@ihor_step
Ihor Stepanov
5 months
It became popular to think that architecture is no longer a thing; we achieved some local minimum in architectural design, and data right now is the limiting factor. This work empirically shows that this is not the case, that these aspects are more interesting and hide deeper.
@wissam_antoun
Wissam Antoun
5 months
ModernBERT or DeBERTaV3?. What's driving performance: architecture or data?. To find out we pretrained ModernBERT on the same dataset as CamemBERTaV2 (a DeBERTaV3 model) to isolate architecture effects. Here are our findings:.
0
0
6
@ihor_step
Ihor Stepanov
5 months
RT @gm8xx8: ModernBERT vs DeBERTaV3. To isolate architecture from pretraining effects, ModernBERT is trained on the same dataset as CamemBE….
0
25
0
@ihor_step
Ihor Stepanov
5 months
RT @knowledgator: ⚡ Introducing truly Flash DeBERTa implementation ⚡. DeBERTa remains a top-performing model in Named Entity Recognition, T….
Tweet card summary image
github.com
Trully flash implementation of DeBERTa disentangled attention mechanism. - Knowledgator/FlashDeBERTa
0
18
0
@ihor_step
Ihor Stepanov
5 months
RT @sijun_tan: Hey @sama, we know you're planning to open-source your reasoning model—but we couldn’t wait. Introducing DeepCoder-14B-Prev….
0
148
0