Filip Graliński Profile
Filip Graliński

@FilipGralinski

Followers
110
Following
60
Media
45
Statuses
203

6502 and Haskell hacker, machine learner, hypopolyglot (many languages, all poor), opposite Pole, skeptical forteanist

Joined September 2021
Don't wanna be here? Send us removal request.
@Snowflake
Snowflake
3 months
Day 2 of #SnowflakeSummit flew by but not before a mountain of announcements from our Platform Keynote! We announced: Adaptive Compute, Snowflake Openflow, Cortex AISQL, Semantic Model Sharing, Snowflake Intelligence, and much more. See what's new: https://t.co/XMRU4FdlU8
3
3
14
@LukaszBorchmann
Łukasz Borchmann
6 months
How can the most accurate SQL be generated for a given question? We propose a method to significantly boost text-to-SQL accuracy while drastically cutting costs.👇 #NLProc #AI #TextToSQL #LLMs
Tweet media one
2
3
56
@datta_cs
Anupam Datta
6 months
Our Snowflake AI Research team just released Arctic Embed’s core training code into the open source ArcticTraining project — making it easier for developers and researchers to reproduce, fine-tune, and build on our embedding models. Arctic Embed is the leading small embedding
4
5
15
@lukemerrick_
Luke Merrick
9 months
Connor Shorten was kind enough to give me the mic for a lot of hot takes on text embedding models in the latest Weaviate podcast.
@CShorten30
Connor Shorten
9 months
Arctic Embed ❄️ has been one of the most impactful open-source text embedding models! In addition to the open model, which has helped a lot of companies kick off their own inference and fine-tuning services (including us), the Snowflake team has also published incredible research
Tweet media one
1
4
11
@AurickQ
Aurick Qiao
9 months
We are excited to share SwiftKV, our recent work at @SnowflakeDB AI Research! SwiftKV reduces the pre-fill compute for enterprise LLM inference by up to 2x, resulting in higher serving throughput for input-heavy workloads. 🧵
Tweet media one
4
16
43
@spacemanidol
Daniel Campos
9 months
🚀 I am thrilled to introduce @SnowflakeDB 's Arctic Embed 2.0 embedding models! 2.0 offers high-quality multilingual performance with all the greatness of our prior embedding models (MRL, Apache-2 license, great English retrieval, inference efficiency) https://t.co/hEcd0niVyr🌍
snowflake.com
8
27
79
@FilipGralinski
Filip Graliński
10 months
Tweet media one
1
2
2
@MichaPietruszka
Michał Pietruszka
11 months
Can AI models help us create better models? 🧵 1/ It's a question that stands at the boundaries of what's possible in data science. We explored how Large Language Models (LLMs) perform as data scientists, especially in the art of feature engineering.
Tweet media one
1
1
3
@poznanAI
Department of Artificial Intelligence AMU Poznan
1 year
A joint study by @poznanAI researchers and Samsung Electronics Polska engineers was presented at @FedCSIS 2024. The paper investigates the impact of augmenting spoken language corpora with domain-specific synthetic samples. https://t.co/NhkDOc7xI2
Tweet media one
0
3
4
@FilipGralinski
Filip Graliński
1 year
Good people out there, please make your Python script more command-line friendly: 1. put this as the first line: #!/usr/bin/env python3 2. set x permission: chmod u+x your_script.py (and commit that to git) Now you I can run your script with ./your_script.py. Thank you!
0
0
1
@spacemanidol
Daniel Campos
1 year
It's fall which means it's intern recruitment time! @SnowflakeDB is widely recruiting research interns to work on all kinds of problems around AI/LLM/Search. If you are interested or know any students who are looking for summer 2025 internships hit me up!
1
9
31
@YuxiangWei9
Yuxiang Wei
1 year
Code LLMs involve multiple stages of training. At Snowflake, we did extensive training ablations across general repo data, high quality filtered data, and synthetic instruction data so you don’t have to. 🧵
Tweet media one
1
6
32
@dk21
Darek Kłeczek
1 year
Pretty wild that @kaggle got me on the cover! Thanks @InezOkulska for the interview! 😻
Tweet media one
Tweet media two
9
4
59
@poznanAI
Department of Artificial Intelligence AMU Poznan
1 year
LLM Bielik v2 on our internal benchmark, based on Polish educational and professional tests, achieves an accuracy score of 58.03%. This is a noticeable improvement over the 41,51% in v.0.1. Congratulations to the entire @Speak_Leash team. More extensive results coming 🔜
0
1
2
@FilipGralinski
Filip Graliński
1 year
This new LLM for Polish looks really interesting, congrats to the team!
@Speak_Leash
SpeakLeash
1 year
The wait is over - Bielik v2 is here!🦅 Here’s what it offers: 💪11B parameters 📈32,768 token context window 🚝Enhanced training data ⌨Improved NLP 🤝Flexible deployment Made possible through our collaboration with @Cyfronet Check it out here:
0
0
1
@FilipGralinski
Filip Graliński
1 year
Some lessons (I) learnt preparing the data mixture for Snowflake Arctic LLM 👨‍🍳
Tweet media one
1
2
9