Pradeep Dasigi Profile
Pradeep Dasigi

@pdasigi

Followers
1K
Following
685
Media
3
Statuses
446

Senior Research Scientist @allen_ai; #NLProc, Post-training for OLMo

Seattle, WA
Joined February 2009
Don't wanna be here? Send us removal request.
@pdasigi
Pradeep Dasigi
15 days
RT @valentina__py: 💡Beyond math/code, instruction following with verifiable constraints is suitable to be learned with RLVR. But the set of….
0
92
0
@pdasigi
Pradeep Dasigi
2 months
RT @JesseDodge: @percyliang @AiEleuther nice! we also recently trained a set of models on 25 different pretraining corpora, each corpus hav….
0
7
0
@pdasigi
Pradeep Dasigi
3 months
RT @allen_ai: We're excited to round out the OLMo 2 family with its smallest member, OLMo 2 1B, surpassing peer models like Gemma 3 1B or L….
0
100
0
@pdasigi
Pradeep Dasigi
4 months
RT @natolambert: A very exciting day for open-source AI! We're releasing our biggest open source model yet -- OLMo 2 32B -- and it beats th….
0
150
0
@pdasigi
Pradeep Dasigi
4 months
RT @allen_ai: Announcing OLMo 2 32B: the first fully open model to beat GPT 3.5 & GPT-4o mini on a suite of popular, multi-skill benchmarks….
0
159
0
@pdasigi
Pradeep Dasigi
5 months
RT @srush_nlp: Really interesting paper on data selection. At this point just read everything Hamish writes.
0
18
0
@pdasigi
Pradeep Dasigi
5 months
How to curate instruction tuning datasets while targeting specific skills? This is a common question developers face while post-training LMs. In this work led by @hamishivi, we found that simple embedding based methods scale much better than fancier computationally intensive.
@hamishivi
Hamish Ivison
5 months
How well do data-selection methods work for instruction-tuning at scale?. Turns out, when you look at large, varied data pools, lots of recent methods lag behind simple baselines, and a simple embedding-based method (RDS) does best!. More below ⬇️ (1/8)
Tweet media one
0
0
12
@pdasigi
Pradeep Dasigi
5 months
RT @allen_ai: Introducing olmOCR, our open-source tool to extract clean plain text from PDFs!. Built for scale, olmOCR handles many documen….
0
266
0
@pdasigi
Pradeep Dasigi
5 months
RT @HannaHajishirzi: Excited to drive innovation and push the boundaries of open, scientific AI research & development! 🚀 Join us at @allen….
0
15
0
@pdasigi
Pradeep Dasigi
5 months
Thanks a lot for having me!.
@savvyRL
Rosanne Liu
5 months
Tülu 3 is being presented at DLCT right now! Led by @pdasigi . The @allen_ai color with 🩷 as author superscript are basically shouting "happy valentine's"
Tweet media one
0
0
9
@pdasigi
Pradeep Dasigi
6 months
RT @hamishivi: One additional thing in the updated Tulu 3 paper that I'd like to highlight is that @pdasigi went back and re-evaluated our….
0
1
0
@pdasigi
Pradeep Dasigi
6 months
Here's a significant update to Tülu 3: we scaled up the post-training recipe to Llama 3.1 405B. Tülu 3 405B beats Llama's 405B instruct model and also Deepseek V3. You can now access the model and the entire post-training pipeline. Huge shoutout to @hamishivi and @vwxyzjn who.
@allen_ai
Ai2
6 months
Here is Tülu 3 405B 🐫 our open-source post-training model that surpasses the performance of DeepSeek-V3! The last member of the Tülu 3 family demonstrates that our recipe, which includes Reinforcement Learning from Verifiable Rewards (RVLR) scales to 405B - with performance on
Tweet media one
0
0
12
@pdasigi
Pradeep Dasigi
7 months
RT @allen_ai: Remember Molmo? The full recipe is finally out!. Training code, data, and everything you need to reproduce our models. Oh, an….
0
91
0
@pdasigi
Pradeep Dasigi
7 months
RT @faeze_brh: Just arrived in 🇨🇦 to attend NeurIPS 2024!. Excited to connect and chat about AI reliability and safety, resource-efficient….
0
3
0
@pdasigi
Pradeep Dasigi
8 months
RT @saumyamalik44: I'm having a great time as a PYI at Ai2! Consider applying for this great program :).
0
10
0
@pdasigi
Pradeep Dasigi
8 months
Our team at Ai2 (OLMo) is looking for a predoctoral researcher. You get to work on exciting research in building open LMs while preparing for a PhD. Apply here:
job-boards.greenhouse.io
0
0
2
@pdasigi
Pradeep Dasigi
8 months
RT @interconnectsai: OpenAI's o1 using "search" was a PSYOP.How to understand OpenAI's o1 models as really just one wacky, wonderful, long….
Tweet media one
www.interconnects.ai
How to understand OpenAI's o1 models as really just one wacky, wonderful, long chain of thought
0
22
0
@pdasigi
Pradeep Dasigi
8 months
RT @allen_ai: Calling all predoctoral candidates: our OLMo team is hiring! Apply to be a Predoctoral Young Investigator today at the link i….
0
14
0
@pdasigi
Pradeep Dasigi
8 months
RT @vwxyzjn: 😆 So happy OLMo 2 is out! We applied the same Tülu 3 RLVR recipe and it worked very nicely for our final 13B instruct model.….
0
4
0
@pdasigi
Pradeep Dasigi
8 months
RT @valentina__py: Applying our Tülu 3 recipe to OLMo 2 resulted in OLMo 2 Instruct, which is competitive with even the best open-weight mo….
0
1
0