
Sewon Min
@sewon__min
Followers
12K
Following
3K
Media
12
Statuses
1K
Assistant professor @Berkeley_EECS @berkeley_ai || Research scientist at @allen_ai || PhD from @uwcse @uwnlp
Seattle, WA
Joined November 2017
It has been great working on the project with support from @allen_ai! I believe there are many meaningful ways different people and orgs can work together to build strong shared models, and data collaboration might be the most impactful form of it. đź“„Paper:.
Introducing FlexOlmo, a new paradigm for language model training that enables the co-development of AI through data collaboration. đź§µ
6
12
189
RT @WeijiaShi2: Can data owners & LM developers collaborate to build a strong shared model while each retaining data control?. Introducing….
0
78
0
Thanks, @willknight, for covering our work!!.
New on @WIRED: A novel type of distributed mixture-of-experts model from Ai2 (called FlexOlmo) allows data can be contributed to a frontier model confidentially, and even revoked after the model is built: .
1
3
22
RT @RulinShao: 🚀 Last year: MassiveDS-1.4T showed great scaling gains with a web-scale datastore but was too heavy for online production.✨….
0
7
0
RT @XinxiLyu: Reasoning benchmarks (e.g., MMLU Pro and GPQA) have seen little benefit from naive RAG. But can we flip this?.🔥Introducing Co….
0
15
0
RT @aryg18: fwiw, I think Prof. @percyliang and the CS336 team nailed this:. Sutton’s Bitter Lesson is often misinterpreted as “scale is al….
0
27
0
RT @svlevine: I always found it puzzling how language models learn so much from next-token prediction, while video models learn so little f….
0
172
0
RT @AiEleuther: Can you train a performant language models without using unlicensed text?. We are thrilled to announce the Common Pile v0.1….
0
131
0
RT @TheOfficialACM: 🎓 Congrats to Ashish Sharma, @UW on receiving the ACM Doctoral Dissertation Award for his dissertation, "Human-AI Colla….
0
11
0
RT @uwcse: Congratulations to @UW #UWAllen Ph.D. grads @sharma_ashish_2 & @sewon__min, @TheOfficialACM Doctoral Dissertation Award honorees….
0
18
0
RT @yizhongwyz: Thrilled to announce that I will be joining @UTAustin @UTCompSci as an assistant professor in fall 2026! . I will continue….
0
54
0
RT @lateinteraction: @TensorTemplar The important breakthrough is that a lot of the “RL just works” noise has little to do with RL and has….
0
3
0
RT @StellaLisy: 🤯 We cracked RLVR with. Random Rewards?!.Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:.- Rando….
0
337
0
RT @MKhalifaaaa: 🚨Announcing SCALR @ COLM 2025 — Call for Papers!🚨. The 1st Workshop on Test-Time Scaling and Reasoning Models (SCALR) is c….
0
17
0
RT @allen_ai: We’re live on Reddit! Ask us Anything about our OLMo family of models. We have six of our researchers on hand to answer all y….
0
25
0
RT @RulinShao: Meet ReasonIR-8B✨the first retriever specifically trained for reasoning tasks! Our challenging synthetic training data unloc….
0
63
0
RT @infoxiao: using a fraction of the compute and building best-in-class models is such aura @danqi_chen . #ICLR25
0
7
0
RT @berkeley_ai: BAIR faculty Stuart Russell, Dan Klein, @alsuhr, @Ken_Goldberg, and @sewon__min weigh in on the future of LLMs, synthetic….
0
5
0
RT @Tim_Dettmers: Happy to announce that I joined the CMU Catalyst with three of my incoming students. Our research will bring the best m….
0
54
0