Dang Nguyen Profile
Dang Nguyen

@dangnth97

Followers
327
Following
5K
Media
0
Statuses
105

PhD Candidate @CS_UCLA | IMO 2015 Silver

Los Angeles, CA, USA
Joined March 2023
Don't wanna be here? Send us removal request.
@dangnth97
Dang Nguyen
17 days
RT @jaseweston: šŸ¤–Introducing OptimalThinkingBench šŸ¤–.šŸ“: - Thinking LLMs use a lot of tokens & overthink; non-thinkin….
0
72
0
@dangnth97
Dang Nguyen
18 days
RT @pratyushmaini: 1/Pretraining is hitting a data wall; scaling raw web data alone leads to diminishing returns. Today @datologyai shares….
0
125
0
@dangnth97
Dang Nguyen
2 months
RT @lyang36: 🚨 Olympiad math + AI:. We ran Google’s Gemini 2.5 Pro on the fresh IMO 2025 problems. With careful prompting and pipeline desi….
0
117
0
@dangnth97
Dang Nguyen
2 months
RT @lmthang: Very excited to share that an advanced version of Gemini Deep Think is the first to have achieved gold-medal level in the Inte….
0
233
0
@dangnth97
Dang Nguyen
2 months
RT @tungnd_13: šŸš€ Introducing PhysiX: One of the first large-scale foundation models for physics simulations!. PhysiX is a 4.5B parameter mo….
0
255
0
@dangnth97
Dang Nguyen
2 months
RT @nouhadziri: šŸ“¢ Can LLMs really reason outside the box in math? Or are they just remixing familiar strategies? . Remember DeepSeek R1, o1….
0
156
0
@dangnth97
Dang Nguyen
3 months
RT @RyanHuynh1108: CVPR-bound! āœˆļø. I'll be presenting CoLLM on Friday, 6/13 (Morning, #364) and looking for my next challenge as a full-tim….
0
6
0
@dangnth97
Dang Nguyen
3 months
RT @xue_yihao65785: šŸŽ‰ Our paper ā€œRepresentations Shape Weak-to-Strong Generalizationā€ is accepted at #ICML2025!.We study weak-to-strong gen….
0
8
0
@dangnth97
Dang Nguyen
3 months
RT @xuandongzhao: šŸš€ Excited to share the most inspiring work I’ve been part of this year:. "Learning to Reason without External Rewards"….
0
510
0
@dangnth97
Dang Nguyen
5 months
RT @sjoshi804: #ICLR2025.Can you pre-train deep models with small, synthetic datasets? 🤯. We introduce the first effective dataset distilla….
0
2
0
@dangnth97
Dang Nguyen
5 months
šŸŽ‰ Achievement unlocked: having papers with all of my labmates and somehow all ended up at ICLR!. I’ll be presenting our work ā€œMini-batch Coresets for Memory-efficient Language Model Training on Data Mixturesā€ at #ICLR2025 šŸ‡øšŸ‡¬. Come by and chat! šŸ‘‹ on Fri, Apr 25 | 10 AM GMT+8.
0
2
28
@dangnth97
Dang Nguyen
6 months
RT @Yihe__Deng: šŸš€Excited to share our latest work: OpenVLThinker, an exploration into enhancing vision-language models with R1 reasoning ca….
0
37
0
@dangnth97
Dang Nguyen
6 months
RT @Yihe__Deng: šŸ¤– I just updated my repository of RL(HF) summary notes to include a growing exploration of new topics, specifically adding….
0
13
0
@dangnth97
Dang Nguyen
7 months
RT @duynguyen772: LLMs must be helpful, unbiased, etc. but optimizing for one attribute can hurt others. šŸš€ We introduce MAT-Steer for st….
0
28
0
@dangnth97
Dang Nguyen
7 months
RT @lmthang: Excited to share details of AlphaGeometry2 (AG2), part of the system that achieved silver-medal standard at IMO 2024 last July….
0
181
0
@dangnth97
Dang Nguyen
7 months
RT @Yihe__Deng: New paper & model release!. Excited to introduce DuoGuard: A Two-Player RL-Driven Framework for Multilingual LLM Guardrails….
0
30
0
@dangnth97
Dang Nguyen
7 months
RT @TheGradient: (1/2) Ever wondered why Sharpness-Aware Minimization (SAM) yields greater generalization gains in vision than in NLP? I'll….
0
10
0
@dangnth97
Dang Nguyen
7 months
RT @natolambert: Since everyone wants to learn RL for language models now post DeepSeek, reminder that I've been working on this book quiet….
0
167
0
@dangnth97
Dang Nguyen
7 months
RT @WenhuChen: Everyone is talking about RL these days. But are we done with SFT? The answer is NO. If we revive SFT in another form, it ca….
0
97
0
@dangnth97
Dang Nguyen
8 months
RT @behrouz_ali: Attention has been the key component for most advances in LLMs, but it can’t scale to long context. Does this mean we need….
0
597
0