Pang Wei Koh Profile
Pang Wei Koh

@PangWeiKoh

Followers
4K
Following
2K
Media
12
Statuses
344

Assistant professor at @uwcse and visiting research scientist at @allen_ai. Formerly @StanfordAILab @GoogleAI @Coursera. šŸ‡øšŸ‡¬

Joined June 2020
Don't wanna be here? Send us removal request.
@PangWeiKoh
Pang Wei Koh
15 days
RT @allen_ai: New updates for olmOCR, our fully open toolkit for transforming documents (PDFs & images) into clean markdown. We released:….
0
40
0
@PangWeiKoh
Pang Wei Koh
15 days
RT @lm4sci: 🚨 Call for Papers: LM4Sci @COLM_conf 2025 🚨. Excited to announce the Large Language Modeling for Scientific Discovery (LM4Sci)….
0
8
0
@PangWeiKoh
Pang Wei Koh
21 days
RT @RulinShao: šŸŽ‰Our Spurious Rewards is available on ArXiv! We added experiments on.- More prompts/steps/models/analysis. - Spurious Prom….
0
40
0
@PangWeiKoh
Pang Wei Koh
24 days
LMs will be much more reliable if we could control the information used in their responses -- a list of approved drugs, facts about a safety recall, etc. -- and not add anything extraneous. Check out @jcqln_h's work on Precise Information Control below!.
@jcqln_h
Jacqueline He
24 days
LMs often output answers that sound right but aren’t supported by input context. This is intrinsic hallucination: the generation of plausible, but unsupported content. We propose Precise Information Control (PIC): a task requiring LMs to ground only on given verifiable claims.
Tweet media one
0
1
14
@PangWeiKoh
Pang Wei Koh
24 days
RT @jcqln_h: LMs often output answers that sound right but aren’t supported by input context. This is intrinsic hallucination: the generati….
0
18
0
@PangWeiKoh
Pang Wei Koh
28 days
RT @niloofar_mire: šŸ“£Thrilled to announce I’ll join Carnegie Mellon University (@CMU_EPP & @LTIatCMU) as an Assistant Professor starting Fal….
0
63
0
@PangWeiKoh
Pang Wei Koh
30 days
RT @yizhongwyz: Thrilled to announce that I will be joining @UTAustin @UTCompSci as an assistant professor in fall 2026! . I will continue….
0
54
0
@PangWeiKoh
Pang Wei Koh
30 days
RT @AkariAsai: ā€˜Bold,’ ā€˜positive’ and ā€˜unparalleled’: Allen School Ph.D. graduates Ashish Sharma and Sewon Min recognized with ACM Doctoral….
0
16
0
@PangWeiKoh
Pang Wei Koh
1 month
RT @zzlccc: We do appreciate their efforts in writing the criticisms, but ā€œturns out that the results in this paper are misreportedā€ is a s….
0
10
0
@PangWeiKoh
Pang Wei Koh
1 month
RT @RulinShao: One more fun thing! .RLVR can elicit existing behaviors like code reasoning. But! If your model is not good at code but thou….
0
24
0
@PangWeiKoh
Pang Wei Koh
1 month
RT @StellaLisy: 🤯 We cracked RLVR with. Random Rewards?!.Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:.- Rando….
0
338
0
@PangWeiKoh
Pang Wei Koh
1 month
Turns out that RL on "verifiable rewards" can work really well even when these rewards are completely random -- but even then, only on some model families! There's still much to understand about RLVR. Check out our analysis on spurious rewards below:.
@StellaLisy
Stella Li
1 month
🤯 We cracked RLVR with. Random Rewards?!.Training Qwen2.5-Math-7B with our Spurious Rewards improved MATH-500 by:.- Random rewards: +21%.- Incorrect rewards: +25%.- (FYI) Ground-truth rewards: + 28.8%.How could this even workā‰ļø Here's why: 🧵.Blogpost:
Tweet media one
1
4
59
@PangWeiKoh
Pang Wei Koh
1 month
RT @rui_xin31: Think PII scrubbing ensures privacy? šŸ¤”Think againā€¼ļø In our paper, for the first time on unstructured text, we show that you….
0
19
0
@PangWeiKoh
Pang Wei Koh
1 month
RT @RulinShao: Super excited to see how ReasonIR data can also help much much smaller models to achieve high reasoning-intensive retrieval….
0
5
0
@PangWeiKoh
Pang Wei Koh
1 month
RT @SitingLi627: Excited to share that our paper "Exploring How Generative MLLMs Perceive More Than CLIP with the Same Vision Encoder" is a….
0
13
0
@PangWeiKoh
Pang Wei Koh
2 months
RT @percyliang: What would truly open-source AI look like? Not just open weights, open code/data, but *open development*, where the entire….
0
193
0
@PangWeiKoh
Pang Wei Koh
2 months
RT @tomchen0: LLMs naturally memorize some verbatim of pre-training data. We study whether post-training can be an effective way to mitigat….
0
31
0
@PangWeiKoh
Pang Wei Koh
2 months
RT @thao_nguyen26: šŸ“¢ Announcing our data-centric workshop at ICML 2025 on unifying data curation frameworks across domains!. šŸ“… Deadline: Ma….
0
21
0
@PangWeiKoh
Pang Wei Koh
2 months
RT @Muennighoff: Reasoning & test-time scaling don't just matter for generating text with LLMs — @RulinShao, @ray_qiaorui & team show how t….
0
8
0
@PangWeiKoh
Pang Wei Koh
2 months
RT @RulinShao: Meet ReasonIR-8B✨the first retriever specifically trained for reasoning tasks! Our challenging synthetic training data unloc….
0
62
0