Hao Peng Profile
Hao Peng

@haopeng_nlp

Followers
604
Following
39
Media
0
Statuses
38

Assistant Professor at UIUC CS

Joined October 2020
Don't wanna be here? Send us removal request.
@haopeng_nlp
Hao Peng
2 months
RT @charlesfornlp: So many works talking about entropy, but what is the **mechanism** of entropy in RL for LLMs? šŸ¤”. Our work gives a princi….
0
18
0
@haopeng_nlp
Hao Peng
2 months
RT @Shivamag12: Can entropy minimization alone improve LLM performance? And how far can they go without any labeled data? This work answers….
0
65
0
@haopeng_nlp
Hao Peng
2 months
RT @saagnikkk: 🚨 Paper Alert: ā€œRL Finetunes Small Subnetworks in Large Language Modelsā€. From DeepSeek V3 Base to DeepSeek R1 Zero, a whopp….
0
126
0
@haopeng_nlp
Hao Peng
8 months
RT @zhaofeng_wu: šŸ’”We find that models ā€œthinkā€ šŸ’­ in English (or in general, their dominant language) when processing distinct non-English or….
0
61
0
@haopeng_nlp
Hao Peng
8 months
RT @AkariAsai: 🚨 I’m on the job market this year! 🚨.I’m completing my @uwcse Ph.D. (2025), where I identify and tackle key LLM limitations….
0
118
0
@haopeng_nlp
Hao Peng
8 months
RT @OfirPress: I'm on the academic job market! .I develop autonomous systems for: programming, research-level question answering, finding s….
0
39
0
@haopeng_nlp
Hao Peng
8 months
RT @lifan__yuan: Wanna train PRMs but process labels, annotated manually or automatically, sound too expensive to youšŸ˜–? .Introduce Implicit….
0
48
0
@haopeng_nlp
Hao Peng
9 months
RT @bingyikang: Curious whether video generation models (like #SORA) qualify as world models?. We conduct a systematic study to answer this….
0
212
0
@haopeng_nlp
Hao Peng
1 year
RT @MKhalifaaaa: What If LLMs can cite the pre-training source(s) supporting their parametric knowledge? Won't this dramatically improve ve….
Tweet card summary image
arxiv.org
Large language models (LLMs) learn a vast amount of knowledge during pretraining, but they are often oblivious to the source(s) of such knowledge. We investigate the problem of intrinsic source...
0
15
0
@haopeng_nlp
Hao Peng
1 year
RT @YangyiChen6666: šŸŽÆ Introducing SOLO, a single Transformer architecture for unified vision-language modeling. SOLO accepts both raw image….
0
53
0
@haopeng_nlp
Hao Peng
1 year
Language models excel at undergraduate exams, but how do they fare in research? SciCode challenges models with real research coding problems. Even the best models solve less than 5%. Very proud of @MinyangTian1 and @luyu_gao for leading the charge!.
@MinyangTian1
Minyang Tian
1 year
SciCode is our new benchmark that challenges LMs to code solutions for scientific problems from advanced papers. The challenges were crafted by PhDs;. ~10% of our benchmark is based on Nobel-winning research. GPT-4 and Sonnet 3.5 get <5% ACC. 🧵 1/6
Tweet media one
0
0
11
@haopeng_nlp
Hao Peng
1 year
RT @YueGuo10: I'm joining the UIUC @UofIllinois this fall as an Assistant Professor in the iSchool, with an affiliation in Computer Science….
0
29
0
@haopeng_nlp
Hao Peng
1 year
RT @Francis_YAO_: From Claude100K to Gemini10M, we are in the era of long context language models. Why and how a language model can utilize….
0
172
0
@haopeng_nlp
Hao Peng
1 year
RT @zhaofeng_wu: Want to train an aligned LM in a new language šŸŒ but don’t have preference data for training the reward model (RM)?. šŸ’” Just….
0
37
0
@haopeng_nlp
Hao Peng
1 year
RT @jyangballin: SWE-agent is our new system for autonomously solving issues in GitHub repos. It gets similar accuracy to Devin on SWE-benc….
0
418
0
@haopeng_nlp
Hao Peng
1 year
Very proud of Eurus. A huge shoutout to @lifan__yuan and @charlesfornlp for leading this!.
@lifan__yuan
Lifan Yuan
1 year
Introducing šŸš€Eurus, a suite of state-of-the-art LLM reasoning generalists powered by a new member of Ultra-Series, UltraInteractšŸŽ‰!. Particularly, Eurus-70B beats GPT-3.5 Turbo in reasoning through a comprehensive benchmarking across 12 tests (mostly OOD) covering five tasks!
Tweet media one
0
2
20
@haopeng_nlp
Hao Peng
1 year
Very proud of Eurus. A huge shoutout to @lifan__yuan and @charlesfornlp for leading this!.
@lifan__yuan
Lifan Yuan
1 year
This is a joint work with @charlesfornlp, @wanghanbin95, @stingning, @xingyaow_, Jia Deng, Boji Shan, Huimin Chen, Ruobing Xie, Yankai Lin, Zhenghao Liu, and advisors Bowen Zhou, @haopeng_nlp, @zibuyu9, Maosong Sun. cc @TsinghuaNLP @uiuc_nlp.
0
0
5
@haopeng_nlp
Hao Peng
1 year
RT @Francis_YAO_: Frontier models all have at least 100k context length, Gemini 1.5 has even 1m context. What about research and open sourc….
0
66
0
@haopeng_nlp
Hao Peng
1 year
RT @xingyaow_: Large Language Model (LLM) agents promise to free us from mundane tasks, but how should they best interact with our world? I….
0
94
0