peixuanhakhan Profile Banner
Peixuan Han (韩沛煊) Profile
Peixuan Han (韩沛煊)

@peixuanhakhan

Followers
80
Following
17
Media
11
Statuses
39

1st year Ph.D. student at UIUC @IllinoisCS Amazon 25Summer Intern LLM researcher

Urbana
Joined September 2024
Don't wanna be here? Send us removal request.
@peixuanhakhan
Peixuan Han (韩沛煊)
1 month
RT @AlexiGlad: How can we unlock generalized reasoning?. ⚡️Introducing Energy-Based Transformers (EBTs), an approach that out-scales (feed-….
0
255
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
Super excited to begin my Applied Scientist Internship at @amazon, which is my first internship in the industry. I'm looking forward to conducting interesting and insightful research on the efficient reasoning of LLMs!
Tweet media one
Tweet media two
0
0
1
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
RT @xiusi_chen: Can LLMs make rational decisions like human experts?. 📖Introducing DecisionFlow: Advancing Large Language Model as Principl….
0
16
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
RT @JiaxunZhang6: ⚠️ Rogue AI scientists? 🛡️ SafeScientist rejects unsafe prompts for ethical discoveries. Check out paper ➡️ ( https://t….
0
7
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
(5/5) Other interesting findings:. ⭐️Without RL, the base model struggles to use ToM info—RL is essential. ⭐️RL Process enables ToMAP to think longer and deeper. ⭐️ToMAP generalizes well to multi-turn long conversations. ⭐️ToMAP frequently adopts logical, opponent-aware tactics
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
0
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
(4/5) After ToM-enhanced RL training, our model—ToMAP (built on Qwen-2.5-3B)—shows strong persuasion performance, surpassing much larger LLMs. The ToM modules and RL training enable ToMAP to generate more diverse arguments and address the opponent’s concerns more effectively!
Tweet media one
1
0
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
(3/5) To address this, we propose two ToM modules that help LLMs infer their opponent’s mindset:. 💡Counterclaim predictor: What reasons might the opponent have to disagree?. 💡Attitude predictor: How confident is the opponent in those reasons?
Tweet media one
1
0
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
(2/5) In real conversations, considering the other side’s thoughts is crucial. This ability—known as "theory of mind" (ToM)—is a core component of human cognition. However, most LLMs lack this capacity. They often argue rigidly, failing to adapt to their opponent's perspective.
Tweet media one
1
0
0
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
(1/5) Want to make your LLM a skilled persuader?. Check out our latest paper: "ToMAP: Training Opponent-Aware LLM Persuaders with Theory of Mind"!. For details:.📄Arxiv: 🛠️GitHub:
Tweet media one
2
6
23
@peixuanhakhan
Peixuan Han (韩沛煊)
2 months
RT @qiancheng1231: 📢 New Paper Drop: From Solving to Modeling!.LLMs can solve math problems — but can they model the real world? 🌍. 📄 arXiv….
0
30
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @xwzliuzijia: 💥Time-R1 is here! Can a 3B LLM truly grasp time? 🤔 YES! . Excited to share our new work, Time-R1: Towards Comprehensive Te….
0
3
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
It seems I underestimated the exploding speed of submissions lol.
0
0
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
NeurIPS will easily reach 25000 submissions this year😯.
2
0
3
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @cursor_ai: Cursor is now free for students. Enjoy!.
0
4K
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @ExplainMiracles: We introduce Gradient Variance Minimization (GVM)-RAFT, a principled dynamic sampling strategy that minimizes gradient….
0
27
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @xiusi_chen: 🚀 Can we cast reward modeling as a reasoning task?. 📖 Introducing our new paper: .RM-R1: Reward Modeling as Reasoning. 📑 Pa….
0
47
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @haofeiyu44: 🧪 Want an AI-generated paper draft in just 1 minute? Or dreaming of building auto-research apps but frustrated with setups?….
Tweet card summary image
github.com
A lightweight framework for building research agents designed for developers - ulab-uiuc/tiny-scientist
0
10
0
@peixuanhakhan
Peixuan Han (韩沛煊)
3 months
RT @Alibaba_Qwen: Introducing Qwen3! . We release and open-weight Qwen3, our latest large language models, including 2 MoE models and 6 den….
0
2K
0
@peixuanhakhan
Peixuan Han (韩沛煊)
5 months
RT @BowenJin13: 🚀 Excited to announce that our paper 𝐒𝐞𝐚𝐫𝐜𝐡-𝐑𝟏. is now live! 📄. We introduce an RL framework (an extension of 𝐃𝐞𝐞𝐩𝐬𝐞𝐞𝐤-𝐑𝟏)….
0
111
0
@peixuanhakhan
Peixuan Han (韩沛煊)
5 months
RT @sundarpichai: Gemma 3 is here! Our new open models are incredibly efficient - the largest 27B model runs on just one H100 GPU. You'd ne….
0
884
0