Junyu Zhang Profile
Junyu Zhang

@jyzhang1208

Followers
70
Following
72
Media
3
Statuses
19

MSCS @UofIllinois | Prev: Undergrad @HuazhongUST | Reinforcement Learning & Foundation Models | Looking for PhD opportunities (Fall 2026)

Joined July 2023
Don't wanna be here? Send us removal request.
@jyzhang1208
Junyu Zhang
5 days
💥Excited to share our paper “AlphaOne: Reasoning Models Thinking Slow and Fast at Test Time” at #EMNLP2025 🚀 this Friday, Nov. 7, during Gather Session 4. Come say hi virtually!👋 📄Paper: https://t.co/CksN8hEuoF 🪩Website & Code: https://t.co/AwMLAQFvtz #AI #LLMs #Reasoning
2
4
14
@chongyiz1
Chongyi Zheng
23 days
1/ How can we model the future rewards (returns) for RL agents? While prior methods round the returns into discrete bins or predict a finite number of quantiles, we use flexible models to predict the fine-grained structure of the full return distribution: https://t.co/YyGEEQnnEH.
2
7
41
@jyzhang1208
Junyu Zhang
5 days
Also huge thanks to all my incredible collaborators and friends @RunpeiDong @HanWang178023 @krystal_ning @HaoranGeng2 @peihao_goblue @Xialin_He @YutongBAI1002 @JitendraMalikCV @_saurabhg @huan_zhang12 @drdhxi. Without them, nothing could be done.
0
0
0
@jyzhang1208
Junyu Zhang
5 days
🎯Key Takeaways 1⃣Slow thinking first, then fast thinking, leads to better LRM reasoning. 2⃣Slow thinking can bring efficient test-time scaling. 3⃣Dense reasoning modulation brings improved reasoning.
1
0
0
@jyzhang1208
Junyu Zhang
5 days
This led us to systematically study test-time reasoning modulation and develop AlphaOne, a universal framework that scale LRM reasoning at test time. 🔬
1
0
0
@jyzhang1208
Junyu Zhang
5 days
Through our research, we found that LRMs struggle to find the optimal schedule between these two thinking modes the way we do effortlessly every day!!!
1
0
0
@jyzhang1208
Junyu Zhang
5 days
Back in April, Runpei and I were fascinated by how AI reasoning differs from human thinking. Humans naturally switch between fast intuition⚡(System 1) and slow deliberation🐢(System 2). But how about large reasoning models?
1
0
2
@delafuentelab
César de la Fuente
4 months
For years I have dreamt of a tool that could neutralize pathogens the moment they emerge. Today we unveil ApexOracle—an AI that, from a pathogen’s genome and phenotypic knowledge alone, predicts which antibiotics will work and invents new molecules for threats it has never seen.
6
38
222
@chongyiz1
Chongyi Zheng
5 months
1/ How should RL agents prepare to solve new tasks? While prior methods often learn a model that predicts the immediate next observation, we build a model that predicts many steps into the future, conditioning on different user intentions: https://t.co/QYPy1Tlv4x.
5
33
189
@jyzhang1208
Junyu Zhang
5 months
EmbodiedBench got an ICML 2025 Oral! Time to challenge your MLLMs on embodied tasks!🤖
@RuiYang70669025
Rui Yang
5 months
Excited to share that EmbodiedBench was selected for an Oral at ICML 2025! We recently added results for new models (InternVL3, Gemma3, Ovis2) and released a large agent trajectory dataset on 🤗: https://t.co/91OLQaaHbT Try training and evaluating your MLLM for embodied agents!
0
0
6
@jyzhang1208
Junyu Zhang
5 months
Huge thanks for sharing our work @_akhaliq! AlphaOne deep dive & code release coming soon 🚀
@_akhaliq
AK
5 months
AlphaOne Reasoning Models Thinking Slow and Fast at Test Time
2
6
33
@omarsar0
elvis
5 months
Reasoning Models Thinking Slow and Fast at Test Time Another super cool work on improving reasoning efficiency in LLMs. They show that slow-then-fast reasoning outperforms other strategies. Here are my notes:
11
60
283
@RuiYang70669025
Rui Yang
8 months
🚀 New model results on EmbodiedBench! 🚀 🔹 Qwen2.5 VL surpasses Qwen2 VL as embodied agents! 🔹 InternVL2_5 MPO leads as the best-performing open-source model! Check out the latest results: https://t.co/91OLQaa9ml Explore the evaluation code: https://t.co/Q0FP6RI2dC
1
4
15
@jyzhang1208
Junyu Zhang
9 months
When we successfully built a framework that enables MLLM-based agents to plan for low-level manipulation tasks (a key component of EmbodiedBench), I was super excited! Could this be a step toward MLLM-based agents becoming so versatile that we no longer need dedicated VLA models?
@RuiYang70669025
Rui Yang
9 months
🤖Can MLLM agents reason about spatial relationships and plan atomic actions for navigation & manipulation? 🔥 Meet EmbodiedBench 🏆—the first fine-grained benchmark for MLLM-based embodied agents! 📄 Paper: https://t.co/zGP6SmBUPk 🌐 Website & code: https://t.co/91OLQaaHbT
1
0
14
@Xingang20
Xingang Guo
1 year
🔍 Can Vision-Language Models (VLMs) truly reason about math, or are they just reflecting patterns they’ve seen before? 🤔 We tested GPT-4o and Claude-3.5 with our new benchmark, DynaMATH, and the results were eye-opening! 🧩 GPT-4o struggles to recognize when a shifted
3
9
22
@jyzhang1208
Junyu Zhang
1 year
Here are the links to our paper and project website. Paper: https://t.co/wTno35WCow Website:
0
0
3
@jyzhang1208
Junyu Zhang
1 year
Thrilled to share our work HERD in #ICLR2024. Thanks much to my co-author Heng @drdhxi for a year of collaboration in exploring Robot Design. Hope to see further research and attention on algorithmic aspects in this field!
@drdhxi
Heng Dong
1 year
Present our novel work “Leveraging Hyperbolic Embeddings for Coarse-Fine Robot Design” in #ICLR2024 in #Vienna. Welcome to visit our website https://t.co/d5nzA2U8mz if you are interested. Also many thanks to my excellent co-author @jyzhang1208. Without her, nothing could be done.
2
0
5
@drdhxi
Heng Dong
2 years
Our work in #ICML2023: Symmetry-Aware Robot Design with Structured Subgroups. https://t.co/JtUsDvALP3
0
1
6