lightetal Profile Banner
Jonathan @SF Profile
Jonathan @SF

@lightetal

Followers
161
Following
296
Media
20
Statuses
80

I’m a PhD researcher at @RPI @Caltech @NEC working in LLM-agents, reasoning, reinforcement learning, and decision making.

Pasadena, CA
Joined June 2023
Don't wanna be here? Send us removal request.
@lightetal
Jonathan @SF
23 days
🧵10/n Huge thanks to my amazing collaborators — Wei Cheng, Benjamin Riviere, @FrankYueWu1 , @stillpedant, @MengdiWang10, @yisongyue , Santiago Paternain, and Haifeng Chen Truly a pleasure collaborating across @NECLabsAmerica , @Caltech , @Princeton, and @rpi . 🙏
1
0
2
@lightetal
Jonathan @SF
23 days
🧵9/n Inference scaling doesn’t need to be brute force. It can be intelligent. By letting models adapt their reasoning granularity, DISC opens a new path toward efficient, self-reflective inference. Paper: https://t.co/rR8NhdRgJl Website:
Tweet card summary image
arxiv.org
Inference scaling methods for LLMs often rely on decomposing problems into steps (or groups of tokens), followed by sampling and selecting the best next steps. However, these steps and their sizes...
1
0
1
@lightetal
Jonathan @SF
23 days
🧵8/n DISC is plug-and-play: ✅ Works with greedy, beam, or MCTS search ✅ Boosts open-source models like Llama, Mistral, and Qwen ✅ Keeps runtime overhead negligible One algorithm, many models — no retraining needed. 🧠
1
0
0
@lightetal
Jonathan @SF
23 days
🧵7/n Reasoning models like DeepSeek-R1 love DISC. ❤️‍🔥 With only 10 samples, it boosts performance by 85% — and even when capped to the same token budget as one vanilla generation, DISC still wins by +33%. Smarter reasoning, not longer decoding. ⚡️
1
0
0
@lightetal
Jonathan @SF
23 days
🧵6/n Up to 4× higher accuracy over base model with open source models, using just 10 samples. 🔥
1
0
0
@lightetal
Jonathan @SF
23 days
🧵5/n So, how well does it work? 📈 Across benchmarks: - APPS: -5.0% error - MATH: -6.7% error - LiveCodeBench: -10.5% error Better scaling with compute budget across the board Beats baselines at any budget
1
0
0
@lightetal
Jonathan @SF
23 days
🧵4/n At its core, DISC teaches models where to think harder. It dynamically allocates compute — 🧩 subdividing challenging steps ⚙️ focusing search on high-reward prefixes 🚀 avoiding wasted effort on trivial tokens All with no extra training, heuristics, or supervision.
1
0
0
@lightetal
Jonathan @SF
23 days
🧵3/n Instead of predefining step sizes, DISC learns to adaptively break down reasoning in real time — taking bigger leaps on easy parts and finer steps where the model struggles. It’s like giving your LLM a dynamic zoom lens for reasoning. 🔍
1
0
0
@lightetal
Jonathan @SF
23 days
🧵2/n Static inference methods split reasoning steps into fixed steps (token, line, or sentence)… but that’s like using the same stride length for every terrain. 🥾 How can we make inference both scalable and adaptive? That’s where DISC comes in. 💡
1
1
0
@lightetal
Jonathan @SF
23 days
🎉 Thrilled to share that our paper “DISC: Dynamic Decomposition Improves LLM Inference Scaling” has been accepted to #neurips2025 ! 🚀 Here’s how we push reasoning and inference scaling to new heights 🧵 1/n
4
6
18
@yisongyue
Yisong Yue
24 days
DISC: Dynamic Decomposition Improves LLM Inference Scaling DISC adaptively partitions reasoning traces during inference so that language models devote more compute to the hardest reasoning steps — leading to faster, more accurate inference across coding and math benchmarks.
2
5
17
@lightetal
Jonathan @SF
27 days
And thanks for the great talk on hallucination and HLE from @ofirnachum, Edwin Zhang, and Long Pham!
0
0
0
@lightetal
Jonathan @SF
27 days
Love seeing research and industry collide in real time at @agihouse_org paper reading roundtables
1
3
10
@lm4sci
LM4SCI @ COLM2025
1 month
📅 TOMORROW is LM4Sci #COLM2025! ⚡🔬 Are you excited? Today's spotlight: Jennifer Sun (Cornell, Google DeepMind) @JenJSun on Accelerating Knowledge & Discovery in Scientific Workflows 🧵
1
3
7
@lightetal
Jonathan @SF
5 months
The Interlink at YC AI SUS was super fun — it matched you with others who shared your interests automatically. Would love to see academic AI/ML conferences like #NeurIPS, #ICLR, and #ICML adopt something similar so that it is easier to find people now that conferences have
0
0
4
@lightetal
Jonathan @SF
5 months
Andrej Karpathy also showed how easy it’s becoming to “vibe code” complex apps — like this AI-powered menu generator he built in a night: https://t.co/DpFMnQkp4K Excited for a future where building powerful tools is more accessible than ever for everyone!
0
0
0
@lightetal
Jonathan @SF
5 months
Y Combinator AI SUS last week was a blast. Got to hear Andrej Karpathy talk about the future of AI-driven software engineering — he's phenomenal at making complex ideas feel intuitive.
0
0
2
@lightetal
Jonathan @SF
5 months
We don’t hardcode pain into games—so why does losing feel so real? And what does that teach us about AI and reward modeling? https://t.co/0NXzcpFJ8B #AI #ReinforcementLearning #MachineLearning
Tweet card summary image
substacktools.com
Why a Game Loss Can Sting More Than a Stubbed Toe
0
0
1