Dongwei__Jiang Profile Banner
Dongwei Jiang Profile
Dongwei Jiang

@Dongwei__Jiang

Followers
510
Following
2K
Media
26
Statuses
142

Working on LLMs, focusing specifically on reasoning and self-improvement. Spent six years in my past life doing research in industry on speech processing

Baltimore, MD
Joined June 2022
Don't wanna be here? Send us removal request.
@Dongwei__Jiang
Dongwei Jiang
1 month
🧵 Recent studies show LLMs can self-improve their responses when given external feedback. But how effectively can they incorporate it?. We tested this systematically—and found they can't fully integrate feedback, even when the feedback is high-quality and backed by ground-truth.
Tweet media one
3
26
155
@Dongwei__Jiang
Dongwei Jiang
1 month
RT @leanprover: Incredibly grateful to @TheOfficialACM SIGPLAN for awarding #LeanLang the Programming Languages Software Award 2025 at #PLD….
0
38
0
@Dongwei__Jiang
Dongwei Jiang
1 month
RT @DanielKhashabi: 🚨🚨 New paper out with @Dongwei__Jiang and team:. Even with near-perfect, ground-truth feedback, LLMs often fail to full….
0
7
0
@Dongwei__Jiang
Dongwei Jiang
1 month
So what's next? A better understanding of this probably involves interactions between how models understand feedback, follow instructions, and update beliefs. We'll continue to investigate this and hope our findings help future research in building truly self-improving AI systems.
1
0
3
@Dongwei__Jiang
Dongwei Jiang
1 month
But what's causing Feedback Friction? Previous work suggested overconfidence, data familiarity, problem complexity might be causing this, but we found no evidence supporting it. We also checked if different models fail on the same problems—and they don't.
1
0
3
@Dongwei__Jiang
Dongwei Jiang
1 month
Training these models is very expensive, so we tried to solve this problem by (1) increasing temperature to add diversity and (2) rejecting previously generated wrong answers to force exploration of new solutions. This helped marginally, but the fundamental problem persists.
Tweet media one
1
0
2
@Dongwei__Jiang
Dongwei Jiang
1 month
We tested three feedback types to see how feedback quality affects performance: simple binary feedback ("wrong"), self-generated reflection, and reflection generated by stronger models. Better feedback definitely helps, but models still plateau below their theoretical ceiling.
Tweet media one
1
0
2
@Dongwei__Jiang
Dongwei Jiang
1 month
We categorize persistent errors after multiple rounds and found that majority are from feedback resistance, where models don't incorporate valid corrections. Only a few come from feedback quality. The main issue isn't bad feedback—solver models just don't integrate good feedback!
Tweet media one
1
0
4
@Dongwei__Jiang
Dongwei Jiang
1 month
We tested across many models: Claude 3.7 (±thinking), Llama-4-Maverick, Llama-4-Scout, Llama-3.3 on 9 diverse tasks. Even the strongest models (Claude 3.7 w/ thinking) plateau well below their target accuracy—the theoretical ceiling if they fully incorporated all feedback.
Tweet media one
1
0
3
@Dongwei__Jiang
Dongwei Jiang
1 month
The setup: A solver model attempts a problem, and when it's wrong, gets targeted feedback generated by a strong feedback generator model with access to the ground-truth answer. The solver model can retry up to 10 times with specific feedback on their mistakes at every iteration.
Tweet media one
1
0
4
@Dongwei__Jiang
Dongwei Jiang
2 months
RT @linxins2: 🚨 We discovered a surprising side effect of Reinforcement Finetuning (RFT): it makes LLMs more confidently wrong on unanswera….
0
38
0
@Dongwei__Jiang
Dongwei Jiang
2 months
Now accepted by #ACL2025!. Thrilled to see our paper also referenced in @lilianweng's latest blog post on reasoning in LLMs! Check it out:
lilianweng.github.io
Special thanks to John Schulman for a lot of super valuable feedback and direct edits on this post. Test time compute (Graves et al. 2016, Ling, et al. 2017, Cobbe et al. 2021) and Chain-of-thought...
@Dongwei__Jiang
Dongwei Jiang
10 months
Process supervision for reasoning is 🔥! While previous approaches often relied on human annotation and struggled to generalize across different reasoning tasks, we're now asking: Can we improve this?. Introducing 𝐑𝐀𝐓𝐈𝐎𝐍𝐀𝐋𝐘𝐒𝐓: a new model pre-trained on implicit
Tweet media one
0
11
58
@Dongwei__Jiang
Dongwei Jiang
3 months
RT @jackjingyuzhang: Current copyright mitigation methods for LLMs typically focus on average-case risks, but overlook worst-case scenarios….
0
9
0
@Dongwei__Jiang
Dongwei Jiang
4 months
RT @iScienceLuvr: Reasoning to Learn from Latent Thoughts. "Motivated by how humans apply deliberate thinking to learn from limited data, w….
0
114
0
@Dongwei__Jiang
Dongwei Jiang
4 months
RT @natolambert: Verification, The Key to AI .Read the archives of Rich Sutton, Turing Award winner :D, has all the major ideas https://t.c….
0
45
0
@Dongwei__Jiang
Dongwei Jiang
5 months
I'll be at #AAAI25 presenting my poster on Self-[In]Correct ( during Session 3 on March 1st at 12:30. Would love to connect if you're attending!.
Tweet card summary image
arxiv.org
Can LLMs consistently improve their previous outputs for better results? For this to be true, LLMs would need to be better at discriminating among previously-generated alternatives, than...
0
2
25