Junlin Wang Profile
Junlin Wang

@JunlinWang3

Followers
229
Following
304
Media
13
Statuses
41

PhD Student @duke_nlp. Interning at @togethercompute. Inference-time scaling, multi-agent systems

Joined December 2013
Don't wanna be here? Send us removal request.
@JunlinWang3
Junlin Wang
26 days
RT @togethercompute: Most AI benchmarks test the past. But real intelligence is about predicting the future. Introducing FutureBench — a….
0
19
0
@JunlinWang3
Junlin Wang
29 days
Work done during my internship at Together AI is being presented at #icml25. Come and check it out! . We propose a new model alignment pipeline that harness collective intelligence from open-source llms!
Tweet media one
0
3
18
@grok
Grok
3 days
Generate videos in just a few seconds. Try Grok Imagine, free for a limited time.
868
3K
10K
@JunlinWang3
Junlin Wang
2 months
RT @NehzUx: LLMs are getting more powerful, but they still struggle with super long documents. A common trick is "Divide and Conquer" - cho….
0
6
0
@JunlinWang3
Junlin Wang
2 months
Curious why backtracking works so well? Dive into our latest study that unpacks the mechanisms behind its success.
@Hongyicai2002
Hongyi James Cai
2 months
🚀Excited to share our new paper: How Much Backtracking is Enough?. 🤔 How many backtracks should your LLM learn to reason better?.Turns out: the harder the task, the more backtracking you need!
Tweet media one
0
0
2
@JunlinWang3
Junlin Wang
3 months
RT @james_y_zou: Our new #icml2025 paper w/@togethercompute shows how to use synthetic data from Mixture-of-Agents to boost LM fine-tuning….
0
20
0
@JunlinWang3
Junlin Wang
3 months
RT @togethercompute: 🚀 Introducing Mixture-of-Agents Alignment (MoAA), a new method to "distill" the collective intelligence of open-source….
0
11
0
@JunlinWang3
Junlin Wang
4 months
Linguistic markers like “hmm” and “wait are a good predictor of response correctness.
Tweet media one
0
0
2
@JunlinWang3
Junlin Wang
4 months
Correctness gap widens with increasing problem difficulty.
Tweet media one
1
0
3
@JunlinWang3
Junlin Wang
4 months
For reasoning models, shorter responses are more likely to be correct. Non-reasoning models show no trend.
Tweet media one
1
0
2
@JunlinWang3
Junlin Wang
4 months
Non-reasoning models augmented by inference-time methods cannot match reasoning models. Find out more in our preprint 👇.📄 Awesome collaboration with @ShangZhu18 @JonSaadFalcon @ben_athi @QingyangWu1 @JueWANG26088228 @Leon75421958 @ce_zhang @bhuwandhingra
Tweet media one
1
1
6
@JunlinWang3
Junlin Wang
4 months
Excited to share work from my @togethercompute internship—a deep dive into inference‑time scaling methods 🧠. We rigorously evaluated verifier‑free inference-time scaling methods across both reasoning and non‑reasoning LLMs. Some key findings:. 🔑 Even with huge rollout budgets,
Tweet media one
1
54
175
@JunlinWang3
Junlin Wang
4 months
RT @iScienceLuvr: Think Deep, Think Fast: Investigating Efficiency of Verifier-free Inference-time-scaling Methods. "This work conducts a c….
0
37
0
@JunlinWang3
Junlin Wang
4 months
RT @LindaHe49140661: Excited to share our work on scaling LLMs to handle million-token contexts! Training models for ultra-long sequences i….
0
46
0
@JunlinWang3
Junlin Wang
4 months
RT @AnnieFeng6: #ICLR2025 Oral. LLMs often struggle with reliable and consistent decisions under uncertainty 😵‍💫 — largely because they can….
0
39
0
@JunlinWang3
Junlin Wang
4 months
RT @togethercompute: Introducing Open Deep Research!. A fully open-source Deep Research tool that:.• writes comprehensive reports.• does mu….
0
73
0
@JunlinWang3
Junlin Wang
5 months
RT @james_y_zou: Happy to share that Mixture of Agents (MoA) is selected as a #ICLR2025 spotlight!⚡️ @togethercompute. Our updated paper ha….
Tweet card summary image
github.com
Together Mixture-Of-Agents (MoA) – 65.1% on AlpacaEval with OSS models - togethercomputer/MoA
0
3
0
@JunlinWang3
Junlin Wang
6 months
RT @bhuwandhingra: **New paper from @RoyXie_ **. Do LLMs know when they have read enough to answer a question?. We show how language models….
0
4
0
@JunlinWang3
Junlin Wang
7 months
RT @ClementDelangue: Our science team has started working on fully reproducing and open-sourcing R1 including training data, training scrip….
0
532
0
@JunlinWang3
Junlin Wang
7 months
To be presented @ICLR2025 !!! Super excited for my first iclr conference😆.
@arankomatsuzaki
Aran Komatsuzaki
1 year
TogetherAI presents Mixture-of-Agents Enhances Large Language Model Capabilities. Achieves SotA performance on AlpacaEval 2.0, MT-Bench and FLASK, surpassing GPT4o.
Tweet media one
0
0
13
@JunlinWang3
Junlin Wang
9 months
f1 is very different from above in that it may scale inference-time compute not through search or self-correct, but through compound AI techniques like Mixture of Agents (. I view methods like this as another pathway of inference-time scaling and can be
Tweet media one
0
0
1