Explore tweets tagged as #Test_Time_Compute
@TheTuringPost
TuringPost
1 month
6+ concepts you should know to master AI:. - Test-time compute and test-time scaling.- AI inference.- RLHF and its variations: DPO, RRHF, RLAIF.- Meta-learning.- Causal AI.- Defense AI. All about them in these guides ->
Tweet media one
Tweet media two
Tweet media three
7
14
137
@TheTuringPost
TuringPost
1 month
6 AI concepts you should know in 2025. - Test-time compute and how to scale it.- AI inference.- RLHF variations: DPO, RRHF, RLAIF.- Meta-learning.- Causal AI.- Defense AI. Find everything from this list in one place:
Tweet media one
Tweet media two
Tweet media three
8
45
173
@asahi_ictrad
朝日新聞社 メディア研究開発センター
13 days
【✨テックブログ更新✨】 . In-Context LearningはTest-Time Computeの恩恵を受けられるか?. 推論時間を増やしてIn-Context Learningの事例選択を改善することで、LLMの性能向上を実現できるか検証しました!.
Tweet media one
0
0
3
@DimitrisPapail
Dimitris Papailiopoulos
17 days
Thinking Less at test-time requires Sampling More at training-time!. GFPO is a new, cool, and simple Policy Opt algorithm is coming to your RL Gym tonite, led by @VaishShrivas and our MSR group:. Group Filtered PO (GFPO) trades off training-time with test-time compute, in order
Tweet media one
Tweet media two
Tweet media three
Tweet media four
19
41
361
@BasedBeffJezos
Beff – e/acc
19 days
General Relativity took 8 years of Einstein's brain's test time compute. Once AI reaches into the task durations of years to decades it will begin to invent whole new theories about the physical world. This is the new scaling axis.
Tweet media one
158
139
1K
@Jeande_d
Jean de Nyandwi
2 months
Reinforcement Learning of Large Language Models, Spring 2025(UCLA). Great set of new lectures on reinforcement learning of LLMs. Covers a wide range of topics related to RLxLLMs such as basics/foundations, test-time compute, RLHF, and RL with verifiable rewards(RLVR).
Tweet media one
Tweet media two
6
236
1K
@HuggingPapers
DailyPapers
50 minutes
New research tackles a core challenge in LLMs. Go beyond memorization to truly.extend multi-step reasoning depth. Leveraging recurrence, memory, and test-time compute scaling is key.
Tweet media one
1
0
1
@ssydasheng
Shengyang Sun
2 months
We built 200k-GPU clusters; .We scaled up & curated higher-quality data;.We scaled compute by 100x;.We developed training & test-time recipes;.We made everything RL native;.We stabilized infrastructure and speeded up;. That's how you turn RL into the pre-training scale. Yet I am
Tweet media one
53
162
1K
@TheTuringPost
TuringPost
10 hours
Chain-of-Layers (CoLa) is the way to make test-time compute controllable. It treats model layers like building blocks that can be rearranged, so you can build custom versions of the model for each input. CoLa allows to:. - Skip layers for faster, simpler tasks.- Recurrently
Tweet media one
7
16
119
@iScienceLuvr
Tanishq Mathew Abraham, Ph.D.
16 days
Noise Hypernetworks: Amortizing Test-Time Compute in Diffusion Models. "we replace reward guided test-time noise optimization in diffusion models with a Noise Hypernetwork that modulates initial input noise.". "We show that our approach recovers a substantial portion of the
Tweet media one
5
27
161
@drmapavone
Marco Pavone
24 days
Our work on test-time scaling for robotics has been accepted to @corl_conf! We show that scaling test-time compute via a generate-and-verify paradigm offers a practical and effective path toward building general-purpose robotics foundation models.
@jackyk02
Jacky Kwok
2 months
✨ Test-Time Scaling for Robotics ✨. Excited to release 🤖 RoboMonkey, which characterizes test-time scaling laws for Vision-Language-Action (VLA) models and introduces a framework that significantly improves the generalization and robustness of VLAs!. 🧵(1 / N). 🌐 Website:
Tweet media one
2
12
93
@_akhaliq
AK
12 days
Noise Hypernetworks. Amortizing Test-Time Compute in Diffusion Models
Tweet media one
3
18
90
@Zhall333
Zara Hall
19 days
Increasing test-time compute can lead to more accurate LLM decisions, but also more unfair. How can we reap the benefits of modern inference techniques while also ensuring unbiased decision making? We explore this question in our new paper! 🧵
Tweet media one
1
1
4
@Azaliamirh
Azalia Mirhoseini
26 days
Happy to share RoboMonkey, a framework for synthetic data generation + scaling test time compute for VLAs: . Turns out generation (via repeated sampling) and verification (via training a verifier on synthetic data) works well for robotics too!. Training the verifier: we sample N
Tweet media one
Tweet media two
5
29
163
@Marktechpost
Marktechpost AI Dev News ⚡
1 month
Too Much Thinking Can Break LLMs: Inverse Scaling in Test-Time Compute. Recent advances in large language models (LLMs) have encouraged the idea that letting models “think longer” during inference usually improves their accuracy and robustness. Practices like chain-of-thought
0
6
14
@fly51fly
fly51fly
1 month
[LG] Inverse Scaling in Test-Time Compute.A P Gema, A Hägele, R Chen, A Arditi. [University of Edinburgh & EPFL & University of Texas at Austin] (2025).
Tweet media one
Tweet media two
Tweet media three
Tweet media four
0
2
8
@theomitsa
Dr. Theophano Mitsa ☦️🇬🇷🇺🇸
17 days
What is test-time compute and how to scale it?.
0
0
2
@chijinML
Chi Jin
24 days
The technical report for Goedel-Prover-V2 is out!. 📌 SOTA among all open-source theorem provers.⚡ Among the best overall—including closed-source—under small test-time compute. Read it here:
6
36
173
@AndrewYNg
Andrew Ng
2 days
Parallel agents are emerging as an important new direction for scaling up AI. AI capabilities have scaled with more training data, training-time compute, and test-time compute. Having multiple agents run in parallel is growing as a technique to further scale and improve.
102
364
2K