Bhavya Agrawalla Profile
Bhavya Agrawalla

@AgrawallaBhavya

Followers
63
Following
1K
Media
0
Statuses
14

Research Interests - Statistics, Deep Reinforcement Learning. PhD student @CMU CS. Prev - Math and CS undergrad at MIT (2021-24) and IISc Bangalore (2020-21).

Cambridge, Massachusetts
Joined November 2022
Don't wanna be here? Send us removal request.
@AgrawallaBhavya
Bhavya Agrawalla
5 months
Suppose we run online SGD for t iterations to estimate linear regression parameter beta^* \in R^d. Consider high-dim regime, so we impose t << d. Can we use SGD output to construct CIs for <v, beta^*> for a new test point v, whose width decays as d^{-alpha} for some alpha > 0?.
1
5
11
@AgrawallaBhavya
Bhavya Agrawalla
22 days
RT @HBCSE_TIFR: Spectacular performance by the Indian team at the International Mathematical Olympiad 2025 held at Sunshine Coast, Australi….
0
56
0
@AgrawallaBhavya
Bhavya Agrawalla
2 months
RT @FahimTajwar10: RL with verifiable reward has shown impressive results in improving LLM reasoning, but what can we do when we do not hav….
0
143
0
@AgrawallaBhavya
Bhavya Agrawalla
4 months
RT @aviral_kumar2: At #ICLR25 workshops, my students+collabs will give many orals talks on newer stuff (don't miss!):. - robot VLA RL fine-….
0
5
0
@AgrawallaBhavya
Bhavya Agrawalla
4 months
RT @QuYuxiao: I am excited to give an oral talk on our work about “Optimizing Test-Time Compute via Meta Reinforcement Fine-Tuning” at #ICL….
0
2
0
@AgrawallaBhavya
Bhavya Agrawalla
4 months
RT @maxsobolmark: I'll be presenting Policy-Agnostic RL: Fine-Tuning of Any Policy Class and Backbone at the Robot Learning (Sunday) and Ge….
0
4
0
@AgrawallaBhavya
Bhavya Agrawalla
4 months
RT @ktiwary2: Yep ! We trained eyeballs from scratch, starting with just light-detecting photoreceptors. 🔬👁️ . Why? To simulate vision evol….
0
3
0
@AgrawallaBhavya
Bhavya Agrawalla
5 months
RT @alirezamh_: With infinite compute, would it make a difference to use Transformers, RNNs, or even vanilla Feedforward nets? They’re all….
0
78
0
@AgrawallaBhavya
Bhavya Agrawalla
5 months
This gives a practically meaningful inference result for SGD in the very high-d regime (samples (n) and dimension (d) -> infty with n/d -> 0) .Arxiv - . Joint work with @krizna_b and Promit Ghosal.
Tweet card summary image
arxiv.org
Stochastic gradient descent (SGD) has emerged as the quintessential method in a data scientist's toolbox. Using SGD for high-stakes applications requires, however, careful quantification of the...
0
1
8
@AgrawallaBhavya
Bhavya Agrawalla
5 months
We construct non-asymptotic, CLT-based CIs for <v, beta^*> (Th 2.3), and show (under mild assumptions on E[XX^T]) that if d^{2/3} << t << d, the lr can be tuned so that both CLT error rate and width of CI decay as d^{-alpha} for some alpha > 0 (Remark 3).
1
1
7
@AgrawallaBhavya
Bhavya Agrawalla
1 year
RT @demishassabis: Advanced mathematical reasoning is a critical capability for modern AI. Today we announce a major milestone in a longsta….
0
581
0