Shubhendu Trivedi
@_onionesque
Followers
9K
Following
27K
Media
2K
Statuses
23K
Cultivated Abandon. Twitter interests: Machine learning research, applied mathematics, mathematical miscellany, ML for physics/chemistry, books.
New York, NY / Cambridge, MA
Joined October 2008
I wrote a review paper about statistical methods in generative AI; specifically, about using statistical tools along with genAI models for making AI more reliable, for evaluation, etc. See here: https://t.co/oNrb4dYe9i! I have identified four main areas where statistical
11
96
467
E. O. Thorp also wrote a paper with an identical title https://t.co/h2nQVxlXpL One difference is that Breiman is writing more for a probability audience, while Thorp is writing as an applied mathematician concerned with constructive betting strategies for real world play.
0
0
3
expected log wealth after n rounds), and shows that the log-utility formulation characterizes the only strategy that both avoids eventual ruin (with prob 1) and achieves maximal asymptotic growth. Also see this survey re: testing, where this is relevant
arxiv.org
Safe anytime-valid inference (SAVI) provides measures of statistical evidence and certainty -- e-processes for testing and confidence sequences for estimation -- that remain valid at all stopping...
1
0
7
Check this ! Sunta Clause Run about to Happen ! $TSLA $NVDA $BTC $GME
1
0
16
Kelly assumed a fixed game with constant odds. This generalizes the optimization of long-run capital growth to a) sequences of different gambles, b) arbitrary distributions, c) stochastic payoffs that may change over time. It frames it in stochastic optimization language (max [+]
1
0
5
Finally got around to reading this paper by Breiman that generalizes Kelly betting to other settings. It gives that seemed like a heuristic, a decision-theoretic foundation. Now very relevant to testing. https://t.co/683BPDFFYX
6
17
138
Basically unusable. Made me a bigger fan of brutalist architecture for similar settings, ironically.
1
1
5
People at MIT (and the leaks, and the mold) could attest.
1
1
21
Sleigh the season with the most personal gift around. Get them a Cameo video!
0
37
604
Behold, the bible of code foundation models. A mixed team from over a dozen institutions put together almost 200 pages of survey on different paradigms for program synthesis. 1000+ references and some nice figures. Seems like a solid resource.
8
62
630
The last bit (in the OP) is also not entirely true. Productivity gains are becoming bottlenecked by deployment, and poor AI strategy in cos. Open research is great, it accelerates scientific/engineering spillover, but does not meaningfully influence value capture at this stage.
0
0
0
The last might seem like a joke, but it's literally true. Everyone needs exposure to "AI," this everyone knows, but returns are not being sought from productivity gains but from exit momentum (betting on secondary markets, roll-ups, or acquisition by incumbents).
1
0
0
This is the legitimizing narrative around valuation. But valuations are less about productivity gains and more about risk capital concentration, crowding (due to the trend-chasing nature of the VC enterprise), and basically momentum-trading in venture markets.
AI valuations are anchored in a promise of productivity gains based on compounding AI improvements. We're now building the products on a bedrock of research from 5 years ago, when everybody was publishing. Open research is the best way to keep compounding the AI improvements.
1
0
9
I don't foresee ever working on the mean field type of stuff, but I like the papers. I like the conceptual neatness of the model moving in Wasserstein space of probability measures and inductive bias becoming the geometry of the flow + activation + architecture.
0
0
9
Looks like they have a bunch of works on the theme now. I was only familiar with
arxiv.org
Viewing Transformers as interacting particle systems, we describe the geometry of learned representations when the weights are not time dependent. We show that particles, representing tokens, tend...
1
1
9
🧵[1/8] Excited to share our NeurIPS 2025 Spotlight paper “Does Object Binding Naturally Emerge in Large Pretrained Vision Transformers?” ✨ To add to the broader discussion of binding in neural networks, we ask whether and how Vision Transformers perform object binding (the
13
100
717
Review article: "The Physics of Soft Adhesion" (by Katharine E. Jensen, Chelsea S. Davis):
arxiv.org
This review provides an introduction to the essential physics of soft adhesion, including the thermodynamics of adhesion and wetting, the mechanics of contact with deformable materials, and the...
1
5
22
Spiking neural networks are surprisingly good for neural decoding. Much better than I thought they could be: https://t.co/fP6nro77au.
arxiv.org
Brain-computer interfaces (BCIs) promise to enable vital functions, such as speech and prosthetic control, for individuals with neuromotor impairments. Central to their success are neural...
1
26
153