adityaarun1 Profile Banner
Aditya Arun Profile
Aditya Arun

@adityaarun1

Followers
343
Following
21K
Media
117
Statuses
8K

Minimalist, Research Scientist @ Adobe, PhD

Noida, India
Joined March 2010
Don't wanna be here? Send us removal request.
@adityaarun1
Aditya Arun
7 years
At times you come across an essay or a discussion which articulates succinctly the fleeting thoughts that have been troubling you for years.
2
0
5
@grantdraws
Grant Snider
3 days
Join me this Tuesday for a virtual workshop on Making Poetry Comics from Nature!
Tweet media one
1
35
123
@grok
Grok
30 days
Generate videos in just a few seconds. Try Grok Imagine, free for a limited time.
535
839
4K
@cosmo_shirley
Shirley Ho
6 days
Extremely proud of our recent "Lost in Latent Space" paper by our amazing @PolymathicAI intern @FrancoisRozet ! How far do you think we can/should compress ? What is needed in the latent space to keep rolling out the physics right across so many scales?
@FrancoisRozet
Franรงois Rozet
6 days
Does a smaller latent space lead to worse generation in latent diffusion models? Not necessarily! We show that LDMs are extremely robust to a wide range of compression rates (10-1000x) in the context of physics emulation. We got lost in latent space. Join us ๐Ÿ‘‡
Tweet media one
7
61
502
@tli104
Tianjian Li
6 days
Language models often produce repetitive responses, and this issue is further amplified by post-training. In this work, we introduce DARLING, a method that explicitly optimizes for both response diversity and quality within online reinforcement learning!
@jaseweston
Jason Weston
6 days
๐ŸŒ€Diversity Aware RL (DARLING)๐ŸŒ€ ๐Ÿ“: https://t.co/MH0tui34Cb - Jointly optimizes for quality & diversity using a learned partition function - Outperforms standard RL in quality AND diversity metrics, e.g. higher pass@1/p@k - Works for both non-verifiable & verifiable tasks ๐Ÿงต1/5
Tweet media one
2
23
90
@GoogleAIStudio
Google AI Studio
14 days
๐ŸŒ nano banana is here โ†’ gemini-2.5-flash-image-preview - SOTA image generation and editing - incredible character consistency - lightning fast available in preview in AI Studio and the Gemini API
Tweet media one
Tweet media two
295
1K
9K
@SebastienBubeck
Sebastien Bubeck
20 days
Claim: gpt-5-pro can prove new interesting mathematics. Proof: I took a convex optimization paper with a clean open problem in it and asked gpt-5-pro to work on it. It proved a better bound than what is in the paper, and I checked the proof it's correct. Details below.
Tweet media one
291
1K
8K
@adityaarun1
Aditya Arun
20 days
Something that I can totally attest to.
@justinskycak
Justin Skycak
21 days
Today is the 1-year anniversary of the best blog post of 2024, "You Are NOT Dumb, You Just Lack the Prerequisites" by @lelouchdaily:
Tweet media one
0
0
2
@omarsar0
elvis
20 days
Chain-of-Agents Interesting idea to train a single model with the capabilities of a multi-agent system. 84.6% reduction in inference cost! Distillation and Agentic RL are no joke! Here are my notes:
Tweet media one
19
163
866
@heinemandavidj
David Heineman
21 days
Evaluating language models is tricky, how do we know if our results are real, or due to random chance? We find an answer with two simple metrics: signal, a benchmarkโ€™s ability to separate models, and noise, a benchmarkโ€™s random variability between training steps ๐Ÿงต
Tweet media one
@allen_ai
Ai2
21 days
๐Ÿ“ข New paper from Ai2: Signal & Noise asks a simple questionโ€”can language model benchmarks detect a true difference in model performance? ๐Ÿงต
Tweet media one
3
44
234
@sedielem
Sander Dieleman
21 days
New survey on diffusion language models: https://t.co/SHicf69gxV (via @NicolasPerezNi1). Covers pre/post-training, inference and multimodality, with very nice illustrations. I can't help but feel a bit wistful about the apparent extinction of the continuous approach after 2023๐Ÿฅฒ
Tweet media one
7
85
547
@jacobaustin132
Jacob Austin
22 days
Today we're putting out an update to the JAX TPU book, this time on GPUs. How do GPUs work, especially compared to TPUs? How are they networked? And how does this affect LLM training? 1/n
Tweet media one
38
528
3K
@fengyao1909
Feng Yao
1 month
Failing on ๐ฅ๐š๐ซ๐ ๐ž-๐ฌ๐œ๐š๐ฅ๐ž ๐‘๐‹ with VeRL? โš ๏ธ Mixing inference backend (๐ฏ๐‹๐‹๐Œ/๐’๐†๐‹๐š๐ง๐ ) with training backends (๐…๐’๐ƒ๐/๐Œ๐ž๐ ๐š๐ญ๐ซ๐จ๐ง) ๐ฌ๐ž๐œ๐ซ๐ž๐ญ๐ฅ๐ฒ ๐ญ๐ฎ๐ซ๐ง๐ฌ ๐ฒ๐จ๐ฎ๐ซ ๐‘๐‹ ๐ข๐ง๐ญ๐จ ๐จ๐Ÿ๐Ÿ-๐ฉ๐จ๐ฅ๐ข๐œ๐ฒ โ€” even if they share the same weights! ๐Ÿ“‰ย Blog:
Tweet media one
13
117
719
@coolfunnytshirt
Keh Ke Peheno
25 days
Brilliant satire by Shraddha Jain on the political language war and language chauvinism! Clean, healthy, bait-free, trigger-free, abuse-free standup comedy! A rarity these days! ๐Ÿ‘Œ๐Ÿฝ https://t.co/7pA0BK1NPl
61
2K
9K
@RoshanKrRaii
Roshan Rai
26 days
This even after 79 Years, remains the best speech by a leader in the world ๐Ÿ‡ฎ๐Ÿ‡ณ Share it with every young Indian who has never listened to Pandit Nehru speaking. Goosebumps. #IndependenceDay2025
94
1K
5K
@p_bojanowski
Piotr Bojanowski
26 days
I am happy to share the work of our team. The outcome of a collaborative effort, by a joyful group of skilled and determined scientists and engineers! Congrats to the team on this amazing milestone!
@AIatMeta
AI at Meta
26 days
Introducing DINOv3: a state-of-the-art computer vision model trained with self-supervised learning (SSL) that produces powerful, high-resolution image features. For the first time, a single frozen vision backbone outperforms specialized solutions on multiple long-standing dense
5
21
220
@arxiv
arXiv.org
26 days
#HBD to arXiv!๐ŸŽˆ On August 14, 1991, the very first paper was submitted to arXiv. That's 34 years of sharing research quickly, freely & openly! Some baby pictures to show how far we've come . . . when we were just a computer under desk . . . & in our 1994 punk phase . . . ๐Ÿ‘ถ๐Ÿ’พ
Tweet media one
15
175
763
@maxseitzer
Max Seitzer
26 days
Introducing DINOv3 ๐Ÿฆ•๐Ÿฆ•๐Ÿฆ• A SotA-enabling vision foundation model, trained with pure self-supervised learning (SSL) at scale. High quality dense features, combining unprecedented semantic and geometric scene understanding. Three reasons why this mattersโ€ฆ
Tweet media one
12
140
1K
@osanseviero
Omar Sanseviero
26 days
Introducing Gemma 3 270M ๐Ÿ”ฅ ๐ŸคA tiny model! Just 270 million parameters ๐Ÿง  Very strong instruction following ๐Ÿค– Fine-tune in just a few minutes, with a large vocabulary to serve as a high-quality foundation https://t.co/E0BB5nlI1k
Tweet media one
124
334
3K
@TanentzapfLab
Tanentzapf Lab
27 days
The goal of a PhD is not to learn some facts or read a few papers or learn a bunch of techniques. The goal of a PhD is to learn independence, problem solving, how to finish things you start, resilience, & gain the ability to adapt & think creatively. Learning these things is hard
25
513
3K
@AnimaAnandkumar
Prof. Anima Anandkumar
28 days
How do we build AI for science? Augment with AI or replace with AI? https://t.co/adSSmJXcFx Popular prescription is to augment AI into existing workflows rather than replace them, e.g., keep the approximate numerical solver for simulations, and use AI only to correct its errors
Tweet media one
2
35
194
@FrnkNlsn
Frank Nielsen
27 days
Two types of universal density approximators: โ‘  Gaussian Mixtures Models โ‘ก Polynomial Exponential Families with sufficient statistics = monomials (Single distrib.) Convert GMMs โ†” PEFs either using moment MLE parameter to natural parameter or score matching (Ref.: slide)
Tweet media one
2
22
202