iarunmah Profile Banner
Arun M Profile
Arun M

@iarunmah

Followers
59
Following
2K
Media
10
Statuses
250

Converses with chatbots more than humans.

Planet Earth
Joined May 2023
Don't wanna be here? Send us removal request.
@iarunmah
Arun M
2 months
This Asia Cup format makes no sense. Why are India & Pak in the same group again? It's a setup to guarantee multiple INDvPAK matches,potentially 3! Same thing happens in World Cups too. Can’t @BCCI request @ICC & ACC to put them in separate groups to minimize the no. of matches?
0
1
2
@keenanisalive
Keenan Crane
2 months
“Everyone knows” what an autoencoder is… but there's an important complementary picture missing from most introductory material. In short: we emphasize how autoencoders are implemented—but not always what they represent (and some of the implications of that representation).🧵
43
410
3K
@rohanpaul_ai
Rohan Paul
2 months
This is probably one of THE most important paper of the last few months. Small language models are sufficiently powerful, operationally suitable, and economical Agentic tasks. - Phi-2 matches 30 billion models running 15x faster. - Serving a 7 billion parameter small language
26
144
888
@VikParuchuri
Vik Paruchuri
2 months
High quality math is the secret sauce for reasoning models. The best math data is in old papers. But OCRing that math is full of insane edge cases. Let's talk about how to solve this, and how you can get better math data than many frontier labs 🧵
19
66
740
@emollick
Ethan Mollick
2 months
Paper from OpenAI says hallucinations are less a problem with LLMs themselves & more an issue with training on tests that only reward right answers. That encourages guessing rather than saying “I don’t know” If this is true, there is a straightforward path for more reliable AI.
140
253
2K
@rohanpaul_ai
Rohan Paul
2 months
Massive proposal in this paper. Argues chips that compute with real physics, not strict digital rules, can relieve the AI compute bottleneck. Early prototypes hit 1000x speed and show steep energy cuts. AI demand keeps rising while data center power, training bills, and
34
78
456
@rohanpaul_ai
Rohan Paul
2 months
A classic paper, collab between @AIatMeta , @GoogleDeepMind , and @NVIDIAAIDev Language models keep personal facts in a measurable amount of “storage”. This study shows how to count that storage—and when models swap memorization for real learning. 📡 The Question Can we
16
170
929
@JeanRemiKing
Jean-Rémi King
2 months
Can AI help understand how the brain learns to see the world? Our latest study, led by @JRaugel from FAIR at @AIatMeta and @ENS_ULM, is now out! 📄 https://t.co/y2Y3GP3bI5 🧵 A thread:
31
303
2K
@ZhiSu22
Zhi Su
2 months
🏓🤖 Our humanoid robot can now rally over 100 consecutive shots against a human in real table tennis — fully autonomous, sub-second reaction, human-like strikes.
118
564
3K
@GoogleAIStudio
Google AI Studio
2 months
🍌 nano banana is here → gemini-2.5-flash-image-preview - SOTA image generation and editing - incredible character consistency - lightning fast available in preview in AI Studio and the Gemini API
294
1K
9K
@DeepLearningAI
DeepLearning.AI
3 months
India launched a program to build native large language models for its many languages, funding startups and pooling compute under the $1.2 billion IndiaAI Mission. The Ministry of Electronics and Information Technology reserved 19,000 GPUs (including 13,000 Nvidia H100s). So
Tweet card summary image
deeplearning.ai
India, which has limited funding and large numbers of languages and dialects, is redoubling its efforts to build native large language models.
33
264
2K
@techwith_ram
𝗿𝗮𝗺𝗮𝗸𝗿𝘂𝘀𝗵𝗻𝗮— 𝗲/𝗮𝗰𝗰
3 months
PyTorch vs TensorFlow: New survey settles the debate. - PyTorch → ~25% faster in training, ~78% faster in eval. Simpler for research. TensorFlow → Stronger in deployment (mobile, browser, server). Read the whole survey here: https://t.co/BHZDfLXdah
6
68
348
@erisaonX
erisa
3 months
so many papers so little time
12
31
217
@AIatMeta
AI at Meta
3 months
Introducing DINOv3: a state-of-the-art computer vision model trained with self-supervised learning (SSL) that produces powerful, high-resolution image features. For the first time, a single frozen vision backbone outperforms specialized solutions on multiple long-standing dense
342
784
5K
@cwolferesearch
Cameron R. Wolfe, Ph.D.
3 months
The gpt-oss models from OpenAI are a synthesis of ideas from prior research. Here are 10 interesting papers that were directly used in gpt-oss… (1) Longformer: Introduces sliding window attention, a form of sparse attention that is utilized in alternating layers of both gpt-oss
7
94
419
@AIatMeta
AI at Meta
3 months
🏆 We're thrilled to announce that Meta FAIR’s Brain & AI team won 1st place at the prestigious Algonauts 2025 brain modeling competition. Their 1B parameter model, TRIBE (Trimodal Brain Encoder), is the first deep neural network trained to predict brain responses to stimuli
148
426
3K
@deedydas
Deedy
3 months
Huge computer science result: A Tsinghua professor JUST discovered the fastest shortest path algorithm for graphs in 40yrs. This improves on Turing award winner Tarjan’s O(m + nlogn) with Dijkstra’s, something every Computer Science student learns in college.
245
2K
22K
@HardCricketpix
Cricket Picture that Goes Hard
3 months
🇮🇳
2
154
3K
@NASAJPL
NASA JPL
3 months
How will NISAR support agriculture and farming? Hear from the scientists and end users how its data will be used to map crop growth, track plant health, and monitor soil moisture.
50
183
911
@deedydas
Deedy
3 months
🚨 AI models just invented better, novel AI models. Chinese researchers fed all LLM research into a model and it discovered 106 novel AI model architectures that converge to lower loss with better benchmarks. ASI-Arch is one of the coolest AI papers this year. En route AGI.
100
463
2K