rajan agarwal
@_rajanagarwal
Followers
5K
Following
14K
Media
219
Statuses
2K
RL @amazonscience, se @uwaterloo, scholar @neo, prev @trykino @cohere_labs
sf / waterloo
Joined March 2018
meet Shadow, a powerful open-source background coding agent! fully featured with a remote environment, codebase indexing/wikis and subagents to understand, write and test your code, directly making PRs to github built in a few weeks w/ @ishaandey_, @ElijahKurien
30
23
284
I’ve been using papiers for the past few weeks, it’s super great! im a huge fan of @cognition deepwiki and I hope this becomes the standard for papers
I built https://t.co/VJ7oHdvf7s, a new interface for arXiv As we enter an era of accelerated scientific discovery, we need better tools that augment human cognition to help us keep up. Try it: visit papiers ai or swap arxiv -> papiers on any paper URL
1
9
120
THIS IS THE MOST UNDERVALUED COMPANY ON NASDAQ, according to Warren Buffett’s analysis. Buffett’s rule is simple: Buy real companies that earn real money at irrationally cheap prices. That description fits NextNRG Inc ($NXXT) perfectly. 📊 Facts: * Market Cap ≈ $200–250
21
17
117
Introducing Nested Learning: A new ML paradigm for continual learning that views models as nested optimization problems to enhance long context processing. Our proof-of-concept model, Hope, shows improved performance in language modeling. Learn more: https://t.co/fpdDlYaleL
137
761
5K
the hidden complexity of video search: why multimodal retrieval isn't just "long text" i just watched rajan agarwal from kino ai break down how they're solving one of ai's trickiest problems: searching through video content effectively. most retrieval work focuses on text and
8
7
95
After two years of work, we’ve made an AI Scientist that runs for days and makes genuine discoveries. Working with external collaborators, we report seven externally validated discoveries across multiple fields. It is available right now for anyone to use. 1/5
109
544
4K
two key problems we noticed were tokenizer fragmentation and cross lingual coupling LLINK is an attempt to make a highly compute efficient alignment between a strong multilingual encoder and a LLaMA base! we notice improvement in both content understanding and Q&A
2
0
8
Three months of confusion. One revelation: God was listening the whole time. Get the rest of the story on my page.
0
15
165
recently, i spent some time working on cross-lingual alignment for LLMs via encoder injection! treating languages as modalities is a compute-efficient way to extend understanding of low-resource languages without extending pre-training or the tokenizer
12
11
111
@windsurf @paulg Full breakdown: https://t.co/1InFCutHYO We all need more AI that turns your brain ON, not OFF. Software development only becomes engineering with *understanding*. Your ability to reason through your most challenging coding tasks is constrained by your mental model of how
cognition.ai
Codemaps is meant to offer a shared understanding of a system between humans and AI, enabling your AI to teach you about the code you are looking at quickly and elegantly. A codemap can be generated...
2
12
211
Wow, language models can talk without words. A new framework, Cache-to-Cache (C2C), lets multiple LLMs communicate directly through their KV-caches instead of text, transferring deep semantics without token-by-token generation. It fuses cache representations via a neural
113
297
3K
there is no reason that anything a llm reads or produces that doesn’t get shown on an interface has to be human readable save tokens and let llms learn their own compression
potentially naive question: why dont we train RL on summarized states so we can have significantly longer convos/not violate MDP? also if the summarizer drops info that later matters, we can penalize it if we want llms to "think for 1-2 months" then we have a context problem
3
0
20
Your body speaks truth. Your brain tells stories. TWO SYSTEMS running your life — end the war. Watch. Touch a leaf. Heal.
5
6
24
I built Tacc — a Tokenization-Aware Compression Codec that efficiently sends LLM outputs and tool calls over low-bandwidth networks. It compresses faster and smaller than gzip, making it a better choice for serving LLM responses over HTTP. Here’s how it works 🧵 (1/7)
11
4
65
i strongly believe tab is the ideal interface for coding, writing and even maybe browser use learning and accelerating your patterns is how u become faster. i find myself rewriting so much code with agent tbh
It would be so useful to have Cursor Tab everywhere: - Notes - Terminal - X - Gmail I feel its one of the most underrated features in Ai products. So useful.
4
0
21
potentially naive question: why dont we train RL on summarized states so we can have significantly longer convos/not violate MDP? also if the summarizer drops info that later matters, we can penalize it if we want llms to "think for 1-2 months" then we have a context problem
7
2
48
It was a magical ride with all of you. Our city, our home, our fans showed up on the grand stage...and we couldn't be prouder.
9K
11K
124K
i've only just started going through this but i can already say with confidence that this is an incredible resource for people of all skill levels to learn about LMs
Training LLMs end to end is hard. Very excited to share our new blog (book?) that cover the full pipeline: pre-training, post-training and infra. 200+ pages of what worked, what didn’t, and how to make it run reliably https://t.co/iN2JtWhn23
1
0
25
Kimi Linear Tech Report is dropped! 🚀 https://t.co/LwNB2sQnzM Kimi Linear: A novel architecture that outperforms full attention with faster speeds and better performance—ready to serve as a drop-in replacement for full attention, featuring our open-sourced KDA kernels! Kimi
huggingface.co
27
192
1K
Generalists are useful, but it’s not enough to be smart. Advances come from specialists, whether human or machine. To have an edge, agents need specific expertise, within specific companies, built on models trained on specific data. We call this Specific Intelligence. It's
107
55
589