Onedroid
@ye_ack
Followers
1
Following
247
Media
1
Statuses
173
Joined June 2017
Introducing ๐ฅEGGROLL ๐ฅ(Evolution Guided General Optimization via Low-rank Learning)! ๐ Scaling backprop-free Evolution Strategies (ES) for billion-parameter models at large population sizes โก100x Training Throughput ๐ฏFast Convergence ๐ขPure Int8 Pretraining of RNN LLMs
21
147
941
Language Models are Provably Injective and Invertible! A groundbreaking paper challenges the long-held belief that LLMs lose information. They prove mathematically and show empirically across billions of tests that inputs map uniquely to representations, making them lossless.
10
25
191
๐ณ๏ธ๐Into the Rabbit Hull โ Part II Continuing our interpretation of DINOv2, the second part of our study concerns the geometry of concepts and the synthesis of our findings toward a new representational phenomenology: the Minkowski Representation Hypothesis
5
67
380
A graph-powered all-in-one RAG system! RAG-Anything is a graph-driven, all-in-one multimodal document processing RAG system built on LightRAG. It supports all content modalities within a single integrated framework. 100% open-source.
30
406
3K
"Aggressive Filtering aint good for larger training" Similar find also at
Second insight: Optimal filtering changes predictably with scale. Smaller models benefit from aggressive filtering (e.g., top 3% at 10ยฒโฐ FLOPs), while larger models prefer larger, more diverse datasets (e.g., top 30% at 10ยฒยณ FLOPs). Specific rates vary by data pool, but the
4
46
353
๐ง How can AI evolve from statically ๐ต๐ฉ๐ช๐ฏ๐ฌ๐ช๐ฏ๐จ ๐ข๐ฃ๐ฐ๐ถ๐ต ๐ช๐ฎ๐ข๐จ๐ฆ๐ด โ dynamically ๐ต๐ฉ๐ช๐ฏ๐ฌ๐ช๐ฏ๐จ ๐ธ๐ช๐ต๐ฉ ๐ช๐ฎ๐ข๐จ๐ฆ๐ด as cognitive workspaces, similar to the human mental sketchpad? ๐ Whatโs the ๐ฟ๐ฒ๐๐ฒ๐ฎ๐ฟ๐ฐ๐ต ๐ฟ๐ผ๐ฎ๐ฑ๐บ๐ฎ๐ฝ from tool-use โ programmatic
0
73
199
UC Berkley has two free courses on LLM Agents for foundational and advanced levels. it also has some of the best lecturers from DeepMind, Meta, and top universities. basically covers all you need to know about agents from the best resources out there.
10
327
2K
๐คCan diffusion models write code competitively? Excited to share our latest 7B coding diffusion LLM!!๐ป With DiffuCoder, we explore how they decode, why temperature๐ฅ matters, and how to improve them via coupled-GRPO that speaks diffusion!!๐ Code: https://t.co/sWsb8a49HL ๐งต
5
112
583
The Ultimate Toolkit for Working with LLMs! Transformer Lab lets you train, fine-tune, and chat with any LLMโ100% locally. Enjoy 1-click LLM downloads and a drag-and-drop UI for RAG. 100% open-source.
7
94
439
LLMs running reinforcement learning shed their randomness almost at once, then their scores stall. This paper shows that randomness drop is predictable and fixable, so bigger gains are still on the table. The authors fit an exponential link between entropy and reward. Two
10
87
510
CVPR 2025 papers pt. 2 - SAMWISE SAMWISE adds language understanding and temporal reasoning to SAM2; you can segment and track objects in videos just by describing them more papers: https://t.co/1VlLn2BWxl โ more
7
56
396
Vision transformers have high-norm outliers that hurt performance and distort attention. While prior work removed them by retraining with โregisterโ tokens, we find the mechanism behind outliers and make registers at โจtest-timeโจโgiving clean features and better performance! ๐งต
16
135
1K
I learnt a lot from O'Reilly books, this is surreal I'm writing a book with amazing people @micuelll @andimarafioti @orr_zohar about VLMs with @huggingface ๐๐ Early Access (first two chapters in raw) are available to everyone, we'd love to have your feedback!
95
236
2K
๐ Excited to share the most inspiring work Iโve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
84
512
3K
Here's Let's Build a Simple Database! It's a bit outdated, but it still perfectly covers the basics on how to get a sqlite clone going, in C. You'll learn a lot about databases and C, enjoy!
4
64
500
OpenMemory MCP provides a persistent memory layer for AI tools like Claude, Cursor and Windsurf. It enables AI Agents to securely read and write to a shared memory. Runs 100% locally on your computer.
12
63
491
A collection of awesome MCP servers for AI Agents:
8
93
728
๐จThis week's top AI/ML research papers: - Absolute Zero - RM-R1 - Seed-Coder - Flow-GRPO - ZeroSearch - Ming-Lite-Uni - A Survey on Large Multimodal Reasoning Models - On Path to Multimodal Generalist - ZeroSearch - HunyuanCustom - Unified Multimodal CoT Reward Model through
8
118
997
The Ultimate Toolkit for Working with LLMs! Transformer Lab lets you train, fine-tune, and chat with any LLMโ100% locally. Enjoy 1-click LLM downloads and a drag-and-drop UI for RAG. 100% open-source.
14
150
702