Onedroid Profile
Onedroid

@ye_ack

Followers
1
Following
247
Media
1
Statuses
173

Joined June 2017
Don't wanna be here? Send us removal request.
@_akhaliq
AK
1 month
discuss:
Tweet card summary image
huggingface.co
0
1
11
@bidiptas13
Bidipta Sarkar
2 months
Introducing ๐ŸฅšEGGROLL ๐Ÿฅš(Evolution Guided General Optimization via Low-rank Learning)! ๐Ÿš€ Scaling backprop-free Evolution Strategies (ES) for billion-parameter models at large population sizes โšก100x Training Throughput ๐ŸŽฏFast Convergence ๐Ÿ”ขPure Int8 Pretraining of RNN LLMs
21
147
941
@HuggingPapers
DailyPapers
3 months
Language Models are Provably Injective and Invertible! A groundbreaking paper challenges the long-held belief that LLMs lose information. They prove mathematically and show empirically across billions of tests that inputs map uniquely to representations, making them lossless.
10
25
191
@thomas_fel_
Thomas Fel
3 months
๐Ÿ•ณ๏ธ๐Ÿ‡Into the Rabbit Hull โ€“ Part II Continuing our interpretation of DINOv2, the second part of our study concerns the geometry of concepts and the synthesis of our findings toward a new representational phenomenology: the Minkowski Representation Hypothesis
5
67
380
@_avichawla
Avi Chawla
5 months
A graph-powered all-in-one RAG system! RAG-Anything is a graph-driven, all-in-one multimodal document processing RAG system built on LightRAG. It supports all content modalities within a single integrated framework. 100% open-source.
30
406
3K
@cloneofsimo
Simo Ryu
6 months
"Aggressive Filtering aint good for larger training" Similar find also at
@dmizrahi_
David Mizrahi
6 months
Second insight: Optimal filtering changes predictably with scale. Smaller models benefit from aggressive filtering (e.g., top 3% at 10ยฒโฐ FLOPs), while larger models prefer larger, more diverse datasets (e.g., top 30% at 10ยฒยณ FLOPs). Specific rates vary by data pool, but the
4
46
353
@May_F1_
Yi R. (May) Fung
6 months
๐Ÿง  How can AI evolve from statically ๐˜ต๐˜ฉ๐˜ช๐˜ฏ๐˜ฌ๐˜ช๐˜ฏ๐˜จ ๐˜ข๐˜ฃ๐˜ฐ๐˜ถ๐˜ต ๐˜ช๐˜ฎ๐˜ข๐˜จ๐˜ฆ๐˜ด โ†’ dynamically ๐˜ต๐˜ฉ๐˜ช๐˜ฏ๐˜ฌ๐˜ช๐˜ฏ๐˜จ ๐˜ธ๐˜ช๐˜ต๐˜ฉ ๐˜ช๐˜ฎ๐˜ข๐˜จ๐˜ฆ๐˜ด as cognitive workspaces, similar to the human mental sketchpad? ๐Ÿ” Whatโ€™s the ๐—ฟ๐—ฒ๐˜€๐—ฒ๐—ฎ๐—ฟ๐—ฐ๐—ต ๐—ฟ๐—ผ๐—ฎ๐—ฑ๐—บ๐—ฎ๐—ฝ from tool-use โ†’ programmatic
0
73
199
@Hesamation
โ„ฮตsam
7 months
UC Berkley has two free courses on LLM Agents for foundational and advanced levels. it also has some of the best lecturers from DeepMind, Meta, and top universities. basically covers all you need to know about agents from the best resources out there.
10
327
2K
@sansa19739319
Sansa Gong
7 months
๐Ÿค–Can diffusion models write code competitively? Excited to share our latest 7B coding diffusion LLM!!๐Ÿ’ป With DiffuCoder, we explore how they decode, why temperature๐Ÿ”ฅ matters, and how to improve them via coupled-GRPO that speaks diffusion!!๐Ÿ“ˆ Code: https://t.co/sWsb8a49HL ๐Ÿงต
5
112
583
@DailyDoseOfDS_
Daily Dose of Data Science
7 months
The Ultimate Toolkit for Working with LLMs! Transformer Lab lets you train, fine-tune, and chat with any LLMโ€”100% locally. Enjoy 1-click LLM downloads and a drag-and-drop UI for RAG. 100% open-source.
7
94
439
@rohanpaul_ai
Rohan Paul
7 months
LLMs running reinforcement learning shed their randomness almost at once, then their scores stall. This paper shows that randomness drop is predictable and fixable, so bigger gains are still on the table. The authors fit an exponential link between entropy and reward. Two
10
87
510
@skalskip92
SkalskiP
7 months
CVPR 2025 papers pt. 2 - SAMWISE SAMWISE adds language understanding and temporal reasoning to SAM2; you can segment and track objects in videos just by describing them more papers: https://t.co/1VlLn2BWxl โ†“ more
7
56
396
@nickhjiang
Nick Jiang
7 months
Vision transformers have high-norm outliers that hurt performance and distort attention. While prior work removed them by retraining with โ€œregisterโ€ tokens, we find the mechanism behind outliers and make registers at โœจtest-timeโœจโ€”giving clean features and better performance! ๐Ÿงต
16
135
1K
@mervenoyann
merve
7 months
I learnt a lot from O'Reilly books, this is surreal I'm writing a book with amazing people @micuelll @andimarafioti @orr_zohar about VLMs with @huggingface ๐Ÿ“–๐Ÿ’— Early Access (first two chapters in raw) are available to everyone, we'd love to have your feedback!
95
236
2K
@xuandongzhao
Xuandong Zhao
8 months
๐Ÿš€ Excited to share the most inspiring work Iโ€™ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
84
512
3K
@RuiCarrilho5
Roy Carrilho
8 months
Here's Let's Build a Simple Database! It's a bit outdated, but it still perfectly covers the basics on how to get a sqlite clone going, in C. You'll learn a lot about databases and C, enjoy!
4
64
500
@Saboo_Shubham_
Shubham Saboo
8 months
OpenMemory MCP provides a persistent memory layer for AI tools like Claude, Cursor and Windsurf. It enables AI Agents to securely read and write to a shared memory. Runs 100% locally on your computer.
12
63
491
@DailyDoseOfDS_
Daily Dose of Data Science
8 months
A collection of awesome MCP servers for AI Agents:
8
93
728
@TheAITimeline
The AI Timeline
8 months
๐ŸšจThis week's top AI/ML research papers: - Absolute Zero - RM-R1 - Seed-Coder - Flow-GRPO - ZeroSearch - Ming-Lite-Uni - A Survey on Large Multimodal Reasoning Models - On Path to Multimodal Generalist - ZeroSearch - HunyuanCustom - Unified Multimodal CoT Reward Model through
8
118
997
@akshay_pachaar
Akshay ๐Ÿš€
8 months
The Ultimate Toolkit for Working with LLMs! Transformer Lab lets you train, fine-tune, and chat with any LLMโ€”100% locally. Enjoy 1-click LLM downloads and a drag-and-drop UI for RAG. 100% open-source.
14
150
702