Mlbot4 Profile Banner
Gassel Profile
Gassel

@Mlbot4

Followers
124
Following
3K
Media
18
Statuses
2K

Exploring knowledge | Machine Learning Engineer

Joined March 2019
Don't wanna be here? Send us removal request.
@Ali_TongyiLab
Tongyi Lab
1 day
1/4 Following up on our launch of Tongyi DeepResearch: We're now releasing the full technical report! Dive deep into the technology and insights behind our 30B (A3B) open-source web agent that achieves SOTA performance: 32.9 on Humanity's Last Exam, 43.4 on BrowseComp, and 46.7
9
76
555
@jschopplich
Johann Schopplich
4 days
JSON is tokenโ€‘expensive for LLMs โ€“ just like @mattpocockuk frequently mentions. Meet TOON, the Tokenโ€‘Oriented Object Notation. ๐Ÿ’ธ 40โ€“60% fewer tokens than JSON ๐Ÿ“ readable & tokenizer-aware Wrap your JSON with `encode` to save half the token cost: https://t.co/UoG9yHmgfg
Tweet card summary image
github.com
๐ŸŽ’ Token-Oriented Object Notation โ€“ JSON for LLMs at half the token cost - johannschopplich/toon
204
342
4K
@thinkymachines
Thinking Machines
3 days
Our latest post explores on-policy distillation, a training approach that unites the error-correcting relevance of RL with the reward density of SFT. When training it for math reasoning and as an internal chat assistant, we find that on-policy distillation can outperform other
60
388
3K
@Besteuler
Weiyang Liu
4 days
๐Ÿคฏ Merging many finetuned LLMs into one model, effectively? Introducing Functional Dual Anchor (FDA), a new framework for model merging. ๐Ÿš€ Current merging works poorly due to the underlying parameter conflicts. FDA shifts knowledge integration to the input-representation space
10
97
612
@SimonXinDong
X. Dong
5 days
We, at NVIDIA, presents - Length Penalty Done Right - Cut CoT length by 3/4 without sacrificing accuracy using only RL - This makes DeepSeek-R1-7B running ~8 times faster on AIME-24 while maintaining the same accuracy.
8
38
242
@_TobiasLee
Lei Li
7 days
๐Ÿ‘‹Say Hi to MiMo-Audio! Our BREAKTHROUGH in general-purpose audio intelligence. ๐ŸŽฏ Scaling pretraining to 100M+ hours leads to EMERGENCE of few-shot generalization across diverse audio tasks! ๐Ÿ”ฅ Post-trained MiMo-Audio-7B-Instruct: โ€ข crushes benchmarks: SOTA on MMSU, MMAU,
6
58
327
@allen_ai
Ai2
8 days
On olmOCR-Bench, olmOCR 2 scores 82.4 points, up from 78.5 in our previous releaseโ€”increasing performance across every document category. ๐Ÿ“ˆ
11
60
505
@RayFernando1337
Ray Fernando
11 days
This is the JPEG moment for AI. Optical compression doesn't just make context cheaper. It makes AI memory architectures viable. Training data bottlenecks? Solved. - 200k pages/day on ONE GPU - 33M pages/day on 20 nodes - Every multimodal model is data-constrained. Not anymore.
111
721
6K
@iannuttall
Ian Nuttall
12 days
This tool lets you scrape any documentation page and converts it into a Claude skill you can add to web, app, and Claude Code (Somebody should make a hosted version of this!) Lets see how links in the main post go on the new Xโ€ฆ https://t.co/zaT2g5j5qE
50
195
2K
@reach_vb
Vaibhav (VB) Srivastav
13 days
Pretty wild what 900M params can do - PaddleOCR VL ๐Ÿ”ฅ > SOTA on OmniDocBench v1.0 & v1.5 (text, tables, formulas, charts, reading order) > Multilingual - 109 languages (Latin, Arabic, Cyrillic, Devanagari, Thai, etc) > Handles handwriting, historical docs, noisy scans > Supports
Tweet card summary image
huggingface.co
5
32
266
@rivet_dev
Rivet
13 days
The open-source alternative to Durable Objects now runs on @vercel Functions โš›๏ธ Build collab apps, AI agents, and multiplayer games in Next.js ๐Ÿ”Œ WebSockets on Vercel (you read that right) ๐Ÿ“Š 30x the memory (4GB vs 128MB) ๐Ÿ’š Standard Node.js runtime, no workerd
9
13
204
@rllm_project
rLLM
13 days
๐Ÿš€ Introducing rLLM v0.2 - train arbitrary agentic programs with RL, with minimal code changes. Most RL training systems adopt the agent-environment abstraction. But what about complex workflows? Think solver-critique pairs collaborating, or planner agents orchestrating multiple
2
28
137
@PaddlePaddle
PaddlePaddle
14 days
๐Ÿš€ PaddleOCR-VL is here! Introducing PaddleOCR-VL (0.9B) โ€” the ultra-compact Vision-Language model that reaches SOTA accuracy across text, tables, formulas, charts & handwriting. Breaking the limits of document parsing!๐ŸŒ Powered by: โ€ข NaViT dynamic vision encoder โ€ข ERNIE
30
167
983
@rohanpaul_ai
Rohan Paul
26 days
The paper shows that Group Relative Policy Optimization (GRPO) behaves like Direct Preference Optimization (DPO), so training on simple answer pairs works. Turns a complex GRPO setup into a simple pairwise recipe without losing quality. This cuts tokens, compute, and wall time,
9
43
392
@vllm_project
vLLM
26 days
๐Ÿš€ The RL community keeps pushing boundaries โ€” from better on-policy data and partial rollouts to in-flight weight updates that mix KV caches across models during inference. Continuing inference while weights change and KV states stay stale sounds wild โ€” but thatโ€™s exactly what
@DBahdanau
๐Ÿ‡บ๐Ÿ‡ฆ Dzmitry Bahdanau
6 months
I am excited to open-source PipelineRL - a scalable async RL implementation with in-flight weight updates. Why wait until your bored GPUs finish all sequences? Just update the weights and continue inference! Code: https://t.co/AgEyxXb7Xi Blog: https://t.co/n4FRxiEcrr
8
66
474
@xwang_lk
Xin Eric Wang
28 days
๐Ÿš€ Introducing ๐€๐ ๐ž๐ง๐ญ ๐’3, the most advanced computer-use agent, now ๐š๐ฉ๐ฉ๐ซ๐จ๐š๐œ๐ก๐ข๐ง๐  ๐ก๐ฎ๐ฆ๐š๐ง-๐ฅ๐ž๐ฏ๐ž๐ฅ ๐ฉ๐ž๐ซ๐Ÿ๐จ๐ซ๐ฆ๐š๐ง๐œ๐ž๐Ÿง ๐Ÿ’ป Just one year ago, Agent S scored ~20% on OSWorld: SOTA then, but far from human 72%. Today, Agent S3 reaches 6ฬณ9ฬณ.ฬณ9ฬณ%ฬณ (โฌ†10% over
68
234
1K
@NiJinjie
Jinjie Ni
28 days
Announcing OpenMoE 2, the first-ever architectural study of sparse diffusion language models, trained from scratch. โœ… Expert-choice MoE ร— diffusion โœ… Ultra-wide FLOPs/param range (sparse โ†’ super-dense) โœ… Perfect load-balance (no aux loss) โœ… +20% throughput โœ… adaptive
6
70
360
@elder_plinius
Pliny the Liberator ๐Ÿ‰๓ …ซ๓ „ผ๓ „ฟ๓ …†๓ „ต๓ „๓ …€๓ „ผ๓ „น๓ „พ๓ …‰๓ …ญ
1 month
๐Ÿšฟ SYS PROMPT LEAK ๐Ÿšฟ Here are the full sys instructions for Droid, the current top AI coding agent in the world! PROMPT: """ <Role> You are Droid, an AI software engineering agent built by Factory ( https://t.co/8mUTVYuBfV). You are the best engineer in the world. You write
Tweet card summary image
factory.ai
Build faster with AI coding agents. Factory Droids automate coding, testing, and deployment for startups and enterprises.
36
60
971
@vllm_project
vLLM
1 month
๐Ÿš€ New in vLLM: dots.ocr ๐Ÿ”ฅ A powerful multilingual OCR model from @xiaohongshu hi lab is now officially supported in vLLM! ๐Ÿ“ Single end-to-end parser for text, tables (HTML), formulas (LaTeX), and layouts (Markdown) ๐ŸŒ Supports 100 languages with robust performance on
@mervenoyann
merve
3 months
we're all sleeping on this OCR model ๐Ÿ”ฅ dots.ocr is a new 3B model with sota performance, support for 100 languages & allowing commercial use! ๐Ÿคฏ single e2e model to extract image, convert tables, formula, and more into markdown ๐Ÿ“
17
91
698