
Apache TVM
@ApacheTVM
Followers
4K
Following
537
Media
53
Statuses
656
Open deep learning compiler stack for CPUs, GPUs and specialized accelerators. Join us for the TVM and Deep Learning Compilation Conference https://t.co/i6MTbWYt87
Joined January 2018
ICYMI, all of the sessions from #tvmcon are available for streaming! Catch up on the latest advances, case studies, and tutorials in #ML acceleration from the @ApacheTVM community.
1
1
11
RT @tqchenml: #MLSys2026 will be led by the general chair @luisceze and PC chairs @JiaZhihao and @achowdhery. The conference will be held i….
0
13
0
RT @JiaZhihao: 📢Exciting updates from #MLSys2025! All session recordings are now available and free to watch at We….
0
31
0
RT @Xinyu2ML: 🚀 Super excited to share Multiverse!. 🏃 It’s been a long journey exploring the space between model design and hardware effici….
0
20
0
RT @BeidiChen: Say hello to Multiverse — the Everything Everywhere All At Once of generative modeling. 💥 Lossless, adaptive, and gloriousl….
0
22
0
RT @InfiniAILab: 🔥 We introduce Multiverse, a new generative modeling framework for adaptive and lossless parallel generation. 🚀 Multivers….
0
82
0
RT @tqchenml: Check out our work on parallel reasoning 🧠; We bring an AI-assisted curator that identifies parallel paths in sequential trac….
0
15
0
RT @yi_xin_dong: @databricks 's Agent Bricks is powered by XGrammar for structured generation, and achieves high quality and efficiency. It….
0
4
0
RT @GT_HaoKang: 🚀📉 A new kind of efficiency challenge: "Win Fast or Lose Slow: Balancing Speed and Accuracy in Latency-Sensitive Decisions….
0
12
0
RT @tqchenml: #MLSys2025 @AnimaAnandkumar will give a keynote Hardware-aware training and inference for large-scale AI at 10:30 today; make….
0
4
0
RT @ying11231: Congrats to @ye_combinator @tqchenml @luisceze! Flashinfer has been the real power behind various inference frameworks! Hope….
0
4
0
RT @ye_combinator: We’re thrilled that FlashInfer won a Best Paper Award at MLSys 2025! 🎉.This wouldn’t have been possible without the comm….
0
37
0
RT @NVIDIAAIDev: 🎉 Congratulations to the FlashInfer team – their technical paper, "FlashInfer: Efficient and Customizable Attention Engine….
0
47
0
RT @ruihanglai: Many congrats to @ye_combinator and the entire team for the best paper award 🏆! Excited to see that FlashInfer is getting w….
0
3
0
RT @tqchenml: FlashInfer won #MLSys2025 best paper🏆, with backing from @NVIDIAAIDev to bring top LLM inference kernels to the community.
0
26
0
RT @tqchenml: #MLSys2025 make sure to attend 10:30am keynote @istoica05 An AI stack: from scaling AI workloads to evaluating LLMs. Checkou….
0
15
0
RT @PyTorch: The PyTorch Foundation is a Gold Sponsor of #MLSys2025 this week in Santa Clara. Visit the booth and explore talks from Soumi….
0
8
0
RT @JoyChew_d: Super excited to release FlexAttention for Inference with a decoding backend, GQA, PagedAttention, trainable bias and more!….
0
7
0