
Cheng Luo
@ChengLuo_lc
Followers
18
Following
57
Media
5
Statuses
20
🌟 Announcing the 1st Workshop on Efficient Reasoning (ER) at @NeurIPSConf 2025 — Dec 6 or 7, San Diego !. 📣 We welcome submissions! Submit your work here: .🗓️ Deadline: September 1, 2025 (AoE) .🔗 Website: 💬 Topics
0
1
1
RT @InfiniAILab: 🔥 We introduce Multiverse, a new generative modeling framework for adaptive and lossless parallel generation. 🚀 Multivers….
0
82
0
RT @InfiniAILab: 🥳 Happy to share our new work – Kinetics: Rethinking Test-Time Scaling Laws. 🤔How to effectively build a powerful reasoni….
0
69
0
RT @AnimaAnandkumar: HeadInfer: Unlocking Long-Context LLM Inference on Consumer GPUs (Million-level Tokens).*long-context inputs require l….
0
14
0
RT @ChengLuo_lc: 🤩🤩 we introduce MST, a memory-efficient transformer, reducing intermediate memory usage and enabling longer sequence train….
0
1
0
RT @rohanpaul_ai: Really 👀 new Paper, MINI-SEQUENCE TRANSFORMER claims to extend the maximum context length of Qwen, Mistral, and Gemma-2 b….
0
4
0
RT @papers_anon: Training Ultra Long Context Language Model with Fully Pipelined Distributed Transformer. Saw a 16x increase in sequence le….
0
7
0
Curious about boosting context length in Llama 3.1 by 16x? 🦙 Our Mini-sequence Transformer (MST) offers insights! 🚀 MST extends context length with no performance drop. 📈 Our paper: and GitHub: #Llama #NLP #AI #llama31.
github.com
Contribute to wdlctc/mini-s development by creating an account on GitHub.
0
1
1
RT @AnimaAnandkumar: Introducing long-context transformer using mini sequences. It is a simple and effective method for highly efficient an….
0
8
0