qx_dong Profile Banner
Qingxiu Dong Profile
Qingxiu Dong

@qx_dong

Followers
2K
Following
470
Media
17
Statuses
128

PhD student @PKU1898. Research Intern @MSFTResearch Asia.

Joined August 2019
Don't wanna be here? Send us removal request.
@qx_dong
Qingxiu Dong
3 months
Thanks to @omarsar0 for sharing our work!
@omarsar0
elvis
3 months
Reinforcement Pre-Training New pre-training paradigm for LLMs just landed on arXiv! It incentivises effective next-token reasoning with RL. This unlocks richer reasoning capabilities using only raw text and intrinsic RL signals. A must-read! Bookmark it! Here are my notes:
Tweet media one
0
1
16
@qx_dong
Qingxiu Dong
3 months
Paper:
Tweet card summary image
huggingface.co
0
3
39
@qx_dong
Qingxiu Dong
3 months
โฐ We introduce Reinforcement Pre-Training (RPT๐Ÿ’) โ€” reframing next-token prediction as a reasoning task using RLVR โœ… General-purpose reasoning ๐Ÿ“‘ Scalable RL on web corpus ๐Ÿ“ˆ Stronger pre-training + RLVR results ๐Ÿš€ Allow allocate more compute on specific tokens
Tweet media one
31
149
957
@qx_dong
Qingxiu Dong
3 months
Paper:
0
0
3
@qx_dong
Qingxiu Dong
5 months
So happy to reunite with old and new friends at ICLR! Had an amazing time exploring Singapore too! ๐ŸŒŸ๐Ÿ‡ธ๐Ÿ‡ฌ #ICLR2025
Tweet media one
Tweet media two
Tweet media three
Tweet media four
2
1
78
@realHongyu_Wang
Hongyu Wang๐Ÿฅ•
5 months
Excited to introduce BitNet b1.58 2B4T โ€” the first large-scale, native 1-bit LLM๐Ÿš€๐Ÿš€ BitNet achieves performance on par with leading full-precision LLMs โ€” and itโ€™s blazingly fastโšก๏ธโšก๏ธuses much lower memory๐ŸŽ‰ Everything is open-sourced โ€” run it on GPU or your Macbook ๐Ÿ–ฅ๏ธโš™๏ธ
Tweet media one
23
98
533
@liangchen5518
Liang Chen
9 months
Proud to introduce our latest work โ€œNext Token Prediction Towards Multimodal Intelligence: A Comprehensive Surveyโ€ as our new year gift for the multimodal learning community! Paper: https://t.co/fpMfOYexbr Github: https://t.co/9g9hq6yYrT
Tweet media one
Tweet media two
Tweet media three
Tweet media four
1
55
256
@ZenMoore1
Zekun Wang (ZenMoore) ๐Ÿ”ฅ
9 months
๐ŸŽ†Survey of the Year: ๐๐ž๐ฑ๐ญ ๐“๐จ๐ค๐ž๐ง ๐๐ซ๐ž๐๐ข๐œ๐ญ๐ข๐จ๐ง ๐“๐จ๐ฐ๐š๐ซ๐๐ฌ ๐Œ๐ฎ๐ฅ๐ญ๐ข๐ฆ๐จ๐๐š๐ฅ ๐ˆ๐ง๐ญ๐ž๐ฅ๐ฅ๐ข๐ ๐ž๐ง๐œ๐ž: ๐€ ๐‚๐จ๐ฆ๐ฉ๐ซ๐ž๐ก๐ž๐ง๐ฌ๐ข๐ฏ๐ž ๐’๐ฎ๐ซ๐ฏ๐ž๐ฒ arXiv: https://t.co/M5H0VoVpSX HugFace: https://t.co/9QqIuoA1AB Github: https://t.co/T4Kac2aDaw
Tweet media one
Tweet media two
Tweet media three
2
64
273
@qx_dong
Qingxiu Dong
10 months
About to arrive in #Miami ๐ŸŒด after a 30-hour flight for #EMNLP2024! Excited to see new and old friends :) Iโ€™d love to chat about data synthesis and deep reasoning for LLMs (or anything else) โ€”feel free to reach out!
0
2
24
@99Solaris
Yifan Song
11 months
๐Ÿš€Introducing MixEval-X, the first real-world, any-to-any benchmark. https://t.co/XJeUAYMDhQ It extends all benefits of MixEval to multi-modal evaluations, including real-world query distribution, fast yet accurate model ranking, high standards evaluation across modalities!
@NiJinjie
Jinjie Ni
11 months
๐Ÿ‡ Frontier players are racing to solve modality puzzles in the quest for AGI. But to get there, we need consistent, high-standard evaluations across all modalities! ๐Ÿš€ Introducing MixEval-X, the first real-world, any-to-any benchmark. Inheriting the philosophy from MixEval,
Tweet media one
1
4
8
@realHongyu_Wang
Hongyu Wang๐Ÿฅ•
11 months
How to deploy a 100B model on your CPU devices? ๐Ÿ”ฅ Excited to introduce bitnet.cpp, our inference framework for BitNet b1.58 ๐Ÿš€๐Ÿš€ https://t.co/nVAGhE8n5A
Tweet media one
12
70
363
@qx_dong
Qingxiu Dong
11 months
Project Lead: Bofei Gao @TpKY09wMKKWdeT3 !
0
0
2
@qx_dong
Qingxiu Dong
11 months
Tweet card summary image
huggingface.co
0
1
4
@qx_dong
Qingxiu Dong
11 months
OpenAI o1 scores 94.8% on MATH dataset๐Ÿ˜ฒ Then...how should we proceed to track and evaluate the next-gen LLMs' math skills? ๐Ÿ‘‰Omni-Math: a new, challenging benchmark with 4k competition-level problems, where OpenAI o1-mini only achieves 60.54 acc Paper: https://t.co/Qggc7paGwe
Tweet media one
10
24
134
@liangchen5518
Liang Chen
11 months
โœจA Spark of Vision-Language Intelligence! We introduce DnD-Transformer, a new auto-regressive image gen model beats GPT/Llama w/o extra cost. AR gen beats diffusion in joint VL modeling in a self-supervised way! Github: https://t.co/NZNA0Xt2mR Paper:
Tweet card summary image
huggingface.co
2
17
77
@qx_dong
Qingxiu Dong
11 months
(Perhaps a bit late) Excited to announce our survey on ICL has been accepted to #EMNLP2024 main conf and been cited 1,000+ times! Thanks to all collaborators and contributors to this field! We've updated the survey https://t.co/Ev7iMn6IHR. Excited to keep pushing boundaries!
Tweet card summary image
arxiv.org
With the increasing capabilities of large language models (LLMs), in-context learning (ICL) has emerged as a new paradigm for natural language processing (NLP), where LLMs make predictions based...
2
43
213
@yeeelow233
Yiming Huang
11 months
๐Ÿค” Are LLMs Ready for Real-World Data Science Challenges? ๐Ÿš€ Weโ€™ve just open-sourced our #EMNLP2024 work DA-Code, a cutting-edge benchmark designed to push LLMs to their limits in real-world data science tasks. Get involved and challenge your models! https://t.co/D3Qvg8OOFS
Tweet media one
3
5
19
@hemingkx
Heming Xia
11 months
๐Ÿค”How much potential do LLMs have for self-acceleration through layer sparsity? ๐Ÿš€ ๐Ÿšจ Excited to share our latest work: SWIFT: On-the-Fly Self-Speculative Decoding for LLM Inference Acceleration. Arxiv: https://t.co/8PfUyyuqN6 ๐Ÿงต1/n
Tweet media one
1
6
16
@qx_dong
Qingxiu Dong
11 months
Paper:
0
0
2
@qx_dong
Qingxiu Dong
11 months
How can we guide LLMs to continually expand their own capabilities with limited annotation? SynPO: a self-boosting paradigm training LLM to auto-learn generative rewards and synthesize preference data. After 4 iterations, Llama3&Mistral achieve over 22.1% win rate improvements
Tweet media one
Tweet media two
3
21
112