Woosuk Kwon Profile
Woosuk Kwon

@woosuk_k

Followers
4K
Following
2K
Media
2
Statuses
259

PhD student at @Berkeley_EECS building @vllm_project

Joined April 2023
Don't wanna be here? Send us removal request.
@woosuk_k
Woosuk Kwon
6 months
As one of the fastest-growing OSS projects, vLLM inevitably accumulated some technical debts. We noticed it, and re-architected vLLM's core with careful engineering. Enjoy simpler code & higher performance with vLLM V1!.
@vllm_project
vLLM
6 months
🚀 With the v0.7.0 release today, we are excited to announce the alpha release of vLLM V1: A major architectural upgrade with 1.7x speedup! .Clean code, optimized execution loop, zero-overhead prefix caching, enhanced multimodal support, and more.
Tweet media one
2
17
210
@woosuk_k
Woosuk Kwon
5 days
RT @zhuohan123: I’ve been fortunate to lead the infra and inference work that brings gpt-oss to life. A year ago, I joined OpenAI after bui….
0
144
0
@woosuk_k
Woosuk Kwon
5 days
Here's the blog post on vLLM's integration:
Tweet card summary image
blog.vllm.ai
We’re thrilled to announce that vLLM now supports gpt-oss on NVIDIA Blackwell and Hopper GPUs, as well as AMD MI300x and MI355x GPUs. In this blog post, we’ll explore the efficient model architecture...
@vllm_project
vLLM
5 days
Thank you @OpenAI for open-sourcing these great models! 🙌.We’re proud to be the official launch partner for gpt-oss (20B & 120B) – now supported in vLLM 🎉.⚡ MXFP4 quant = fast & efficient.🌀 Hybrid attention (sliding + full).🤖 Strong agentic abilities.🚀 Easy deployment.👉🏻.
0
9
73
@woosuk_k
Woosuk Kwon
5 days
RT @vllm_project: Thank you @OpenAI for open-sourcing these great models! 🙌.We’re proud to be the official launch partner for gpt-oss (20B….
0
63
0
@woosuk_k
Woosuk Kwon
2 months
RT @vllm_project: The model is supported in vLLM, welcome to try this powerful model on your own🚀.
0
11
0
@woosuk_k
Woosuk Kwon
2 months
RT @vllm_project: vLLM has just reached 50K github stars! Huge thanks to the community!🚀.Together let's bring easy, fast, and cheap LLM ser….
0
21
0
@woosuk_k
Woosuk Kwon
2 months
RT @p_nawrot: We built sparse-frontier — a clean abstraction that lets you focus on your custom sparse attention implementation while autom….
0
50
0
@woosuk_k
Woosuk Kwon
3 months
RT @AurickQ: Excited to share our work on Speculative Decoding @Snowflake AI Research!. 🚀 4x faster LLM inference for coding agents like Op….
0
38
0
@woosuk_k
Woosuk Kwon
3 months
RT @JustinLin610: Thanks for the quick merge and instant support for our models! Users of vllm and Qwen, feel free to try it out to see whe….
0
12
0
@woosuk_k
Woosuk Kwon
4 months
RT @OpenAIDevs: Announcing the first Codex open source fund grant recipients:. ⬩vLLM - inference serving engine @vllm_project.⬩OWASP Nettac….
0
149
0
@woosuk_k
Woosuk Kwon
4 months
RT @vllm_project: perf update: we are continuing to see benefits with vLLM V1 engine’s highly performant design. on 8xH200, vLLM leads in t….
0
42
0
@woosuk_k
Woosuk Kwon
4 months
RT @vllm_project: vLLM🤝🤗! You can now deploy any @huggingface language model with vLLM's speed. This integration makes it possible for one….
Tweet card summary image
blog.vllm.ai
The Hugging Face Transformers library offers a flexible, unified interface to a vast ecosystem of model architectures. From research to fine-tuning on custom dataset, transformers is the go-to...
0
127
0
@woosuk_k
Woosuk Kwon
4 months
RT @vllm_project: 🙏 @deepseek_ai's highly performant inference engine is built on top of vLLM. Now they are open-sourcing the engine the ri….
Tweet card summary image
github.com
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation - deepseek-ai/open-infra-index
0
349
0
@woosuk_k
Woosuk Kwon
4 months
Huge congrats to all the @googlecloud and @RedHat_AI team members who drove this effort!.
@vllm_project
vLLM
4 months
spotted @vllm_project at @googlecloud next keynote today!
Tweet media one
0
2
63
@woosuk_k
Woosuk Kwon
4 months
RT @hmellor_: Another month, another open-source milestone for the @vllm_project 🎉. now has 1000 contributors 🚀 htt….
0
4
0
@woosuk_k
Woosuk Kwon
4 months
RT @Agentica_: Introducing DeepCoder-14B-Preview - our fully open-sourced reasoning model reaching o1 and o3-mini level on coding and math.….
0
210
0
@woosuk_k
Woosuk Kwon
4 months
RT @yi_xin_dong: XGrammar is accepted to MLSys 2025🎉🎉🎉.It is a widely adopted library for structured generation with LLMs—output clean JSON….
0
18
0
@woosuk_k
Woosuk Kwon
4 months
RT @Ahmad_Al_Dahle: Introducing our first set of Llama 4 models!. We’ve been hard at work doing a complete re-design of the Llama series. I….
0
934
0
@woosuk_k
Woosuk Kwon
4 months
RT @robertnishihara: If you're using vLLM + Ray for batch inference or online serving, check this out. We're investing heavily in making th….
0
1
0
@woosuk_k
Woosuk Kwon
5 months
RT @casper_hansen_: TRL now handles multi-node training with vLLM for GRPO🤯
Tweet media one
0
23
0