ZhaozhuoX Profile Banner
Zhaozhuo Xu Profile
Zhaozhuo Xu

@ZhaozhuoX

Followers
242
Following
596
Media
3
Statuses
102

Assistant Professor, @FollowStevens. Ph.D. @RiceCompSci

Houston, TX
Joined September 2017
Don't wanna be here? Send us removal request.
@ZhaozhuoX
Zhaozhuo Xu
9 months
🚀 Join us at #AAAI2025 for our tutorial: TQ08: KV Cache Compression for Efficient Long Context LLM Inference 📍 Room 116 ⏰ 4:15 PM - 6:00 PM Learn how to compress KV cache for faster, scalable LLM inference! Don't miss it! #AI #LLM #Efficiency #AAAI25
3
21
64
@Anshumali_
Anshumali
4 months
Grateful for everyone’s support and encouragement throughout this journey — family, colleagues, mentors, friends. A special shoutout to all my Ph.D. students — the real heroes behind the scenes (watch out for them, they’re headed for big things)
@RiceCompSci
Rice Computer Science
4 months
Congrats to Rice CS' @Anshumali_ Shrivastava, who has been promoted to full professor. Shrivastava is well on his way to revolutionizing how LLMs & other deep learning models are trained & stored, using new algorithms to make AI scalable & more accessible. https://t.co/8VpFk371gp
0
1
15
@denghui_zhang
Denghui Zhang
4 months
We’re grateful that our recent work on the Theory-of-Mind of LLMs was featured by MIT Technology Review China 🙏 🔗 https://t.co/Evg7wwoDRy In two new preprints, we explore: Sparsity and ToM: How extremely sparse patterns in LLMs shape their ability to infer others’ beliefs and
1
6
14
@yuxin_tang_
Yuxin Tang 🦉
4 months
📣📣📣Excited to meet @Anshumali_ at ICML 2025! I want to promote their work on parameter-efficient fine-tuning (PEFT) for LLMs using sketches. This method offers great results with fewer parameters. @ZhaozhuoX @Tianyi_zha @Apd10Desai
0
2
4
@ran_yide42201
Yide Ran
5 months
🚨 Call for Papers: VISTA Workshop @ ICDM 2025 🚨 📅 Nov 12, 2025 | 📍 Washington, DC Explore GenAI standards, legal constraints, copyright risks, & compliance. Submit by Sep 5! 🔗 https://t.co/MjmZx8UunI Speakers: V. Braverman, D. Atkinson, A. Li #ICDM2025 #GenAI #AIStandards
1
2
5
@InfiniAILab
Infini-AI-Lab
5 months
🔥 We introduce Multiverse, a new generative modeling framework for adaptive and lossless parallel generation. 🚀 Multiverse is the first open-source non-AR model to achieve AIME24 and AIME25 scores of 54% and 46% 🌐 Website: https://t.co/J9osByhWUf 🧵 1/n
6
85
223
@ChaoChenSBU
Chao Chen
6 months
Attending #CVPR2025? Please check out our paper on TopoCellGen. We propose the first diffusion network for cell topology generation, based on topological data analysis (TDA). The paper was selected for an oral presentation, an honor awarded to only about 0.7% of submissions.
0
1
9
@InfiniAILab
Infini-AI-Lab
6 months
🥳 Happy to share our new work –  Kinetics: Rethinking Test-Time Scaling Laws 🤔How to effectively build a powerful reasoning agent? Existing compute-optimal scaling laws suggest 64K thinking tokens + 1.7B model > 32B model. But, It only shows half of the picture! 🚨 The O(N²)
7
73
245
@Anshumali_
Anshumali
6 months
If you are starting at Rice, you can choose to be an AIist. I am very excited by this "uniquely designed" AI curriculum offered by @RiceCompSci. I have penned down my thoughts in here. Feedback welcome. https://t.co/VUp0vGjvuO @RiceEngineering @RiceUniversity
Tweet card summary image
medium.com
I am sharing an exciting update — I’m part of the core design team that is launching one of the handful first few Bachelor of Science…
0
5
18
@HuaweiLin7
Huawei Lin
6 months
Why does autoregressive (AR) image generation still suck — even in the era of LLMs? The core issue might be the Visual Tokenizer (VT). We introduce VTBench, a comprehensive benchmark designed to evaluate 20+ state-of-the-art VTs in the context of AR image generation. Our
1
3
4
@adamse
Adam Seligman
7 months
llama fine tuning is great but have you tried it on https://t.co/DjeYB9EMqI with just a couple of clicks?
2
2
2
@Anshumali_
Anshumali
7 months
Any approximation of LLMs—or loss in reliability—for efficiency gains simply isn’t worth the risk. Introducing "perfect compression of LLMs" Blog: https://t.co/qgs2xSpe3c arxiv:
Tweet card summary image
medium.com
BFloat16 is 5-bits larger than needed!
1
9
31
@ZhaozhuoX
Zhaozhuo Xu
9 months
Excited to give my New Faculty Highlight Talk at AAAI 2025 tomorrow! 🎤 📌 Compression-Aware Computing for Scalable & Sustainable ML 🗓 March 2, 10:00 AM | 📍 Room 121B If you’re at #AAAI2025, let’s connect! 🚀
2
5
43
@henryzhongsc
Henry Zhong
9 months
We really appreciate the great turnout and engagement during our AAAI 25 tutorial! Given that it was a highly technical talk, we’re genuinely glad that the audience received our message well and asked so many great questions. https://t.co/CRUUbZfBs3 Here's the slides as promised,
Tweet card summary image
github.com
Official implementation for Yuan & Liu & Zhong et al., KV Cache Compression, But What Must We Give in Return? A Comprehensive Benchmark of Long Context Capable Approaches. EMNLP Fin...
@ZhaozhuoX
Zhaozhuo Xu
9 months
🚀 Join us at #AAAI2025 for our tutorial: TQ08: KV Cache Compression for Efficient Long Context LLM Inference 📍 Room 116 ⏰ 4:15 PM - 6:00 PM Learn how to compress KV cache for faster, scalable LLM inference! Don't miss it! #AI #LLM #Efficiency #AAAI25
0
2
6
@ZhaozhuoX
Zhaozhuo Xu
9 months
Join us at #AAAI2025 for TQ05: LLMs and Copyright Risks: Benchmarks and Mitigation Approaches! Discover how LLMs intersect with copyright challenges, learn benchmark evaluations, and explore effective mitigation strategies. See you in Room 117, 2:00–3:45 pm!
0
7
14
@InfiniAILab
Infini-AI-Lab
9 months
🚀 RAG vs. Long-Context LLMs: The Real Battle ⚔️ 🤯Turns out, simple-to-build RAG can match million-dollar long-context LLMs (LC LLMs) on most existing benchmarks. 🤡So, do we even need long-context models? YES. Because today’s benchmarks are flawed: ⛳ Too Simple –
6
39
189
@BiologyAIDaily
Biology+AI Daily
10 months
Lemur: A Single-Cell Foundation Model with Fine-Tuning-Free Hierarchical Cell-Type Generation for Drosophila melanogaster 1/ Lemur is a novel single-cell foundation model designed for Drosophila melanogaster, offering a fine-tuning-free approach for accurate cell-type
0
4
8
@RiceCompSci
Rice Computer Science
10 months
How do LLMs work? What is it about the recently-announced DeepSeek technology that has changed the way people think about building large AI models? Join us for the answers to these questions and more. Co-sponsored by Rice CS & @ricekenkennedy. RSVP: https://t.co/jVovSorinR
0
4
8
@hengjinlp
Heng Ji
10 months
A bit late to mention this work led by the amazing Prof. Denghui Zhang @denghui_zhang - invented a novel approach to steer LLM’s internal states to prevent unsafe output. https://t.co/YpxnfUF2KL https://t.co/ezGwn89aGu
0
17
109
@RiceCompSci
Rice Computer Science
10 months
Rice CS' @Anshumali_ Shrivastava & team have developed algorithms to make AI models more efficient & customizable. Their work is critical to how AI could be integrated into solving societal problems while meeting growing energy & computing demands. https://t.co/s0whUDaxDS
0
3
8