Xiaochuang Han Profile
Xiaochuang Han

@XiaochuangHan

Followers
692
Following
1K
Media
16
Statuses
103

PhD student at the University of Washington

Seattle, WA
Joined February 2016
Don't wanna be here? Send us removal request.
@XiaochuangHan
Xiaochuang Han
2 days
RT @shangbinfeng: 👀 How to find more difficult/novel/salient evaluation data?.✨ Let the data generators find it for you!. Introducing Data….
0
15
0
@XiaochuangHan
Xiaochuang Han
1 month
RT @orevaahia: 🎉 We’re excited to introduce BLAB: Brutally Long Audio Bench, the first benchmark for evaluating long-form reasoning in audi….
0
47
0
@XiaochuangHan
Xiaochuang Han
1 month
Check out our work led by @Cumquaaa on a hybrid autoregressive-diffusion architecture for image generation -- it flexibly balances the number of autoregressive and diffusion layers for optimal generation quality and inference speed! Autoregressive vs. diffusion -- you don't have.
@Cumquaaa
Junhao Chen
2 months
🚀 Training an image generation model and picking sides between autoregressive (AR) and diffusion? Why not both? Check out MADFormer with half of the model layers for AR and half for diffusion. AR gives a fast guess for the next patch prediction while diffusion helps refine the
Tweet media one
0
6
21
@XiaochuangHan
Xiaochuang Han
1 month
RT @WeijiaShi2: Can data owners & LM developers collaborate to build a strong shared model while each retaining data control?. Introducing….
0
87
0
@XiaochuangHan
Xiaochuang Han
2 months
RT @ZEYULIU10: LLMs trained to memorize new facts can’t use those facts well.🤔. We apply a hypernetwork to ✏️edit✏️ the gradients for fact….
0
66
0
@XiaochuangHan
Xiaochuang Han
3 months
RT @lpjiang97: I'm excited to share our latest work — "Data Heterogeneity Limits the Scaling Effect of Pretraining in Neural Data Transform….
Tweet card summary image
biorxiv.org
A key challenge in analyzing neuroscience datasets is the profound variability they exhibit across sessions, animals, and data modalities–i.e., heterogeneity. Several recent studies have demonstrated...
0
13
0
@XiaochuangHan
Xiaochuang Han
4 months
RT @liujc1998: Today we're unveiling OLMoTrace, a tool that enables everyone to understand the outputs of LLMs by connecting to their train….
0
48
0
@XiaochuangHan
Xiaochuang Han
6 months
RT @michiyasunaga: 📢 Introducing Multimodal RewardBench:. A holistic, human-annotated benchmark for evaluating VLM reward models or judges….
0
37
0
@XiaochuangHan
Xiaochuang Han
8 months
RT @WeijiaShi2: Introducing 𝐋𝐥𝐚𝐦𝐚𝐅𝐮𝐬𝐢𝐨𝐧: empowering Llama 🦙 with diffusion 🎨 to understand and generate text and images in arbitrary sequen….
0
180
0
@XiaochuangHan
Xiaochuang Han
11 months
RT @alisawuffles: very excited to attend the first @COLM_conf!🤩 I will be presenting proxy-tuning as a spotlight talk🔦 on Wed @ 10am. I wo….
0
10
0
@XiaochuangHan
Xiaochuang Han
11 months
RT @vidhisha_b: On some personal news: I joined MSR AI Frontiers a few months ago and am very excited to share my first work with this amaz….
Tweet card summary image
arxiv.org
Rigorous and reproducible evaluation is critical for assessing the state of the art and for guiding scientific advances in Artificial Intelligence. Evaluation is challenging in practice due to...
0
11
0
@XiaochuangHan
Xiaochuang Han
1 year
RT @shangbinfeng: Check out JPEG-LM pleaseeeeeee.
0
2
0
@XiaochuangHan
Xiaochuang Han
1 year
RT @tsvetshop: Check out JPEG-LM, amazing and transformative idea!.
0
2
0
@XiaochuangHan
Xiaochuang Han
1 year
RT @PangWeiKoh: Check out JPEG-LM, a fun idea led by @XiaochuangHan -- we generate images simply by training an LM on raw JPEG bytes and sh….
0
4
0
@XiaochuangHan
Xiaochuang Han
1 year
Check out our paper ( for more details! Apart from the experiments and analyses of JPEG-LM, we also include proof-of-concept video generation experiments using canonical video codecs (AVC/H.264). Joint work with @gh_marjan , @PangWeiKoh , and Yulia.
Tweet card summary image
arxiv.org
Recent work in image and video generation has been adopting the autoregressive LLM architecture due to its generality and potentially easy integration into multi-modal systems. The crux of...
0
1
8
@XiaochuangHan
Xiaochuang Han
1 year
The style of this work follows ImageGPT ( but the use of compression algorithms like JPEG greatly reduces the length constraint (e.g., by 40x). Compared to vector quantization methods, JPEG-LM is extremely simple (without any sophisticated tokenizer
Tweet media one
1
0
2
@XiaochuangHan
Xiaochuang Han
1 year
👽Have you ever accidentally opened a .jpeg file with a text editor (or a hex editor)?. Your language model can learn from these seemingly gibberish bytes and generate images with them!. Introducing *JPEG-LM* - an image generator that uses exactly the same architecture as LLMs
Tweet media one
Tweet media two
4
39
157
@XiaochuangHan
Xiaochuang Han
1 year
RT @violet_zct: Introducing *Transfusion* - a unified approach for training models that can generate both text and images. .
0
209
0
@XiaochuangHan
Xiaochuang Han
1 year
RT @liliyu_lili: 🚀 Excited to share our latest work: Transfusion! A new multi-modal generative training combining language modeling and ima….
0
17
0