Anji Liu Profile
Anji Liu

@liu_anji

Followers
254
Following
72
Media
7
Statuses
39

Incoming Assistant Professor (Presidential Young Professor, PYP) at the National University of Singapore (NUS).

Singapore
Joined October 2019
Don't wanna be here? Send us removal request.
@liu_anji
Anji Liu
2 months
🎓 Looking for PhD students, postdocs & interns!.I’m recruiting for my new lab at @NUSComputing, focusing on generative modeling, reasoning, and tractable inference. 💡 Interested? Learn more here: 🗓️ PhD application deadline: June 15, 2025.
0
3
28
@liu_anji
Anji Liu
5 months
RT @danielmisrael: “That’s one small [MASK] for [MASK], a giant [MASK] for mankind.” – [MASK] Armstrong. Can autoregressive models predict….
0
8
0
@liu_anji
Anji Liu
5 months
RT @Vinh_Suhi: 🚀 Exciting news! Our paper "Learning to Discretize Diffusion ODEs" has been accepted as an Oral at #ICLR2025! 🎉. [1/n].We pr….
0
6
0
@liu_anji
Anji Liu
8 months
RT @zhezeng0908: 📢 I’m recruiting PhD students @CS_UVA for Fall 2025!.🎯 Neurosymbolic AI, probabilistic ML, trustworthiness, AI for science….
0
71
0
@liu_anji
Anji Liu
8 months
RT @StephanMandt: My group will be hiring several PhD students next year. I can’t reveal the details before the official announcement at Ne….
0
42
0
@liu_anji
Anji Liu
9 months
RT @gengala13: I’ll be attending #NeurIPS2024, where I’ll present our spotlight: Scaling Continuous Latent Variable Models as Probabilistic….
0
21
0
@liu_anji
Anji Liu
9 months
RT @EmilevanKrieken: @MinkaiX Very cool work! It sounds quite similar to Discrete Copula Diffusion by @liu_anji, in that it combines both p….
0
2
0
@liu_anji
Anji Liu
9 months
Shout-out to my amazing collaborators @OliverBroadrick @Mniepert @guyvdb !!.
0
0
4
@liu_anji
Anji Liu
9 months
[9/n] In addition to presenting an effective discrete diffusion generation algorithm, we emphasize the importance of modeling inter-variable dependencies in discrete diffusion. Please check the details in our paper
1
0
4
@liu_anji
Anji Liu
9 months
[8/n] In particular, by combining a recent discrete diffusion model ( with GPT-2 (small) as the copula model, we achieve better (un)conditioned generation performance in 8-32 fewer denoising steps.
1
0
4
@liu_anji
Anji Liu
9 months
[7/n] Since discrete diffusion models learn all univariate marginals well, we use a copula model to additionally capture the dependency information. As shown on the right side of the figure, by combining information from both models, we generate better samples in fewer steps.
Tweet media one
1
0
4
@liu_anji
Anji Liu
9 months
[6/n] Given the prompt: “Let’s do outdoor sports! How about <X> <Y>?” The copula indicates the phrases “alpine skiing” and “scuba diving” happen more often than the others (e.g., “alpine diving”), while the univariate marginals indicate which phrase appears more often.
Tweet media one
1
0
4
@liu_anji
Anji Liu
9 months
[5/n] How to fix this problem? We resort to the classical work of Sklar ( that every distribution can be decomposed into two disjoint sets of information: (i) a set of univariate marginals and (ii) a copula.
1
0
7
@liu_anji
Anji Liu
9 months
[4/n] We quantify the performance drop by showing that there is an irreducible component in the negative ELBO stemming from this independent denoising assumption, which prevents the model from generating high-quality samples in just a few steps even if the model is perfect.
Tweet media one
1
0
7
@liu_anji
Anji Liu
9 months
[3/n] As shown in the left side of the figure, discrete diffusion models predict the token probabilities individually for each variable. When multiple “edits” are made in a single denoising step, the model fails to consider the probability of these changes happening jointly.
Tweet media one
1
0
10
@liu_anji
Anji Liu
9 months
[2/n] This problem can be mitigated by supplementing the missing dependency information with another *copula model* (a deep generative model). Check out details about this problem and our solution in our paper Discrete Copula Diffusion
1
0
12
@liu_anji
Anji Liu
9 months
[1/n] 🚀Diffusion models for discrete data excel at modeling text, but they need hundreds to thousands of diffusion steps to perform well. We show that this is caused by the fact that discrete diffusion models predict each output token *independently* at each denoising step.
Tweet media one
5
33
212
@liu_anji
Anji Liu
11 months
RT @PoorvaGarg11: Are you looking for an inference algorithm that supports your discrete-continuous probabilistic program? Look no further!….
0
11
0
@liu_anji
Anji Liu
1 year
RT @KareemYousrii: I'm very happy to share PyJuice, the culmination of years of scaling up probabilistic & logical reasoning for neural net….
0
1
0
@liu_anji
Anji Liu
1 year
Interested in the technical details of PyJuice? Our paper has been accepted to #ICML24 and is available on Arxiv Many thanks to my amazing collaborators @KareemYousrii and @guyvdb, without whom this project would not be possible.
0
0
4