Yonglong Tian Profile
Yonglong Tian

@YonglongT

Followers
3K
Following
315
Media
4
Statuses
91

Research Scientist @OpenAI. Previously Researcher @GoogleDeepMind, PhD @MIT. Opinions are my own.

Boston, MA
Joined June 2019
Don't wanna be here? Send us removal request.
@YonglongT
Yonglong Tian
7 months
RT @shobsund: Personal vision tasks–like detecting *your mug*-are hard; they’re data scarce and fine-grained. In our new paper, we show y….
0
59
0
@YonglongT
Yonglong Tian
9 months
RT @lijie_fan: 🚀 Excited to share our latest work Fluid!. We've developed a scalable autoregressive text-to-image model without VQ. We trai….
0
22
0
@YonglongT
Yonglong Tian
9 months
We name our Fluid model from 150M upto 10B! Surprisingly, Fluid with only 300M achieves similar FID as prior model with billions of parameters, e.g. Parti-20B. Joint work with @lijie_fan, @TianhongLi6, Siyang Qin, Yuanzhen Li, @jesu9, @MikiRubinstein, @DeqingSun, and Kaiming He.
2
1
21
@YonglongT
Yonglong Tian
9 months
Do we still need codebook/quantization for scalable autoregressive visual generation?. No! Thrilled to share our latest work on scaling w/ continuous tokens. We observe power-law scaling behavior on val loss, and obtain SOTA coco FID and GenEval score.
Tweet media one
Tweet media two
6
46
280
@YonglongT
Yonglong Tian
1 year
RT @JiaweiYang118: Very excited to get this out: “DVT: Denoising Vision Transformers”. We've identified and combated those annoying positio….
0
80
0
@YonglongT
Yonglong Tian
1 year
RT @phillip_isola: Our computer vision textbook is released!. Foundations of Computer Vision.with Antonio Torralba and Bill Freeman.https:/….
0
401
0
@YonglongT
Yonglong Tian
2 years
Thank you @_akhaliq for featuring our work!.
@_akhaliq
AK
2 years
Denoising Vision Transformers. paper page: identify crucial artifacts in ViTs caused by positional embeddings and propose a two-stage approach to remove these artifacts, which significantly improves the feature quality of different pre-trained ViTs
Tweet media one
0
6
41
@YonglongT
Yonglong Tian
2 years
HNY! Excited to share SynCLR, that rivals CLIP and Dino v2 but uses pure synthetic data. The interesting part - it can outperform models (e.g. CLIP) directly trained on LAION-2B, which was the dataset used to train SD 1.5 that we used to generate images.
Tweet media one
5
42
282
@YonglongT
Yonglong Tian
2 years
RT @lijie_fan: 🚀 Is the future of vision models Synthetic? Introducing SynCLR: our new pipeline leveraging LLMs & Text-to-image models to t….
0
41
0
@YonglongT
Yonglong Tian
2 years
I had the joy of working with Olivier (and Aaron) at DeepMind. My best internship experience. Strongly recommended!.
@olivierhenaff
Olivier Hénaff
2 years
Thrilled to announce that we have an opening for a Student Researcher to come work with us at @GoogleDeepMind!. If you’re interested in multimodal learning, in-context adaptation, memory-augmented perception, or active learning, do consider applying:
0
0
6
@YonglongT
Yonglong Tian
2 years
Thank you @_akhaliq for covering our work!.
@_akhaliq
AK
2 years
Leveraging Unpaired Data for Vision-Language Generative Models via Cycle Consistency. paper page: Current vision-language generative models rely on expansive corpora of paired image-text data to attain optimal performance and generalization capabilities.
Tweet media one
0
4
29
@YonglongT
Yonglong Tian
2 years
RT @TongzhouWang: Quasimetric RL code is now on GitHub: Instead of deleting 80% of the dev repo, I rewrote the alg….
0
24
0
@YonglongT
Yonglong Tian
2 years
RT @sangnie: Join us at the WiML Un-Workshop breakout session on "Role of Mentorship and Networking"! Do not miss the chance to talk with l….
0
22
0
@YonglongT
Yonglong Tian
2 years
RT @Jing36645824: 🎉(1/6) Exciting News:🐑LAMM is online!. ⭐️Features:.① 200k 2D/3D Instruction tuning dataset.② Benchmark on 14 high-level 2….
0
2
0
@YonglongT
Yonglong Tian
2 years
Our new work led by elegant @xuyilun2 , Mingyang and Xiang.
@xuyilun2
Yilun Xu
2 years
In diffusion models, samplers are primarily ODE-centric, overlooking slower stochastic methods. However, we show that stochastic sampler can outperform previous samplers on Stable Diffusion, if we use stochasticity correctly!. check out Restart Sampling:
0
1
12
@YonglongT
Yonglong Tian
2 years
MIT is a place for serious research.
@jacobandreas
Jacob Andreas
2 years
Tweet media one
0
2
50
@YonglongT
Yonglong Tian
2 years
RT @dilipkay: New paper!! We show that pre-training language-image models *solely* on synthetic images from Stable Diffusion can outperform….
0
112
0
@YonglongT
Yonglong Tian
2 years
This paper is jointly done w/ @lijie_fan, @dilipkay, @phillip_isola, and Huiwen Chang.
@YonglongT
Yonglong Tian
2 years
Today marks the official ending of my PhD life at MIT. So grateful to this journey. Coincidentally, we arXiv a paper today: It shows the potential of learning from synthetic data. This coincidence nicely concludes my PhD life in an academic manner.
Tweet media one
Tweet media two
0
0
6
@YonglongT
Yonglong Tian
2 years
Today marks the official ending of my PhD life at MIT. So grateful to this journey. Coincidentally, we arXiv a paper today: It shows the potential of learning from synthetic data. This coincidence nicely concludes my PhD life in an academic manner.
Tweet media one
Tweet media two
23
20
318
@YonglongT
Yonglong Tian
2 years
RT @dilipkay: New paper! We show how to leverage pre-trained LLMs (ChatGPT, Bard, LLaMa) to rewrite captions, and significantly improve ove….
0
28
0