Sai__Bi Profile Banner
Sai Bi Profile
Sai Bi

@Sai__Bi

Followers
423
Following
215
Media
7
Statuses
108

Research Scientist @ Adobe Research

San Jose, CA
Joined October 2011
Don't wanna be here? Send us removal request.
@Sai__Bi
Sai Bi
29 days
Check out the cool work by our intern @HanshengCh on policy-based distillation for few-step generation.
@HanshengCh
Hansheng Chen
29 days
Excited to announce a new track of accelerating Generative AI: pi-Flow: Policy-Based Few-Step Generation via Imitation Distillation https://t.co/6ro55E1XGP Distill 20B flow models now using just an L2 loss via imitation learning for SOTA diversity and teacher-aligned quality.
0
0
7
@bowei_chen_19
Bowei Chen
2 months
We found that visual foundation encoder can be aligned to serve as tokenizers for latent diffusion models in image generation! Our new paper introduces a new tokenizer training paradigm that produces a semantically rich latent space, improving diffusion model performancešŸš€šŸš€.
7
71
529
@percyliang
Percy Liang
5 months
Wrapped up Stanford CS336 (Language Models from Scratch), taught with an amazing team @tatsu_hashimoto @marcelroed @neilbband @rckpudi. Researchers are becoming detached from the technical details of how LMs work. In CS336, we try to fix that by having students build everything:
46
595
5K
@Sai__Bi
Sai Bi
5 months
I am going to give a talk on scalable 3D reconstructions today at the 3D-LLM/VLA workshop at CVPR at 10:55am today at Room 106A. Welcome to attend!
Tweet card summary image
3d-llm-vla.github.io
Bridging Language, Vision and Action in 3D Environments. Join us at CVPR 2025 in Nashville, TN, USA to explore the integration of language and 3D perception.
1
1
28
@flycooler_zd
Zhao Dong
5 months
šŸš€ Excited to announce our CVPR 2025 Workshop: 3D Digital Twin: Progress, Challenges, and Future Directions šŸ—“ June 12, 2025 Ā· 9:00 AM–5:00 PM šŸ“¢ Incredible lineup: @rapideRobot, Andrea Vedaldi @Oxford_VGG,@richardzhangsfu,@QianqianWang5,Dr. Xiaoshuai Zhang @Hillbot_AI,
2
24
58
@tianyuanzhang99
Tianyuan Zhang
6 months
Bored of linear recurrent memories (e.g., linear attention) and want a scalable, nonlinear alternative? Our new paperĀ ā€œTest-Time Training Done Rightā€ proposeĀ LaCT (Large Chunk Test-Time Training) — a highly efficient, massively scalable nonlinear memory with: šŸ’”Ā Pure PyTorch
5
85
426
@Haian_Jin
Haian Jin
7 months
Excited to attend #ICLR2025 in person this year! I’ll be presenting two papers: 1. LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias šŸ”¹ Oral Presentation: Session 3C (Garnet 216-218) — Apr 25 (Fri), 11:06–11:18 a.m. šŸ”¹ Poster: Hall 3 + Hall 2B, Poster #593 — Apr
1
3
27
@Sai__Bi
Sai Bi
7 months
I will be attending ICLR in Singapore this week. Feel free to reach out and chat!
0
0
23
@Sai__Bi
Sai Bi
7 months
Check out the fantastic work by our intern @HanshengCh at Adobe Research. The code and model are publicly available!
@HanshengCh
Hansheng Chen
7 months
Excited to share our work: Gaussian Mixture Flow Matching Models (GMFlow) https://t.co/XWAy2VCJlg GMFlow generalizes diffusion models by predicting Gaussian mixture denoising distributions, enabling precise few-step sampling and high-quality generation.
1
0
18
@Haian_Jin
Haian Jin
7 months
Our paper LVSM has been accepted as an oral presentation at #ICLR2025! See you in Singapore! We’ve just released the code and checkpoints—check it out here: https://t.co/07Px6Rt2Jn.šŸš€
Tweet card summary image
github.com
[ICLR 2025 Oral] Official code for "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias" - Haian-Jin/LVSM
@Haian_Jin
Haian Jin
1 year
Novel view synthesis has long been a core challenge in 3D vision. But how much 3D inductive bias is truly needed? —Surprisingly, very little! Introducing "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias"—a fully transformer-based approach that enables scalable,
2
23
129
@Sai__Bi
Sai Bi
11 months
The speaker was fully aware of the implications of her words and the damage they would cause. Yet, instead of preventing harm, she chose to inflict it first and then attempt to repair it with some 'nice' words. That’s not acceptable!
@sunjiao123sun_
Jiao Sun
11 months
Mitigating racial bias from LLMs is a lot easier than removing it from humans! Can’t believe this happened at the best AI conference @NeurIPSConf We have ethical reviews for authors, but missed it for invited speakers? 😔
0
0
20
@Sai__Bi
Sai Bi
11 months
Check out the latest work led by @hanzhe_hu. Turbo3D achieves high-quality text-to-3D generation within 0.35 seconds.
@hanzhe_hu
Hanzhe Hu
11 months
Text-to-image generation can already generate high-quality results in the blink of an eye, while text-to-3D still requires a much longer time. How do we bridge this gap? Introducing "Turbo3D: Ultra-fast Text-to-3D Generationā€ — Ultra-fast high-quality text-to-3D generation in
0
0
9
@gene_ch0u
Gene Chou
1 year
We've released our paper "Generating 3D-Consistent Videos from Unposed Internet Photos"! Video models like Luma generate pretty videos, but sometimes struggle with 3D consistency. We can do better by scaling them with 3D-aware objectives. 1/N page: https://t.co/Hgu8uo3tvu
6
47
230
@Haian_Jin
Haian Jin
1 year
Novel view synthesis has long been a core challenge in 3D vision. But how much 3D inductive bias is truly needed? —Surprisingly, very little! Introducing "LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias"—a fully transformer-based approach that enables scalable,
23
95
577
@zhenjun_zhao
Zhenjun Zhao
1 year
LVSM: A Large View Synthesis Model with Minimal 3D Inductive Bias @Haian_Jin, @hanwenjiang1, @HaoTan5, @KaiZhang9546, @Sai__Bi, @tianyuanzhang99, @fujun_luan, @Jimantha, @zexiangxu tl;dr: purely transformer-based large view synthesis https://t.co/bMmqX4fbq1
2
10
59
@KaiZhang9546
Kai Zhang
1 year
We have formed a foundation team in Adobe to work on video foundation models with @jianming_zhang_@Sai__Bi@fujun_luan. I’m excited to see the non-parametric side of 3D: an AI model with strong spatial-temporal capability, besides the existing parametric 3d representations!
3
8
129
@chenziwee
Ziwen Chen
1 year
Hate waiting 10 minutes for 3D GS to render your favorite indoor or outdoor scenes? ā³ Our feed-forward solution, Long-LRM, cuts it down to just 1 second! āš”ļø With a straightforward mix of Mamba2 and transformer, it scales up to 32 high-res input images. https://t.co/brAgawmtV3
5
38
209
@Sai__Bi
Sai Bi
1 year
I will be presenting our work on applying large reconstruction model for Gaussian Splsting ( https://t.co/RyXzojjMb0) from sparse images at #ECCV2024 in Milan. Welcome to come by our poster at stand 320 on Thursday morning!
2
3
52
@YaoQin_UCSB
Yao Qin
1 year
🄰 Super excited to share this new work on benchmarking LLMs for carbohydrate estimation, which is a huge daily burden that every patient with diabetes needs to deal with multiple times every day. šŸ‘šŸ‘Proud of my students for starting to investigate the potential of LLMs in
2
9
38