
Weixin Liang
@liang_weixin
Followers
1K
Following
107
Media
29
Statuses
155
CS Ph.D. @Stanford | @StanfordAILab | TA for CS224C: NLP for Computational Social Science | Exploring AI & NLP | https://t.co/pOjcCS4gUk
Palo Alto, CA
Joined November 2019
RT @ShirleyYXWu: Even the smartest LLMs can fail at basic multiturn communication. Ask for grocery help โ without asking where you liveโฆ.
0
47
0
Thank you, @VictoriaLinML , for the write-up.
Let's talk about Mixture-of-Transformers (MoT) and heterogeneous omni-model training. 1. Inspired by prior architectures consisting of modality-specific parametersโsuch as Flamingo, CogVLM, BEIT-3, and MoMAโMoT ( pushes this idea further by using.
1
0
10
RT @xuandongzhao: ๐ Excited to share the most inspiring work Iโve been part of this year:. "Learning to Reason without External Rewards"โฆ.
0
510
0
RT @VoyageAI: Thanks @liang_weixin We all enjoyed reading the paper! And we appreciate your paper for helping the community gain a deeperโฆ.
0
1
0
RT @JunhongShen1: We introduce Mixture-of-Mamba, a multi-modal SSM that leverages modality-aware sparsity for efficient multi-modal pretraiโฆ.
0
5
0
RT @zhang677: ๐ ML library development is crucial but requires expertise in ML algorithms & architecture-specific programming languages (ASโฆ.
0
6
0
RT @Zhang_Yu_hui: ๐ Vision language models are getting better - but how do we evaluate them reliably? Introducing AutoConverter: transformiโฆ.
0
73
0
RT @WeijiaShi2: Introducing ๐๐ฅ๐๐ฆ๐๐
๐ฎ๐ฌ๐ข๐จ๐ง: empowering Llama ๐ฆ with diffusion ๐จ to understand and generate text and images in arbitrary sequenโฆ.
0
175
0
RT @Zhang_Yu_hui: ๐ค Why are VLMs (even GPT-4V) worse at image classification than CLIP, despite using CLIP as their vision encoder?. Presenโฆ.
0
21
0
RT @SiyouPei: Iโm open to academia & industry in 2025. My work in #XR ๐ฅฝ + #HCI ๐ฉโ๐ป enables low-friction XR experience thru #EmbodiedInteracโฆ.
0
23
0