Weixin Liang
@liang_weixin
Followers
1K
Following
103
Media
29
Statuses
154
CS Ph.D. @Stanford | @StanfordAILab | TA for CS224C: NLP for Computational Social Science | Exploring AI & NLP | https://t.co/pOjcCS4gUk
Palo Alto, CA
Joined November 2019
๐ Excited to share: "๐๐ข๐ฑ๐ญ๐ฎ๐ซ๐-๐จ๐-๐๐ซ๐๐ง๐ฌ๐๐จ๐ซ๐ฆ๐๐ซ๐ฌ (๐๐จ๐)" has been officially accepted to TMLR (March 2025) and the code is now open-sourced! ๐ GitHub repo: https://t.co/KiDbxpDWt0 ๐ Paper: https://t.co/KQoZ3cunEf How can we reduce pretraining costs for
2
83
435
Even the smartest LLMs can fail at basic multiturn communication Ask for grocery help โ without asking where you live ๐คฆโโ๏ธ Ask to write articles โ assumes your preferences ๐คท๐ปโโ๏ธ โญ๏ธCollabLLM (top 1%; oral @icmlconf) transforms LLMs from passive responders into active collaborators.
9
65
209
Thank you, @VictoriaLinML , for the write-up.
Let's talk about Mixture-of-Transformers (MoT) and heterogeneous omni-model training. 1. Inspired by prior architectures consisting of modality-specific parametersโsuch as Flamingo, CogVLM, BEIT-3, and MoMAโMoT ( https://t.co/1LMdVZkZdN) pushes this idea further by using
1
0
11
๐ Excited to share the most inspiring work Iโve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
84
512
3K
๐ On United Nations (UN) adoption: Even the world's most prominent international bodies are embracing LLMs! UN press releases showed a rapid initial surge (3.1% to 10.1%) in early 2023, then steadily climbing to 13.7% by Q3 2024.
1
1
13
0
0
3
๐ Key findings: - Lower education areas showed higher LLM adoption in consumer complaints - Urban areas have higher LLM usage (18.2% vs 10.9%) - Science & tech companies lead in corporate adoption - Younger firms (post-2015) use LLMs 3x more than older ones (pre-1980)
0
0
9
๐จ New research: We analyzed 1.5M+ documents to track LLM-assisted writing adoption across society from 2022-2024. The results? ๐By late 2024, LLMs assist in writing: - 18% of financial consumer complaints - 24% of corporate press releases - Up to 15% of job postings (esp. in
5
37
122
Thanks @liang_weixin We all enjoyed reading the paper! And we appreciate your paper for helping the community gain a deeper understanding of the modality gap ๐ฅฐ
Glad to see our Modality Gap paper's insights reflected in Voyage AI's new state-of-the-art multimodal embedding model! @VoyageAI @kaidicao
https://t.co/DN2IMQ9BAi
1
1
9
We are excited to announce that Voyage AI is officially joining @MongoDB ! Joining @MongoDB enables us to bring our cutting-edge AI retrieval technology to a broader audience and seamlessly integrate it into mission-critical applications. Learn more: https://t.co/V8PTq3v5ZM
4
11
91
Update: check out https://t.co/U94zIRGSoj for our code, data, and model!
github.com
The official implementation of "Self-play LLM Theorem Provers with Iterative Conjecturing and Proving" - kfdong/STP
and SoTA among whole-proof generation methods on miniF2F, ProofNet, and PutnamBench, and double the previous best results on LeanWorkBook. (reposting because it seems that this table has much more views ๐)
1
9
42
We introduce Mixture-of-Mamba, a multi-modal SSM that leverages modality-aware sparsity for efficient multi-modal pretraining! At the core of Mixture-of-Mamba: ๐นModality-aware sparsity to optimize efficiency ๐นMixture-of-SSMs to enable cross-modal interactions ๐นScales
๐ Want 2x faster pretraining for your multi-modal LLM? ๐งต Following up on Mixture-of-Transformers (MoT), we're excited to share Mixture-of-Mamba (MoM)! https://t.co/OTTpAlB4Vq ๐ฅ Why it matters: MoM applies modality-aware sparsity across image, text, and speechโmaking
0
8
38
๐ Want 2x faster pretraining for your multi-modal LLM? ๐งต Following up on Mixture-of-Transformers (MoT), we're excited to share Mixture-of-Mamba (MoM)! https://t.co/OTTpAlB4Vq ๐ฅ Why it matters: MoM applies modality-aware sparsity across image, text, and speechโmaking
0
2
19
๐ข Can LLMs program themselves to run faster? ๐โฑ๏ธ LLM self-taught to code for next-gen AI hardware! https://t.co/wiwgiPEpeH 1/ Programming AI accelerators is a major bottleneck in ML. Our self-improving LLM agent learns to write optimized code for new hardware, achieving 3.9x
2
7
38
๐ ML library development is crucial but requires expertise in ML algorithms & architecture-specific programming languages (ASPLs). ๐ค LLM agents can enable better automation. We propose an adaptive self-improvement agentic system for generating ML libraries in STePโa
2
7
26
๐ Vision language models are getting better - but how do we evaluate them reliably? Introducing AutoConverter: transforming open-ended VQA into challenging multiple-choice questions! Key findings: 1๏ธโฃ Current open-ended VQA eval methods are flawed: rule-based metrics correlate
3
73
154
Introducing ๐๐ฅ๐๐ฆ๐๐
๐ฎ๐ฌ๐ข๐จ๐ง: empowering Llama ๐ฆ with diffusion ๐จ to understand and generate text and images in arbitrary sequences. โจ Building upon Transfusion, our recipe fully preserves Llamaโs language performance while unlocking its multimodal understanding and
15
188
897
๐ค Why are VLMs (even GPT-4V) worse at image classification than CLIP, despite using CLIP as their vision encoder? Presenting VLMClassifier at #NeurIPS2024: โฐ Dec 11 (Wed), 11:00-14:00 ๐ East Hall #3710 Key findings: 1๏ธโฃ VLMs dramatically underperform CLIP (>20% gap) 2๏ธโฃ After
2
21
87
Iโm open to academia & industry in 2025. My work in #XR ๐ฅฝ + #HCI ๐ฉโ๐ป enables low-friction XR experience thru #EmbodiedInteraction, unlocking potential for all -- tech-savvy or not ๐ Design+Science+Engineering. Let's shape the future of spatial computing โจ RT appreciated! (1/8)
3
22
89
Honored that @Nature has highlighted our work again in their latest piece examining #ChatGPT's transformative impact on scientific research and academia over the past two years. h/t @Nature
https://t.co/wK4ayZYH9w
1
1
17