liang_weixin Profile Banner
Weixin Liang Profile
Weixin Liang

@liang_weixin

Followers
1K
Following
103
Media
29
Statuses
154

CS Ph.D. @Stanford | @StanfordAILab | TA for CS224C: NLP for Computational Social Science | Exploring AI & NLP | https://t.co/pOjcCS4gUk

Palo Alto, CA
Joined November 2019
Don't wanna be here? Send us removal request.
@liang_weixin
Weixin Liang
6 months
๐ŸŽ‰ Excited to share: "๐Œ๐ข๐ฑ๐ญ๐ฎ๐ซ๐ž-๐จ๐Ÿ-๐“๐ซ๐š๐ง๐ฌ๐Ÿ๐จ๐ซ๐ฆ๐ž๐ซ๐ฌ (๐Œ๐จ๐“)" has been officially accepted to TMLR (March 2025) and the code is now open-sourced! ๐Ÿ“Œ GitHub repo: https://t.co/KiDbxpDWt0 ๐Ÿ“„ Paper: https://t.co/KQoZ3cunEf How can we reduce pretraining costs for
2
83
435
@ShirleyYXWu
Shirley Wu
5 months
Even the smartest LLMs can fail at basic multiturn communication Ask for grocery help โ†’ without asking where you live ๐Ÿคฆโ€โ™€๏ธ Ask to write articles โ†’ assumes your preferences ๐Ÿคท๐Ÿปโ€โ™€๏ธ โญ๏ธCollabLLM (top 1%; oral @icmlconf) transforms LLMs from passive responders into active collaborators.
9
65
209
@liang_weixin
Weixin Liang
5 months
Thank you, @VictoriaLinML , for the write-up.
@VictoriaLinML
Victoria X Lin
5 months
Let's talk about Mixture-of-Transformers (MoT) and heterogeneous omni-model training. 1. Inspired by prior architectures consisting of modality-specific parametersโ€”such as Flamingo, CogVLM, BEIT-3, and MoMAโ€”MoT ( https://t.co/1LMdVZkZdN) pushes this idea further by using
1
0
11
@xuandongzhao
Xuandong Zhao
5 months
๐Ÿš€ Excited to share the most inspiring work Iโ€™ve been part of this year: "Learning to Reason without External Rewards" TL;DR: We show that LLMs can learn complex reasoning without access to ground-truth answers, simply by optimizing their own internal sense of confidence. 1/n
84
512
3K
@liang_weixin
Weixin Liang
8 months
๐ŸŒ On United Nations (UN) adoption: Even the world's most prominent international bodies are embracing LLMs! UN press releases showed a rapid initial surge (3.1% to 10.1%) in early 2023, then steadily climbing to 13.7% by Q3 2024.
1
1
13
@liang_weixin
Weixin Liang
8 months
Work done in collaboration w/@yaohuiz3, @m_codreanu, Jiayu Wang, @CaoHancheng, @james_y_zou
0
0
3
@liang_weixin
Weixin Liang
8 months
๐Ÿ” Key findings: - Lower education areas showed higher LLM adoption in consumer complaints - Urban areas have higher LLM usage (18.2% vs 10.9%) - Science & tech companies lead in corporate adoption - Younger firms (post-2015) use LLMs 3x more than older ones (pre-1980)
0
0
9
@liang_weixin
Weixin Liang
8 months
๐Ÿšจ New research: We analyzed 1.5M+ documents to track LLM-assisted writing adoption across society from 2022-2024. The results? ๐Ÿ“ŠBy late 2024, LLMs assist in writing: - 18% of financial consumer complaints - 24% of corporate press releases - Up to 15% of job postings (esp. in
5
37
122
@VoyageAI
Voyage AI by MongoDB
1 year
Thanks @liang_weixin We all enjoyed reading the paper! And we appreciate your paper for helping the community gain a deeper understanding of the modality gap ๐Ÿฅฐ
@liang_weixin
Weixin Liang
1 year
Glad to see our Modality Gap paper's insights reflected in Voyage AI's new state-of-the-art multimodal embedding model! @VoyageAI @kaidicao https://t.co/DN2IMQ9BAi
1
1
9
@VoyageAI
Voyage AI by MongoDB
8 months
We are excited to announce that Voyage AI is officially joining @MongoDB ! Joining @MongoDB enables us to bring our cutting-edge AI retrieval technology to a broader audience and seamlessly integrate it into mission-critical applications. Learn more: https://t.co/V8PTq3v5ZM
4
11
91
@kefandong
Kefan Dong
9 months
Update: check out https://t.co/U94zIRGSoj for our code, data, and model!
Tweet card summary image
github.com
The official implementation of "Self-play LLM Theorem Provers with Iterative Conjecturing and Proving" - kfdong/STP
@tengyuma
Tengyu Ma
9 months
and SoTA among whole-proof generation methods on miniF2F, ProofNet, and PutnamBench, and double the previous best results on LeanWorkBook. (reposting because it seems that this table has much more views ๐Ÿ˜)
1
9
42
@JunhongShen1
Junhong Shen
9 months
We introduce Mixture-of-Mamba, a multi-modal SSM that leverages modality-aware sparsity for efficient multi-modal pretraining! At the core of Mixture-of-Mamba: ๐Ÿ”นModality-aware sparsity to optimize efficiency ๐Ÿ”นMixture-of-SSMs to enable cross-modal interactions ๐Ÿ”นScales
@liang_weixin
Weixin Liang
9 months
๐Ÿš€ Want 2x faster pretraining for your multi-modal LLM? ๐Ÿงต Following up on Mixture-of-Transformers (MoT), we're excited to share Mixture-of-Mamba (MoM)! https://t.co/OTTpAlB4Vq ๐Ÿ”ฅ Why it matters: MoM applies modality-aware sparsity across image, text, and speechโ€”making
0
8
38
@liang_weixin
Weixin Liang
9 months
๐Ÿš€ Want 2x faster pretraining for your multi-modal LLM? ๐Ÿงต Following up on Mixture-of-Transformers (MoT), we're excited to share Mixture-of-Mamba (MoM)! https://t.co/OTTpAlB4Vq ๐Ÿ”ฅ Why it matters: MoM applies modality-aware sparsity across image, text, and speechโ€”making
0
2
19
@liang_weixin
Weixin Liang
9 months
๐Ÿ“ข Can LLMs program themselves to run faster? ๐Ÿƒโฑ๏ธ LLM self-taught to code for next-gen AI hardware! https://t.co/wiwgiPEpeH 1/ Programming AI accelerators is a major bottleneck in ML. Our self-improving LLM agent learns to write optimized code for new hardware, achieving 3.9x
2
7
38
@zhang677
Genghan Zhang
9 months
๐Ÿ” ML library development is crucial but requires expertise in ML algorithms & architecture-specific programming languages (ASPLs). ๐Ÿค– LLM agents can enable better automation. We propose an adaptive self-improvement agentic system for generating ML libraries in STePโ€”a
2
7
26
@Zhang_Yu_hui
Yuhui Zhang
10 months
๐Ÿ” Vision language models are getting better - but how do we evaluate them reliably? Introducing AutoConverter: transforming open-ended VQA into challenging multiple-choice questions! Key findings: 1๏ธโƒฃ Current open-ended VQA eval methods are flawed: rule-based metrics correlate
3
73
154
@WeijiaShi2
Weijia Shi
11 months
Introducing ๐‹๐ฅ๐š๐ฆ๐š๐…๐ฎ๐ฌ๐ข๐จ๐ง: empowering Llama ๐Ÿฆ™ with diffusion ๐ŸŽจ to understand and generate text and images in arbitrary sequences. โœจ Building upon Transfusion, our recipe fully preserves Llamaโ€™s language performance while unlocking its multimodal understanding and
15
188
897
@Zhang_Yu_hui
Yuhui Zhang
11 months
๐Ÿค” Why are VLMs (even GPT-4V) worse at image classification than CLIP, despite using CLIP as their vision encoder? Presenting VLMClassifier at #NeurIPS2024: โฐ Dec 11 (Wed), 11:00-14:00 ๐Ÿ“ East Hall #3710 Key findings: 1๏ธโƒฃ VLMs dramatically underperform CLIP (>20% gap) 2๏ธโƒฃ After
2
21
87
@SiyouPei
Siyou Pei (on job market)
11 months
Iโ€™m open to academia & industry in 2025. My work in #XR ๐Ÿฅฝ + #HCI ๐Ÿ‘ฉโ€๐Ÿ’ป enables low-friction XR experience thru #EmbodiedInteraction, unlocking potential for all -- tech-savvy or not ๐ŸŒ Design+Science+Engineering. Let's shape the future of spatial computing โœจ RT appreciated! (1/8)
3
22
89
@liang_weixin
Weixin Liang
11 months
Honored that @Nature has highlighted our work again in their latest piece examining #ChatGPT's transformative impact on scientific research and academia over the past two years. h/t @Nature https://t.co/wK4ayZYH9w
1
1
17