
Nan Jiang
@nanjiangwill
Followers
72
Following
111
Media
5
Statuses
31
research at UChicagoCS / at Amazon AGI SF Lab
San Francisco, CA
Joined January 2018
RT @SonglinYang4: 📢 (1/16) Introducing PaTH 🛣️ — a RoPE-free contextualized position encoding scheme, built for stronger state tracking, be….
arxiv.org
The attention mechanism is a core primitive in modern large language models (LLMs) and AI more broadly. Since attention by itself is permutation-invariant, position encoding is essential for...
0
91
0
RT @lmsysorg: Awesome collaboration between our SGLang team @lmsysorg, @verl_project, LinkedIn, and UCLA AGI Lab. Thanks so much for the….
0
5
0
RT @wzhao_nlp: Some personal news: I'll join @UMassAmherst CS as an assistant professor in fall 2026. Until then, I'll postdoc at @Meta nyc….
0
32
0
amazing Jason, amazing Nexad, please check this out!.
Let’s be real—ads have annoyed me for years. Pop-ups, spam, etc… while the world is moving towards AGI, the ad world felt stuck in the past. So I decided to flip the script. Today, I’m proud to share: Nexad has raised a $6M seed round, led by @a16z SR04, @Prosus_Ventures ,
1
0
1
RT @wzhao_nlp: Coding agents can debug their own outputs, but what if none of the fixes are correct? We overcome sparse rewards by making t….
0
25
0
RT @srush_nlp: I teach a class where students code up an ML library from scratch in Python. Wenting showed me that a Claude Agent (with int….
0
37
0
So. can agents now build a package from scratch? Test them on Commit0!. This is an amazing and fun project this summer! Huge thanks to Wenting and to everyone in the lab for their support and guidance! 🚀👏.
Introducing the commit0 interactive environment for coding agents. Challenge: generate Python libraries from scratch. Commit0 is designed with interactivity, dependencies, and specifications as first-class considerations. We include a benchmark with 50+ challenging libraries.
0
2
9
RT @xiuyu_l: Handling long context in LLMs is expensive, but can we cut the cost by learning them offline for a specific set/genre of docum….
0
52
0
RT @onjas_buidl: 🚀 Introducing RouterBench, the first comprehensive benchmark for evaluating LLM routers! 🎉.A collaboration between @withma….
0
30
0
RT @wkvong: 1/ Today in Science, we train a neural net from scratch through the eyes and ears of one child. The model learns to map words t….
0
689
0
We're excited to contribute to the exploration of alternative architectures and emergent capabilities!!. 🎉🎉🎉 Huge congrats and many thanks to Ivan Lee and Prof. Taylor Berg-Kirkpatrick @BergKirkpatrick. 🧵[9/9].
0
0
1
❓Are attention-based models needed for In-Context Learning(ICL)?.🤔Can emerging architectures perform ICL?. 🎉Check out our #ICLR2024 paper "Exploring the Relationship Between Model Architecture and In-Context Learning Ability" 🎉 #LLM. Paper: 🧵[1/9].
arxiv.org
What is the relationship between model architecture and the ability to perform in-context learning? In this empirical study, we take the first steps toward answering this question. We evaluate...
1
29
123
RT @marcusjmin: 🚨 #GPT4 doesn't understand the code/specification written by itself!? 🚨. 🥳 Check out our #ICLR2024 paper "Beyond Accuracy:….
0
5
0