
Tianjian Li
@tli104
Followers
277
Following
1K
Media
7
Statuses
175
PhD student @jhuclsp, research scientist intern @AIatMeta FAIR. I work on data curation for language models. Previously @nyuniversity.
Baltimore, MD
Joined November 2022
I have written a blogpost offering an explanation of why both the chosen and the rejected log-probability decreases during DPO, and more interestingly, why it is a desired phenomenon to some extent. Link:
0
5
12
RT @_zhihuixie: ๐ Thrilled to announce Dream-Coder 7B โ the most powerful open diffusion codeโฏ LLM to date.
0
33
0
RT @PranjalAggarw16: Super excited to see L1 accepted to #COLM2025!. We are further open-sourcing 5 new models & a dataset:.1. L1-7B & L1-โฆ.
0
3
0
RT @DanielKhashabi: Whatโs really going on inside LLMs when they handle non-English queries?. @BafnaNiyati's recent work introduces the **tโฆ.
0
4
0
RT @Happylemon56775: Excited to share what I worked on during my time at Meta. - We introduce a Triton-accelerated Transformer with *2-simโฆ.
0
96
0
RT @BafnaNiyati: ๐ขWhen LLMs solve tasks with a mid-to-low resource input/target language, their output quality is poor. We know that. But cโฆ.
0
9
0
RT @gui_penedo: We have finally released the ๐paper for ๐ฅFineWeb2, our large multilingual pre-training dataset. Along with general (and exโฆ.
0
96
0
RT @arnal_charles: โHow to balance negative and positive rewards in off-policy RLโ. In Asymmetric REINFORCE for off-Policy RL, we show thatโฆ.
0
28
0
RT @nouhadziri: ๐ข Can LLMs really reason outside the box in math? Or are they just remixing familiar strategies? . Remember DeepSeek R1, o1โฆ.
0
159
0
RT @YLiiiYLiii: ๐ง ๐คCan LLMs outperform online therapists in single-turn counseling? Meet CounselBench - 100 mental-health professionals rateโฆ.
0
6
0
RT @natolambert: A common trend across recent research in using reinforcement learning to train reasoning models is that the clipping operaโฆ.
0
54
0
RT @aviral_kumar2: Our view on test-time scaling has been to train models to discover algos that enable them to solve harder problems. @seโฆ.
0
30
0
RT @alexdmartin314: Talk to me at #CVPR2025 about Multimodal RAG topics! I'll be presenting two papers on video retrieval: Video-ColBERT (lโฆ.
0
7
0
RT @jaehunjung_com: Data curation is crucial for LLM reasoning, but how do we know if our dataset is not overfit to one benchmark and generโฆ.
0
33
0
RT @_vaishnavh: ๐ข New paper on creativity & multi-token prediction! We design minimal open-ended tasks to argue:. โ LLMs are limited in creโฆ.
0
40
0
RT @May_F1_: ๐ Data to pre-train LLMs on are reaching critical bottleneck. ๐ฟ๐ค๐๐จ ๐ข๐ค๐๐๐ก-๐๐๐ฃ๐๐ง๐๐ฉ๐๐ ๐จ๐ฎ๐ฃ๐ฉ๐๐๐ฉ๐๐ ๐๐๐ฉ๐ ๐ฌ๐ค๐ง๐ ๐จ๐๐ข๐๐ก๐๐ง๐ก๐ฎ ๐ฌ๐๐ก๐ก ๐๐ค๐ง ๐จ๐๐๐กโฆ.
0
27
0
RT @Mengyue_Yang_: Curious how training data order impacts LLMs without retraining?. Introducing FUT:. ๐Estimate the effects of any sampleโฆ.
0
44
0