tli104 Profile Banner
Tianjian Li Profile
Tianjian Li

@tli104

Followers
277
Following
1K
Media
7
Statuses
175

PhD student @jhuclsp, research scientist intern @AIatMeta FAIR. I work on data curation for language models. Previously @nyuniversity.

Baltimore, MD
Joined November 2022
Don't wanna be here? Send us removal request.
@tli104
Tianjian Li
8 months
I have written a blogpost offering an explanation of why both the chosen and the rejected log-probability decreases during DPO, and more interestingly, why it is a desired phenomenon to some extent. Link:
0
5
12
@tli104
Tianjian Li
5 days
RT @yanawei_: ๐Ÿš€ Check out our โœˆ๏ธ ICML 2025 work: Perception in Reflection!. A reasonable perception paradigm for LVLMs should be iterativeโ€ฆ.
0
3
0
@tli104
Tianjian Li
8 days
RT @_zhihuixie: ๐Ÿš€ Thrilled to announce Dream-Coder 7B โ€” the most powerful open diffusion codeโ€ฏ LLM to date.
Tweet media one
0
33
0
@tli104
Tianjian Li
9 days
RT @PranjalAggarw16: Super excited to see L1 accepted to #COLM2025!. We are further open-sourcing 5 new models & a dataset:.1. L1-7B & L1-โ€ฆ.
0
3
0
@tli104
Tianjian Li
16 days
RT @DanielKhashabi: Whatโ€™s really going on inside LLMs when they handle non-English queries?. @BafnaNiyati's recent work introduces the **tโ€ฆ.
0
4
0
@tli104
Tianjian Li
17 days
RT @Happylemon56775: Excited to share what I worked on during my time at Meta. - We introduce a Triton-accelerated Transformer with *2-simโ€ฆ.
0
96
0
@tli104
Tianjian Li
19 days
RT @BafnaNiyati: ๐Ÿ“ขWhen LLMs solve tasks with a mid-to-low resource input/target language, their output quality is poor. We know that. But cโ€ฆ.
0
9
0
@tli104
Tianjian Li
25 days
RT @gui_penedo: We have finally released the ๐Ÿ“paper for ๐Ÿฅ‚FineWeb2, our large multilingual pre-training dataset. Along with general (and exโ€ฆ.
0
96
0
@tli104
Tianjian Li
26 days
RT @arnal_charles: โ“How to balance negative and positive rewards in off-policy RLโ“. In Asymmetric REINFORCE for off-Policy RL, we show thatโ€ฆ.
0
28
0
@tli104
Tianjian Li
28 days
RT @nouhadziri: ๐Ÿ“ข Can LLMs really reason outside the box in math? Or are they just remixing familiar strategies? . Remember DeepSeek R1, o1โ€ฆ.
0
159
0
@tli104
Tianjian Li
28 days
RT @YLiiiYLiii: ๐Ÿง ๐Ÿค–Can LLMs outperform online therapists in single-turn counseling? Meet CounselBench - 100 mental-health professionals rateโ€ฆ.
0
6
0
@tli104
Tianjian Li
1 month
RT @_akhaliq: POLARIS. A POst-training recipe for scaling reinforcement Learning on Advanced ReasonIng modelS. 4B model achieves 81.2% Passโ€ฆ.
0
63
0
@tli104
Tianjian Li
1 month
RT @natolambert: A common trend across recent research in using reinforcement learning to train reasoning models is that the clipping operaโ€ฆ.
0
54
0
@tli104
Tianjian Li
1 month
RT @aviral_kumar2: Our view on test-time scaling has been to train models to discover algos that enable them to solve harder problems. @seโ€ฆ.
0
30
0
@tli104
Tianjian Li
1 month
RT @alexdmartin314: Talk to me at #CVPR2025 about Multimodal RAG topics! I'll be presenting two papers on video retrieval: Video-ColBERT (lโ€ฆ.
0
7
0
@tli104
Tianjian Li
2 months
RT @fly51fly: [LG] Rewarding the Unlikely: Lifting GRPO Beyond Distribution Sharpening.A He, D Fried, S Welleck [CMU] (2025)..
0
10
0
@tli104
Tianjian Li
2 months
RT @wellecks: New paper by Andre He:. Rewarding the Unlikely: Lifting GRPO Beyond Distribution Sharpening. Tired oโ€ฆ.
0
53
0
@tli104
Tianjian Li
2 months
RT @jaehunjung_com: Data curation is crucial for LLM reasoning, but how do we know if our dataset is not overfit to one benchmark and generโ€ฆ.
0
33
0
@tli104
Tianjian Li
2 months
RT @_vaishnavh: ๐Ÿ“ข New paper on creativity & multi-token prediction! We design minimal open-ended tasks to argue:. โ†’ LLMs are limited in creโ€ฆ.
0
40
0
@tli104
Tianjian Li
2 months
RT @May_F1_: ๐Ÿš€ Data to pre-train LLMs on are reaching critical bottleneck. ๐˜ฟ๐™ค๐™š๐™จ ๐™ข๐™ค๐™™๐™š๐™ก-๐™œ๐™š๐™ฃ๐™š๐™ง๐™–๐™ฉ๐™š๐™™ ๐™จ๐™ฎ๐™ฃ๐™ฉ๐™๐™š๐™ฉ๐™ž๐™˜ ๐™™๐™–๐™ฉ๐™– ๐™ฌ๐™ค๐™ง๐™  ๐™จ๐™ž๐™ข๐™ž๐™ก๐™–๐™ง๐™ก๐™ฎ ๐™ฌ๐™š๐™ก๐™ก ๐™›๐™ค๐™ง ๐™จ๐™˜๐™–๐™กโ€ฆ.
0
27
0
@tli104
Tianjian Li
2 months
RT @Mengyue_Yang_: Curious how training data order impacts LLMs without retraining?. Introducing FUT:. ๐Ÿ”Estimate the effects of any sampleโ€ฆ.
0
44
0