
Hitesh Golchha
@hitesh_golchha
Followers
80
Following
568
Media
6
Statuses
300
Applied Scientist @Amazon, ML Research @UMassAmherst,
Massachusetts
Joined November 2017
🥳 Excited to announce that our work “Language Guided Exploration for RL Agents in Text Environments” was accepted into #NAACL2024 Findings! 🚀 .Paper: 🧵(1/n).
4
5
23
RT @HannaHajishirzi: Excited to release our newest, largest, and best Tulu yet. Our RLVR recipe works at scale, outperforming Deepseek V3.….
0
5
0
RT @natolambert: The DeepSeek R1 recipe, what questions we need to answer to train an o1 replication ourselves at home, and what it means f….
0
86
0
RT @SonglinYang4: I've created slides for those curious about the recent rapid progress in linear attention: from linear attention to Light….
0
164
0
RT @cgarciae88: Google Cloud just recently released "The PyTorch developer's guide to JAX fundamentals". Contains a side-by-side implement….
0
171
0
RT @nrehiew_: How to train a 670B parameter model. Let's talk about the DeepSeek v3 report + some comparisons with what Meta did with Lla….
0
527
0
RT @cocoweixu: We wrapped up CS 8803 "Large Language Model" class at @GeorgiaTech for Fall 2024. Here is the reading list:. • learning fr….
0
173
0
RT @natolambert: I've spent the last two years scouring all available resources on RLHF specifically and post training broadly. Today, with….
0
142
0
RT @simon_jegou: 🚀 Excited to announce KVPress — our open-source library for efficient #LLM KV cache compression!.👉 Check it out (and drop….
0
8
0
RT @currying: Around ten years ago, I started studying inverse problems in Topological Data Analysis (TDA). For decades people in computati….
0
130
0
RT @jxmnop: Top-rated papers from ICLR 2025. Scaling In-the-Wild Training for Diffusion-based Illumination Harmonization and Editing by Imp….
0
88
0
RT @abeirami: RLHF provably can't teach models any new knowledge. If you need to teach new skills, you need to look at pre-training and SFT….
0
21
0
RT @jaseweston: 🚨 Self-Consistency Preference Optimization (ScPO)🚨.- New self-training method without human labels - learn to make the mode….
0
107
0
RT @kayembruno: Diffusion models are so ubiquitous, but it's difficult to find an introduction that is concise, simple and comprehensive.….
0
116
0
RT @wellingmax: “We are on the brink of an irreversible climate disaster. This is a global emergency beyond any doubt. Much of the very fab….
0
31
0
RT @srush_nlp: If you know Torch, I think you can code for GPU now with OpenAI's Triton language. We made some puzzles to help you rewire….
0
197
0
The work was done during my Masters degree with super amazing co-authors - @sahil_yerawar, @_dhruveshp from IESL Lab @UMass_NLP at UMass Amherst and Soham Dan, @keerthi166 from @IBMResearch! Congratulations to all! 🎉 🧵(6/n).
0
0
1