
Prasann Singhal
@prasann_singhal
Followers
285
Following
151
Media
13
Statuses
92
4th-year undergrad #NLProc Researcher at UT Austin, advised by @gregd_nlp
Austin Texas
Joined January 2021
Labeling preferences online for LLM alignment improves DPO vs using static prefs. We show we can use online prefs to train a reward model and label *even more* preferences to train the LLM. D2PO: discriminator-guided DPO. Work w/ @natolambert @scottniekum @tanyaagoyal @gregd_nlp
3
28
121
RT @LiyanTang4: Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts!. ✍🏻Entirely human-written questions by….
0
28
0
RT @thomlake: Interested in how alignment changes the response distribution defined by LLMs? Come check out my poster at 2 PM at #NAACL2025….
0
6
0
RT @gregd_nlp: Check out Ramya et al.'s work on understanding discourse similarities in LLM-generated text! We see this as an important ste….
0
2
0
RT @ManyaWadhwa1: Evaluating language model responses on open-ended tasks is hard! 🤔. We introduce EvalAgent, a framework that identifies n….
0
35
0
RT @gregd_nlp: Check out Manya's work on evaluation for open-ended tasks! The criteria from EvalAgent can be plugged into LLM-as-a-judge or….
0
3
0
RT @SriramPad05: Are LMs sensitive to suspicious coincidences? Our paper finds that, when given access to knowledge of the hypothesis space….
0
6
0
RT @AnirudhKhatry: 🚀Introducing CRUST-Bench, a dataset for C-to-Rust transpilation for full codebases 🛠️.A dataset of 100 real-world C repo….
0
18
0
RT @gregd_nlp: New work led by @LiyanTang4 with a strong new model for chart understanding! Check out the blog post, model, and playground!….
0
8
0
RT @gregd_nlp: Check out Anirudh's work on a new benchmark for C-to-Rust transpilation! 100 realistic-scale C projects, plus target Rust in….
0
1
0
RT @madiator: Happy to announce Bespoke-Minichart-7B!. This was a tough cookie to crack, and involved a lot of data curation and modeling w….
0
7
0
RT @tanishqkumar07: trained a nanoGPT? feeling behind before o4-mini?. 🚨🚨i'm open-sourcing beyond-nanoGPT, an internal codebase to help peo….
0
47
0
RT @jacspringer: Training with more data = better LLMs, right? 🚨. False! Scaling language models by adding more pre-training data can decre….
0
174
0
RT @jessyjli: 🌟Job ad🌟 We (@gregd_nlp, @mattlease and I) are hiring a postdoc fellow within the CosmicAI Institute, to do galactic work wit….
0
23
0
RT @ZayneSprague: Interesting perspective, thanks for sharing! As one of the authors of the “CoT mainly helps on math/logic paper”, I agree….
0
8
0
RT @ManyaWadhwa1: I'll be presenting this work at #EMNLP2024 🌴on Tuesday, 4-5:30pm, Poster Session C in Jasmine Hall ! . Stop by or reach o….
0
11
0
RT @RyoKamoi: We will present our survey on self-correction of LLMs (TACL) at #EMNLP2024 in person!. Oral: Nov 12 (Tue) 11:00- (Language Mo….
0
11
0
RT @gregd_nlp: I won't be at #EMNLP2024, but my students & collaborators are presenting:.🔍 Detecting factual errors from LLMs @LiyanTang4….
0
16
0
RT @LiyanTang4: Great to see an independent measurement of MiniCheck in hallucination detection!. Repo: Model: http….
0
5
0
RT @gregd_nlp: At COLM w/.@ManyaWadhwa1: work on 🧑⚖️ using NL explanations in evaluation.@prasann_singhal: work on (1) 📈 length correlatio….
0
11
0