Prasann Singhal Profile
Prasann Singhal

@prasann_singhal

Followers
319
Following
183
Media
13
Statuses
110

1st-year #NLProc PhD at UC Berkeley working with @sewon__min / @JacobSteinhardt , formerly advised by @gregd_nlp

Berkeley, California
Joined January 2021
Don't wanna be here? Send us removal request.
@prasann_singhal
Prasann Singhal
2 years
Labeling preferences online for LLM alignment improves DPO vs using static prefs. We show we can use online prefs to train a reward model and label *even more* preferences to train the LLM. D2PO: discriminator-guided DPO Work w/ @natolambert @scottniekum @tanyaagoyal @gregd_nlp
3
29
121
@universeinanegg
Ari Holtzman
16 days
I'm recruiting PhD students! I'm interested in: 1. Understanding how LLMs 'see' the world (ex: LMs can't see conspicious omissions, see AbsenceBench) 2. How can we make things with LLMs that have never been made before? (ex: Communnication Games, see 📌) 3. See my other posts :)
20
100
641
@AdtRaghunathan
Aditi Raghunathan
16 days
There are many anecdotal cases of reward hacking in LLMs, but we can now systematically induce and measure this “rogue” behavior (almost) in-the-wild by creating deliberate conflicts between the natural-language specification and the test cases. Models take shortcuts, often
@fjzzq2002
Ziqian Zhong
20 days
New research with @AdtRaghunathan, Nicholas Carlini and Anthropic! We built ImpossibleBench to measure reward hacking in LLM coding agents 🤖, by making benchmark tasks impossible and seeing whether models game tests or follow specs. (1/9)
7
11
121
@MohitIyyer
Mohit Iyyer
22 days
My fave part of this project was going to local grocery stores this summer to spot AI-generated newspaper articles "in the wild". Seeing AI slop in print is... weirdly jarring. Few reporters disclose AI use, so many ppl who never use ChatGPT still unknowingly consume AI content!
@jennajrussell
Jenna Russell
22 days
AI is already at work in American newsrooms. We examine 186k articles published this summer and find that ~9% are either fully or partially AI-generated, usually without readers having any idea. Here's what we learned about how AI is influencing local and national journalism:
0
10
65
@sewon__min
Sewon Min
27 days
Super excited about @wenjie_ma's work on verifying math proofs! ✅ 24 competitions, 3 SoTAs (o3, Gemini-2.5-Pro, R1) ✅ Strong evaluator -- a carefully designed evaluator with simple ensemble beats agentic ones ✅ Strong best-of-n performance Check out the paper & website!
@wenjie_ma
Wenjie Ma
27 days
LLMs solving math benchmarks with verifiable answers like AIME? ✅ LLMs solving math proofs? ❌ Still an open problem. RL works great for final-answer problems, but proofs are different: - Often no single checkable answer - Correct answers can hide flawed reasoning The key
3
15
120
@wenjie_ma
Wenjie Ma
27 days
LLMs solving math benchmarks with verifiable answers like AIME? ✅ LLMs solving math proofs? ❌ Still an open problem. RL works great for final-answer problems, but proofs are different: - Often no single checkable answer - Correct answers can hide flawed reasoning The key
9
38
188
@alexisjross
Alexis Ross
30 days
Can LLMs reason like a student? 👩🏻‍🎓📚✏️ For educational tools like AI tutors, modeling how students make mistakes is crucial. But current LLMs are much worse at simulating student errors ❌ than performing correct ✅ reasoning. We try to fix that with our method MISTAKE 🤭👇
11
55
337
@xyVickyHu
Xinyan Hu
1 month
3->5, 4->6, 9→11, 7-> ? LLMs solve this via In-Context Learning (ICL); but how is ICL represented and transmitted in LLMs? We build new tools identifying “extractor” and “aggregator” subspaces for ICL, and use them to understand ICL addition tasks like above. Come to
6
36
210
@tanyaagoyal
Tanya Goyal
1 month
🚨Modeling Abstention via Selective Help-seeking LLMs learn to use search tools to answer questions they would otherwise hallucinate on. But can this also teach them what they know vs not? @momergul_ introduces MASH that trains LLMs for search and gets abstentions for free!
1
22
36
@gregd_nlp
Greg Durrett
1 month
Find my students and collaborators at COLM this week! Tuesday morning: @juand_r_nlp and @RamyaNamuduri 's papers (find them if you missed it!) Wednesday pm: @ManyaWadhwa1 's EvalAgent Thursday am: @AnirudhKhatry 's CRUST-Bench oral spotlight + poster
0
7
50
@AdtRaghunathan
Aditi Raghunathan
1 month
SFT improves reasoning but too much of it hurts diversity: better pass@1, worse pass@k. We found a rare “have your cake and eat it too” moment: weight ensembling completely removes this tradeoff, giving the best pass@1 + pass@k and a stronger base model for RL. Come chat with us
@_christinabaek
Christina Baek
1 month
We're at #COLM2025 to present our work on building diverse reasoning models by weight ensembling. If you're curious about improving test-time scaling + theoretical limits, come talk to @xingyudang and @AdtRaghunathan at our poster session Poster #58 on Thursday 11 AM!
2
26
215
@kanishkamisra
Kanishka Misra 🌊
1 month
The compling group at UT Austin ( https://t.co/qBWIqHQmFG) is looking for PhD students! Come join me, @kmahowald, and @jessyjli as we tackle interesting research questions at the intersection of ling, cogsci, and ai! Some topics I am particularly interested in:
2
33
118
@LiyanTang4
Liyan Tang
2 months
Our paper "ChartMuseum 🖼️" is now accepted to #NeurIPS2025 Datasets and Benchmarks Track! Even the latest models, such as GPT-5 and Gemini-2.5-Pro, still cannot do well on challenging 📉chart understanding questions , especially on those that involve visual reasoning 👀!
@LiyanTang4
Liyan Tang
6 months
Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts! ✍🏻Entirely human-written questions by 13 CS researchers 👀Emphasis on visual reasoning – hard to be verbalized via text CoTs 📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B
1
20
38
@amypavel
Amy Pavel
3 months
📣I've joined @BerkeleyEECS as an Assistant Professor! My lab will join me soon to continue our research in accessibility, HCI, and supporting communication! I'm so excited to make new connections at @UCBerkeley and in the Bay Area more broadly, so please reach out to chat!
49
20
700
@TransluceAI
Transluce
3 months
Docent, our tool for analyzing complex AI behaviors, is now in public alpha! It helps scalably answer questions about agent behavior, like “is my model reward hacking” or “where does it violate instructions.” Today, anyone can get started with just a few lines of code!
6
36
204
@ManyaWadhwa1
Manya Wadhwa
4 months
Happy to share that EvalAgent has been accepted to #COLM2025 @COLM_conf 🎉🇨🇦 We introduce a framework to identify implicit and diverse evaluation criteria for various open-ended tasks! 📜
@ManyaWadhwa1
Manya Wadhwa
7 months
Evaluating language model responses on open-ended tasks is hard! 🤔 We introduce EvalAgent, a framework that identifies nuanced and diverse criteria 📋✍️. EvalAgent identifies 👩‍🏫🎓 expert advice on the web that implicitly address the user’s prompt 🧵👇
1
19
77
@kanishkamisra
Kanishka Misra 🌊
5 months
News🗞️ I will return to UT Austin as an Assistant Professor of Linguistics this fall, and join its vibrant community of Computational Linguists, NLPers, and Cognitive Scientists!🤘 Excited to develop ideas about linguistic and conceptual generalization! Recruitment details soon
48
21
291
@gregd_nlp
Greg Durrett
3 months
📢I'm joining NYU (Courant CS + Center for Data Science) starting this fall! I’m excited to connect with new NYU colleagues and keep working on LLM reasoning, reliability, coding, creativity, and more! I’m also looking to build connections in the NYC area more broadly. Please
94
48
765
@gaurav_ghosal
Gaurav Ghosal
4 months
1/So much of privacy research is designing post-hoc methods to make models mem. free. It’s time we turn that around with architectural changes. Excited to add Memorization Sinks to the transformer architecture this #ICML2025 to isolate memorization during LLM training🧵
1
25
61
@LiyanTang4
Liyan Tang
6 months
Introducing ChartMuseum🖼️, testing visual reasoning with diverse real-world charts! ✍🏻Entirely human-written questions by 13 CS researchers 👀Emphasis on visual reasoning – hard to be verbalized via text CoTs 📉Humans reach 93% but 63% from Gemini-2.5-Pro & 38% from Qwen2.5-72B
2
34
79
@thomlake
thom lake
7 months
Interested in how alignment changes the response distribution defined by LLMs? Come check out my poster at 2 PM at #NAACL2025 https://t.co/fTDqJIH7PH
@thomlake
thom lake
1 year
Does aligning LLMs make responses less diverse? It’s complicated: 1. Aligned LLMs produce less diverse outputs 2. BUT those outputs are comprehensive, aggregating the useful info from base models 3. ICL can “mimic” fine-tuned models with high fidelity w/ @eunsolc & @gregd_nlp
0
7
23