peterbhase Profile Banner
Peter Hase Profile
Peter Hase

@peterbhase

Followers
3K
Following
2K
Media
57
Statuses
470

Visiting Scientist at Schmidt Sciences. Visiting Researcher at the Stanford NLP Group Previously: Anthropic, AI2, Google, Meta, UNC Chapel Hill

New York, NY
Joined April 2019
Don't wanna be here? Send us removal request.
@peterbhase
Peter Hase
1 year
My last PhD paper šŸŽ‰: fundamental problems with model editing for LLMs!. We present *12 open challenges* with definitions/benchmarks/assumptions, inspired by work on belief revision in philosophy. To provide a way forward, we test model editing against Bayesian belief revision.🧵
Tweet media one
3
74
307
@peterbhase
Peter Hase
1 month
Shower thought: LLMs still have very incoherent notions of evidence, and they update in strange ways when presented with information in-context that is relevant to their beliefs. I really wonder what will happen when LLM agents start doing interp on themselves and see the source.
5
5
23
@grok
Grok
4 days
Join millions who have switched to Grok.
175
335
3K
@peterbhase
Peter Hase
1 month
RT @hannahrosekirk: My team at @AISecurityInst is hiring! This is an awesome opportunity to get involved with cutting-edge scientific resea….
0
24
0
@peterbhase
Peter Hase
2 months
RT @nouhadziri: Current agents are highly unsafe, o3-mini one of the most advanced models in reasoning score 71% in executing harmful reque….
0
15
0
@peterbhase
Peter Hase
2 months
RT @milesaturpin: New @Scale_AI paper! 🌟. LLMs trained with RL can exploit reward hacks but not mention this in their CoT. We introduce ver….
0
77
0
@peterbhase
Peter Hase
2 months
Overdue job update -- I am now:.- A Visiting Scientist at @schmidtsciences, supporting AI safety and interpretability.- A Visiting Researcher at the Stanford NLP Group, working with @ChrisGPotts. I am so grateful I get to keep working in this fascinating and essential area, and.
15
22
174
@peterbhase
Peter Hase
2 months
RT @FazlBarez: Excited to share our paper: "Chain-of-Thought Is Not Explainability"! . We unpack a critical misconception in AI: models exp….
0
136
0
@peterbhase
Peter Hase
2 months
RT @JustenMichel: really interesting to see just how gendered excitement about AI is, even among AI experts
Tweet media one
0
47
0
@peterbhase
Peter Hase
3 months
RT @farairesearch: šŸ¤” Can lie detectors make AI more honest? Or will they become sneakier liars?. We tested what happens when you add decept….
0
10
0
@peterbhase
Peter Hase
3 months
RT @jiaxinwen22: New Anthropic research: We elicit capabilities from pretrained models using no external supervision, often competitive or….
0
157
0
@peterbhase
Peter Hase
3 months
RT @dongkeun_yoon: šŸ™ LLMs are overconfident even when they are dead wrong. 🧐 What about reasoning models? Can they actually tell us ā€œMy an….
0
49
0
@peterbhase
Peter Hase
3 months
colab: For aficionados, the post also contains some musings on ā€œtuning the random seedā€ and how to communicate uncertainty associated with this process.
colab.research.google.com
Colab notebook
0
0
0
@peterbhase
Peter Hase
3 months
Are p-values missing in AI research?. Bootstrapping makes model comparisons easy!. Here's a new blog/colab with code for:.- Bootstrapped p-values and confidence intervals.- Combining variance from BOTH sample size and random seed (eg prompts).- Handling grouped test data. Link ā¬‡ļø.
1
3
9
@peterbhase
Peter Hase
4 months
RT @ysu_nlp: New AI/LLM Agents Track at #EMNLP2025! . In the past few years, it feels a bit odd to submit agent work to *CL venues because….
0
24
0
@peterbhase
Peter Hase
4 months
RT @vaidehi_patil_: 🚨 Introducing our @TmlrOrg paper ā€œUnlearning Sensitive Information in Multimodal LLMs: Benchmark and Attack-Defense Eva….
0
37
0
@peterbhase
Peter Hase
4 months
RT @EliasEskin: Extremely excited to announce that I will be joining @UTAustin @UTCompSci in August 2025 as an Assistant Professor! šŸŽ‰. I’m….
0
65
0
@peterbhase
Peter Hase
4 months
RT @rowankwang: New Anthropic Alignment Science blog post: Modifying LLM Beliefs with Synthetic Document Finetuning. We study a technique f….
0
46
0
@peterbhase
Peter Hase
4 months
RT @amuuueller: Lots of progress in mech interp (MI) lately! But how can we measure when new mech interp methods yield real improvements ov….
0
38
0
@peterbhase
Peter Hase
5 months
RT @sydneymlevine: šŸ”†Announcement time!šŸ”†In Spring 2026, I will be joining the NYU Psych department as an Assistant Professor! My lab will s….
0
13
0
@peterbhase
Peter Hase
5 months
RT @maksym_andr: Excited to present our recent work on AI safety at this event!. If you're coming to ICLR 2025 in S….
0
9
0
@peterbhase
Peter Hase
5 months
RT @yanda_chen_: My first paper @AnthropicAI is out!. We show that Chains-of-Thought often don’t reflect models’ true reasoning—posing chal….
0
87
0