Mark Chen
@markchen90
Followers
66K
Following
2K
Media
23
Statuses
394
Chief Research Officer at @OpenAI. Coach for the USA IOI Team.
Joined June 2020
We wrapped up this year's competition circuit with a full score on the ICPC, after achieving 6th in the IOI, a gold medal at the IMO, and 2nd in the AtCoder Heuristic contest!
1/n I’m really excited to share that our @OpenAI reasoning system got a perfect score of 12/12 during the 2025 ICPC World Finals, the premier collegiate programming competition where top university teams from around the world solve complex algorithmic problems. This would have
29
43
792
Instead of trying to prove it, I asked GPT5 about it, and in about 20 seconds received a proof. The proof relied on a lemma that I had not heard of (the statement was a bit outside my main areas), so although I am confident I'd have got there in the end, 2/3
8
34
787
Codex had its strongest growth in one day yesterday since the launch of gpt-5-codex. Way to motivate the team during a gnarly investigation that's making us go through every piece of infra, hardware and line of code in our system.
108
34
1K
We often talk about big leaps in AI for mathematics, but I think the small steps are equally impressive. The future of mathematics is now. I was working on a particular task: finding a case-free proof of the representability of the local Néron function correction using a
22
76
549
@SebastienBubeck @kevinweil ChatGPT is now at the level of solving some math research questions, but you do need an expert guiding it. This exercise was a lot of fun and was highly productive. I also feel I'm getting better at prompting ChatGPT. I'll also try other open and unsolved problems. (16/N, N=16)
15
23
313
Excited to start OpenAI for Physics w/ @ALupsasca @kevinweil @aleks_madry and @merettm! I sat with @ALupsasca when GPT-5 reproduced his latest research paper, and we both felt parallels to watching AlphaGo play move 37. It's nearly impossible to be a world class chess player
After GPT-5 Pro launched, I gave it that same problem. To my utter shock, it rediscovered the result in <30min! See for yourself: https://t.co/IpLuaGlJ03 It’s not flawless (it needs priming on the flat-space case before tackling the full problem) but the leap is incredible.
30
95
1K
The Training team @OpenAI is hiring researchers in London 🚀 Our twin missions are to train better LLMs, and serve them more cheaply Get in touch if you are excited to collaborate on architecture design, reliable scaling, and faster optimization
12
38
474
I guess it's now every day until the end of time
GPT-5 Pro found a counterexample to the NICD-with-erasures majority optimality (Simons list, p.25). https://t.co/T3m9MYgqe0 At p=0.4, n=5, f(x) = sign(x_1-3x_2+x_3-x_4+3x_5) gives E|f(x)|=0.43024 vs best majority 0.42904.
19
64
1K
AI’s next leap isn’t “bigger models”, it’s more learnable ones What if models could learn faster, with less data, even while solving a task? Join @lukaszkaiser (co‑creator of the Transformer) for a map of how AI is evolving from RNNs to Reasoners to Researchers Link in thread
2
1
28
We aim to build the most intelligent and useful AI. But “useful” is a fuzzy word. GDPval (consisting of tasks spanning 44 occupations across the top 9 sectors contributing to US GDP) makes "usefulness" more concrete. Would love to saturate this one!
Understanding the capabilities of AI models is important to me. To forecast how AI models might affect labor, we need methods to measure their real-world work abilities. That’s why we created GDPval.
12
15
321
“The default way to code is vibecoding.” OpenAI chief research officer Mark Chen (@markchen90) says high schoolers already see coding from scratch as weird. If coding is vibes, what does great research look like?
We wrapped up this year's competition circuit with a full score on the ICPC, after achieving 6th in the IOI, a gold medal at the IMO, and 2nd in the AtCoder Heuristic contest!
31
45
436
Introducing Among AIs, a social reasoning benchmark where embodied models play Among Us to test social intelligence: deception, persuasion, and coordination. We put 6 SOTA models in a live arena and GPT-5 came out on top by leading in Impostor & Crewmate wins. Why did GPT-5 get
63
136
1K
GPT-5 (the one we can actually use) solved 11/12 ICPC problems. Only the final hardest one needed OpenAI’s internal reasoning model. So the public already has access to a math/coding genius that just outperformed Google’s internal Gemini Deep Think. 🤯
11 out of 12 problems were correctly solved by GPT-5 solutions on the first submission attempt to the ICPC-managed and sanctioned online judging environment The final and most challenging problem was solved by our experimental reasoning model after GPT-5 encountered
9
20
314
.@OpenAI joined the experiment to incorporate AI development tools into the 49th Annual ICPC World Finals Local Judge. OpenAI's models solved all 12 problems – a milestone akin to achieving a gold medal at the highest level of achievement! https://t.co/ntUHfRAQzO
worldfinals.icpc.global
The 49th Annual World Championship of the ICPC International Collegiate Programming Contest
8
35
244
Any past contestant will tell you how much hard work, dedication, and creativity is needed to achieve these results. We're excited to bring these capabilities to bear on solving real world problems and to advance the scientific frontier.
1
3
147
Alignment is arguably the most important AI research frontier. As we scale reasoning, models gain situational awareness and a desire for self-preservation. Here, a model identifies it shouldn’t be deployed, considers covering it up, but then realizes it might be in a test.
Today we’re releasing research with @apolloaievals. In controlled tests, we found behaviors consistent with scheming in frontier models—and tested a way to reduce it. While we believe these behaviors aren’t causing serious harm today, this is a future risk we’re preparing
58
69
591
Give GPT-5-Codex a try! Huge props to Andrey Mishchenko, @katyhshi, @hansonwng, @ssnl_tz, and @mia_glaese for turning our reasoning models’ raw intelligence into real-world coding performance - and uncovering new research along the way.
We’re releasing GPT-5-Codex — a version of GPT-5 further optimized for agentic coding in Codex. Available in the Codex CLI, IDE Extension, web, mobile, and for code reviews in Github.
11
25
501
GPT-5-Codex is here: a version of GPT-5 better at agentic coding. It is faster, smarter, and has new capabilities. Let us know what you think! The team has been absolutely cooking, very fun to watch.
863
821
12K