
Dan Hendrycks
@DanHendrycks
Followers
41K
Following
2K
Media
251
Statuses
1K
• Center for AI Safety Director • xAI and Scale AI advisor • GELU/MMLU/MATH/HLE • PhD in AI • Analyzing AI models, companies, policies, and geopolitics
San Francisco
Joined August 2009
Superintelligence is destabilizing. If China were on the cusp of building it first, Russia or the US would not sit idly by—they'd potentially threaten cyberattacks to deter its creation. @ericschmidt @alexandr_wang and I propose a new strategy for superintelligence. 🧵
77
135
739
RT @jim_mitre: In a new paper about AGI and preventive war, @RANDCorporation colleagues argue that the probability of war is low in absolut….
0
12
0
That said, they still are worse than humans when the puzzles are represented pictorially (image below). The paper: Does Spatial Cognition Emerge in Frontier Models? (ICLR 2025). Thanks to @longphan3110 for running the evaluation
2
4
86
This is a strawman. We don't use the phrase "AGI" in the MAIM paper (Superintelligence Strategy). In fact, we discuss how the concept of AGI is too vague to be useful in the appendix. We make it clear that the first thing we want to deter is an intelligence recursion---thousands.
The idea of a clear "AGI threshold" for preventive actions (MAIM paper) misses a challenge: we'll never agree when something becomes "superintelligent" or AGI. @ylecun will say, "It lacks autonomy!" while @GaryMarcus declares it's hitting a wall the next day. Some thoughts 1/.
12
5
120
We can prevent gradual disempowerment by AI. We delegate to doctors and lawyers yet can stay in charge because they must earn our informed consent. Requiring AIs to obtain informed consent and have foresight into long-term consequences helps ensure human control isn't eroded.
What to do about gradual disempowerment? We laid out a research agenda with all the concrete and feasible research projects we can think of: 🧵. with @raymondadouglas @jankulveit @DavidSKrueger.
10
6
77