EpochAIResearch Profile Banner
Epoch AI Profile
Epoch AI

@EpochAIResearch

Followers
31K
Following
482
Media
583
Statuses
2K

Investigating the trajectory of AI for the benefit of society.

Joined May 2022
Don't wanna be here? Send us removal request.
@EpochAIResearch
Epoch AI
1 month
Announcing our Frontier Data Centers Hub! The world is about to see multiple 1 GW+ AI data centers. We mapped their construction using satellite imagery, permits & public sources — releasing everything for free, including commissioned satellite images. Highlights in thread!
88
99
624
@EpochAIResearch
Epoch AI
6 hours
Explore the interactive insight on our website: https://t.co/4lFnZTyAEZ Or dig into our Frontier Data Centers hub yourself:
Tweet card summary image
epoch.ai
Open database of AI data centers using satellite and permit data to show compute, power use, and construction timelines.
0
0
7
@EpochAIResearch
Epoch AI
6 hours
Our compute estimates for these data centers are based on satellite imagery of the cooling infrastructure, permitting documents, and company disclosures.
1
0
8
@EpochAIResearch
Epoch AI
6 hours
Using data from our Frontier Data Centers and AI Models hubs, we calculated the number of GPT-4-scale training runs achievable on the largest data center in a single month. To keep things comparable with GPT-4, we assume training is done at 16-bit precision.
1
0
12
@EpochAIResearch
Epoch AI
6 hours
Microsoft's Fairwater Atlanta (today's largest data center) could likely train over 20 models the size of GPT-4 in the course of a month. This computational power will enable AI companies to increase the number and scale of both experiments and training runs.
6
16
88
@EpochAIResearch
Epoch AI
7 hours
Using data from our Frontier Data Centers and AI Models hubs, we calculated the number of GPT-4-scale training runs achievable in a single month. To keep things comparable with GPT-4, we assume training is done at 16-bit precision.
0
0
2
@EpochAIResearch
Epoch AI
8 hours
Learn more about Alibaba Zhangbei and other data centers using our free Satellite Explorer!
Tweet card summary image
epoch.ai
See how satellite imagery, permits, and public disclosures are used to track the power capacity and performance of frontier data centers.
0
0
8
@EpochAIResearch
Epoch AI
8 hours
In total, we estimate between 200–500 MW of total facility power is operational for Alibaba in Zhangbei today. This power capacity is close to leading US data centers, but the compute capacity is significantly lower, due to the older buildings with slower chips.
1
1
17
@EpochAIResearch
Epoch AI
8 hours
We also noticed a mid-flight change to the design of 6 buildings from 2024 to 2025. The updated design puts many more chillers on the roof, roughly doubling the power density. This is an unusual development, and suggests that more advanced chips are being installed.
1
0
13
@EpochAIResearch
Epoch AI
8 hours
How do we know the power capacity of these buildings? Cooling equipment gives it away. For example, each building shown below has 40 air-cooled chillers with 12 fans. Plugging this into our model, we estimate 38 MW of cooling capacity, which is closely related to IT capacity.
2
2
14
@EpochAIResearch
Epoch AI
8 hours
We first found out about an Alibaba facility located in Zhangbei from a web search about Chinese data centers. We then purchased high-resolution satellite images of the region and identified 20 data center buildings, many more than we expected.
1
0
14
@EpochAIResearch
Epoch AI
8 hours
We’ve identified an Alibaba data center in Zhangbei, China with an estimated 200–500 MW capacity. Roughly half of that predates the AI boom and likely hosts little modern compute. But several newer buildings show a high power density consistent with advanced AI chips. 🧵
2
8
81
@EpochAIResearch
Epoch AI
2 days
This project was funded by Google DeepMind. It was done in collaboration with researchers from @EpochAIResearch and @GoogleDeepMind@ansonwhho, @js_denain, @DJAtanasov, @SamuelAlbanie, and @rohinmshah.
1
0
35
@EpochAIResearch
Epoch AI
2 days
And it also allows us to study accelerations in AI capabilities. We ran simulations with synthetic benchmark data, where capabilities accelerated two-fold in 2027. Using our framework, we’re able to detect this speed-up within 2-3 months.
1
2
39
@EpochAIResearch
Epoch AI
2 days
It lets us study improvements in AI software, where better algorithms and data need less training compute to reach the same estimated capability.
2
0
27
@EpochAIResearch
Epoch AI
2 days
This opens several promising directions of research. For example, it lets us project future improvements in AI capabilities:
1
5
32
@EpochAIResearch
Epoch AI
2 days
Using real-world data from our benchmark hub, we estimate these parameters to get capability scores for different models. So we can compare models even if they’re not evaluated on the same benchmarks! And we also see a clear trend in capabilities over time:
1
0
24
@EpochAIResearch
Epoch AI
2 days
We then relate these parameters to benchmark scores with an S-curve, with 3 regimes: 1. Capability << Difficulty: performance is close to random 2. Capability ≈ Difficulty: capability grows linearly with performance 3. Capability >> Difficulty: the benchmark is saturated
1
0
21
@EpochAIResearch
Epoch AI
2 days
The core idea is similar to Item Response Theory: assume each model has a latent capability and each benchmark has a latent difficulty This is like how chess players and puzzles both have Elo scores Finally, each benchmark has a latent slope that tells us how fast it saturates.
1
0
26
@EpochAIResearch
Epoch AI
2 days
Most AI benchmarks share a common flaw: they saturate too quickly to study long-run trends. Our solution: “stitch” many benchmarks together. This lets us compare models across a wide range of capabilities on a single unified scale. Here’s how this works.🧵
9
28
201