From chip to hall, at four scales. The chip does the work. Everything else keeps it cool, fed, and connected.
A single accelerator. Tens of thousands of cores doing matrix maths in parallel.
Eight GPUs in one chassis at terabit speed. The smallest unit a frontier model can train on.
Eight servers in one cabinet, drawing the power of a small suburb. One tile of a much larger cluster.
Many rows of racks. Megawatts of grid feed, megalitres of cooling water, kilometres of fibre, all for the chips.
Tens of thousands of GPUs in one fabric, working as one machine. Network speed inside the building decides whether it converges.
The trained model copied across many small clusters, each serving live queries. Milliseconds, then the chip is free. Distributable across sites.
An AI data centre is a life-support system for chips. The compute layer ages on a three-to-four year life; the shell, the substation, and the grid right do not. Durable asset: the building and the connection. Wasting asset: what sits on the floor.