The data centre was not designed for AI

Share this article

Artificial intelligence is being scaled inside buildings conceived for a different era of computing. What looks like a question of performance is turning into something more fundamental, a confrontation between the physical logic of the old data centre and the thermal realities of the new machine.

The data centre industry has spent years talking as if artificial intelligence were simply the next workload to arrive, larger, faster, more demanding perhaps, but still something that could be absorbed into the familiar logic of capacity planning, rack design and mechanical resilience. That assumption is starting to look thin. AI does not simply consume more infrastructure. It behaves in ways that expose the underlying compromises of the infrastructure it enters, and nowhere is that becoming clearer than in the way heat is now beginning to dictate what can be built, run and scaled.

There is a tendency in the market to treat this as a technical challenge that can be solved by adding more of everything. More power into the building, more fans through the hall, more engineering layered onto an architecture that was designed for a different thermal order. Yet that instinct risks missing the deeper point. The modern data centre was built on an arrangement of systems that assumed relative separation between compute, power and cooling, with each treated as a discipline that could be optimised largely within its own boundaries. AI is beginning to erode that separation because it does not arrive as a polite extension of conventional compute. It arrives as a concentration of energy, heat and operational volatility that makes those boundaries harder to defend.

That matters because once heat stops being a secondary effect and starts shaping performance directly, the discussion changes. The issue is no longer whether a facility can be made to accommodate a new generation of servers. The issue is whether the underlying design logic of the facility still makes sense when the economics, density and behaviour of AI systems are driving toward a far tighter relationship between processing and thermodynamics. In that context, cooling is no longer a service wrapped around the machine. It becomes part of the machine itself.

A building with the wrong assumptions

The older logic of the data centre was built around distribution. Heat loads were spread more evenly, airflow could be treated as a facility-wide discipline, and efficiency gains could be pursued through improvements in containment, circulation and plant performance. Even when workloads intensified, the model broadly held because the physical behaviour of the infrastructure remained legible. The building could still be understood as a controlled environment in which IT equipment sat inside a larger mechanical order.

AI is disrupting that arrangement because the density of the new systems is not simply higher, it is differently concentrated. Processing power is being compressed into smaller footprints, and the resulting heat is generated in ways that are more intense, more localised and more variable than the environments for which mainstream data centre design was optimised. At that point, airflow starts to look less like a durable answer and more like a legacy habit, one that can still be stretched, but only by forcing ever greater effort into preserving assumptions that are becoming less useful by the year.

The problem is not that air suddenly ceases to function. It is that the cost, complexity and inefficiency involved in making it function at the edge of these new requirements become increasingly difficult to justify. Large volumes of moving air, carefully engineered pathways and ever more assertive facility controls can keep buying time, but they do not alter the underlying imbalance. Heat is still being generated in places and at intensities that make distance, dilution and generalised environmental management less persuasive as a long-term strategy. What is being exposed is not a failure of engineering competence, but a mismatch between a previous design philosophy and a new computational reality.

This is why some of the most interesting work in the market is no longer asking how to cool the next generation of AI hardware within the conventions of yesterday’s data centre. It is asking a more awkward question. What if the conventions themselves are the problem? What if the real issue is that the industry has been trying to fit a new machine inside an old building concept, rather than accepting that the machine is now forcing a redesign of the building from the inside out.

Closer to the heat

Once that question is taken seriously, the debate changes in tone. Immersion cooling stops looking like a niche or specialist answer aimed at exotic deployments and starts to appear as part of a broader architectural correction. The significance of that shift is often misunderstood. It is easy to describe immersion simply as a more efficient method of removing heat, which it is. It is harder, and more important, to see that it also represents a different philosophy of infrastructure.

Instead of trying to manage thermal behaviour at one remove, across aisles, rooms and mechanical systems, immersion brings the cooling medium directly into the environment where heat is generated. That reduces distance, reduces ambiguity and reduces the reliance on air as an intermediary that must be pushed, directed and controlled across a much larger volume. The effect is not merely better thermal performance. The effect is a more contained and more coherent relationship between compute and cooling, one in which the infrastructure begins to act less like a layered facility and more like an integrated system.

Immersion has often been discussed in terms of incremental efficiency, yet that framing understates what is changing. The more consequential shift is architectural, a move away from managing heat at a distance toward designing systems in which thermal behaviour is controlled at source.

The more interesting argument is not that fluid is somewhat better than air at dealing with thermal loads. It is that immersion points toward a different way of thinking about the data centre itself, one in which the physical environment surrounding compute is no longer treated as an afterthought or external condition, but as part of the design logic of the system. That distinction shifts the conversation away from incremental efficiency and toward a more fundamental question of how AI infrastructure is designed.

Seen in that light, the rise of immersion is not just a response to hot chips and crowded racks. It is a challenge to the inherited architecture of digital infrastructure. The industry has been accustomed to buildings that host compute. AI is encouraging the emergence of facilities that are shaped by compute at a much more fundamental level, with thermal containment, energy movement and physical integration becoming central to the design rather than subordinate to it. That change will not happen all at once, and it will not happen evenly, but the direction of travel is already difficult to miss.

The new argument about efficiency

There is another reason this matters, and it goes beyond engineering purity. The economics of AI infrastructure are becoming harder, not softer. Power is constrained, land is constrained, timelines are compressed, and the appetite for inefficiency is declining rapidly now that scale has moved from ambition to expenditure. In that environment, thermal management is not just a technical matter. It becomes a question about whether capital is being translated into usable performance or dissipated through the effort of preserving an increasingly strained operating model.

That is where immersion begins to take on a broader strategic meaning. By rethinking the thermal environment more radically, it offers a route not only to higher densities but to a different balance between compute and overhead. The point is not that every AI environment will immediately abandon established approaches in favour of full immersion. The point is that the old assumption, that cooling is essentially a background utility whose job is to keep the room within range, is no longer sufficient for serious discussion about AI infrastructure economics.

This also helps explain why the conversation is shifting from equipment to architecture. The companies that will matter most in this part of the market are not simply those offering components within a crowded cooling ecosystem. They are the ones articulating a more convincing idea of what the AI data centre is becoming. That is a more strategic contest, because it touches the entire shape of future infrastructure, from the layout of halls and the design of pods through to the relationship between power delivery, thermal control and operational resilience. It is a conversation about the next physical model of compute, not just the next product category.

Framing immersion as a marginal improvement understates what is taking place. The more important development is a rethinking of how infrastructure is constructed around thermal behaviour. Those claims are already abundant. What matters more is understanding how immersion fits within a wider shift, one in which the assumptions underpinning the data centre are being reworked under AI pressure, and a more integrated physical logic is beginning to emerge. That is the level at which a company becomes part of a structural industry narrative rather than a narrow technology discussion.

Not an extension but a redesign

The temptation in fast-moving markets is always to describe change as an extension of what already exists. It sounds calmer, more manageable and more commercially digestible. Yet AI infrastructure is beginning to resist that language because the stresses are now too visible. Racks are denser, thermal tolerances are tighter, and the interaction between compute ambition and physical reality is becoming more confrontational. This is no longer just a story of faster chips and larger clusters. It is a story about whether the buildings and systems around those chips still embody the right assumptions.

That is why the most serious cooling conversation in AI is not ultimately about cooling at all. It is about design authority. The question is who gets to define the next infrastructure model. The facility tradition, which sees the data centre as a controlled building in which IT can be housed, still exerts a powerful influence. But an alternative view is gathering weight, one in which AI systems demand a more intimate alignment between physical environment and computational function, with thermodynamics treated not as a background constraint but as a governing principle.

The companies likely to shape that future will not be the ones merely helping yesterday’s architecture endure a little longer. They will be the ones showing that AI infrastructure must be reimagined from first principles, with the management of heat brought into the centre of the design. That does not guarantee that one method will dominate every deployment or that legacy environments will suddenly become obsolete. It does mean, however, that the argument has moved beyond whether established facilities can cope. The more important question now is whether coping is an ambitious enough objective for the scale of what is being built.

If AI continues on its present trajectory, the answer will become increasingly difficult to avoid. The data centre was not designed for this era of computing. The next generation of infrastructure will have to be.

Related Posts
Others have also viewed

The processor everyone forgot is now running the AI economy

The AI boom has been framed as a triumph of acceleration, yet the system is ...

The network is no longer infrastructure it is the constraint on AI

AI is not failing at the model layer, it is failing in motion, in the ...

The data centre was not designed for AI

Artificial intelligence is being scaled inside buildings conceived for a different era of computing. What ...

The real limit of AI infrastructure is not compute, it is heat

AI infrastructure is being designed around performance metrics that assume unlimited scaling. The reality is ...