The data center industry is standing on the shoreline of a massive tidal shift. For decades, we have relied on the steady, predictable breeze of air cooling to keep our digital engines running. It was a comfortable era where raised floors, CRAC units, and hot-aisle containment were sufficient to tame the heat generated by general-purpose compute. But the wind has died down, and a tsunami is approaching.
That tsunami is Artificial Intelligence, and it brings with it a thermal density that air simply cannot carry.
We are no longer discussing a “trend” or an “emerging technology.” We are facing a hard physical wall. The era of air-cooled facilities for high-performance computing is effectively ending. As we move deep into late 2025 and 2026, operators are realizing that the liquid cooling mandate is not just about changing plumbing; it’s about survival in the age of the AI factory.
This isn’t just a facility upgrade, it is a complete reimagining of mission-critical thermal management.
The End of the Air Age
To understand why the industry is pivoting so violently, we have to look at the physics. Air is an insulator, not a conductor. It is famously inefficient at moving heat away from a source. For years, we compensated for this by moving more air, faster and colder. We built containment systems to stop mixing; we optimized airflow tiles; we installed massive fan walls.
But silicon has outpaced physics.
The latest generation of GPUs and AI accelerators are pushing thermal envelopes that would melt a traditional server chassis. We are seeing individual chips drawing over 1,000 watts. When you stack these into a high-density NVL72 rack, you aren’t dealing with the comfortable 10kW or even 20kW densities of the past. You are staring down the barrel of 100kW to 132kW per rack.
At these densities, air cooling ceases to be an engineering challenge and becomes a physical impossibility. To cool a 100kW rack with air, you would need a hurricane-force gale blowing through the server, which would vibrate the components to pieces and consume more energy in fan power than the compute itself.
This reality has birthed the liquid cooling mandate. It is no longer an optional “science project” for hyperscalers; it is the baseline requirement for any facility that intends to host modern AI workloads. If you cannot support liquid, you cannot support AI.
Direct-to-Chip & Immersion: The New Standard
The industry’s response to this thermal crisis has been a rapid maturation of liquid technologies. What was once niche is now standard.

Direct-to-Chip (DtC): The Surgical Approach
Direct-to-chip, or cold plate cooling, is currently the dominant form of this transition. By routing liquid directly to the hottest components, the CPU and GPU, we can capture 70-80% of the heat at the source. This allows facilities to push rack densities well past the 100kW mark while still utilizing some air cooling for the remaining low-heat components (like memory and storage).
DtC is favored because it feels familiar. It fits into standard racks and allows for serviceability that technicians recognize. However, it introduces a new complexity: the “plumbing” of the server. The risk of leaks, while statistically low with modern couplers, changes the psychological landscape of the data center floor.
Immersion Cooling: The Deep Dive
For those willing to go further, immersion cooling offers the holy grail of thermal management. By submerging the entire server in a dielectric fluid, you capture 100% of the heat. This eliminates fans entirely, creating a silent data center. While the operational changes are drastic, requiring cranes to lift servers out of “baths”, the efficiency gains are unmatched.
Both DtC and Immersion are rapidly moving from “testing” to “production.” The liquid cooling mandate is driving manufacturers to standardize quick-disconnect couplings, fluid chemistries, and manifold designs, making these systems as plug-and-play as possible.
The Rise of the Coolant Distribution Unit (CDU)
If the server is the engine, the Coolant Distribution Unit (CDU) is the heart.
In an air-cooled world, the CRAC unit was the king. In a liquid-cooled world, the CDU takes the throne. These units are responsible for exchanging heat between the facility water loop (FWS) and the technology cooling system (TCS) that goes into the rack. They manage flow rate, pressure, and temperature with surgical precision.
A trending operational focus is the installation and management of CDUs and the plumbing infrastructure required to bring liquid to the rack. This is where the rubber meets the road—or rather, where the water meets the chip.

Managing a CDU is fundamentally different from managing a CRAC.
- Pressure Management: You are now monitoring differential pressure to ensure coolant flows evenly across all nodes. A drop in pressure could indicate a leak; a spike could indicate a blockage.
- Chemistry Control: The quality of the fluid matters. Filtration, pH levels, and inhibitor concentrations must be monitored to prevent corrosion or biological growth inside the expensive cold plates.
- Redundancy: If a CDU fails, the rack overheats in seconds, not minutes. The thermal ride-through time of a liquid-cooled rack is virtually zero compared to the thermal mass of an air-cooled hall.
This shift places CDUs at the center of the DCIM (Data Center Infrastructure Management) conversation. You cannot manage the liquid cooling mandate with spreadsheets and clipboard walkthroughs. You need real-time telemetry from these CDUs integrated into a unified management platform.
The Management Gap: Facilities vs. IT
Perhaps the biggest risk in the liquid cooling mandate is not the plumbing, but the people.
Liquid cooling blurs the line between Facilities and IT. In the past, Facilities delivered cold air to the room, and IT placed servers in the rack. As long as the room was cold, everyone was happy. The two teams rarely needed to coordinate on a rack-by-rack basis.
Liquid cooling destroys that silo. The cooling is now inside the rack, physically tethered to the IT equipment.
- Who owns the leak? If a cold plate drips on a motherboard, is that a Facilities problem or an IT hardware failure?
- Capacity Planning: You can no longer just “roll in a rack.” You need to know if the specific CDU row has the hydraulic headroom (flow rate) to support the specific TDP (Thermal Design Power) of the new nodes.
- Maintenance: Swapping a server now involves unplugging fluid lines. Does the IT technician have the training to handle dripless connectors, or does a Facilities engineer need to be present?
This convergence creates a “management gap.” Without a unified view, Facilities might be optimizing the chiller plant for one set of parameters while IT is deploying hardware that requires completely different flow rates. The liquid cooling mandate requires a single source of truth—a platform that bridges the gap between the grey space (infrastructure) and the white space (IT).
Mastering Mission-Critical Thermal Management
So, how do operators survive the liquid cooling mandate? The answer lies in mastering the entire thermal chain.

It is no longer enough to monitor the room temperature. You must monitor the temperature of the fluid entering the CDU, the pressure at the manifold, the flow rate through the cold plate, and the junction temperature of the GPU itself.
1. Unified Visibility
You need a DCIM solution that speaks both “Facilities” (Modbus, BACnet) and “IT” (SNMP, Redfish). Nlyte, for example, excels at this intersection. By pulling data from the CDUs and correlating it with the power consumption of the servers, you get a complete picture of efficiency. You can see exactly how much cooling energy is required to support a specific AI job.
2. Predictive Modeling
With the margins for error shrinking, you cannot rely on reactive alarms. You need predictive modeling. If you add 10 more NVL72 racks to Row 4, what happens to the coolant pressure in Row 8? sophisticated thermal modeling and CFD (Computational Fluid Dynamics) simulations are essential for planning deployments in a liquid-cooled environment.
3. Automated Response
In a high-density AI environment, things happen fast. If a CDU fails, software needs to instantly identify the affected workloads and potentially trigger a graceful shutdown or migration before thermal damage occurs. “Agentic AI” in the management layer will become a critical ally, making micro-adjustments to flow valves to optimize cooling without human intervention.
Sustainability Considerations
The good news is that the liquid cooling mandate is also a sustainability win. Liquid is thousands of times more efficient at capturing heat than air. This allows for:
- Higher Inlet Temperatures: You can run liquid loops at 40°C (104°F) or higher, drastically reducing the need for energy-intensive mechanical chillers. In many climates, you can use free cooling (dry coolers) year-round.
- Heat Reuse: Because liquid captures heat so effectively, the return water is hot enough to be useful. Instead of venting low-grade heat into the atmosphere, data centers can pipe 60°C water into district heating systems, warming nearby homes or offices.
This turns the data center from a consumer of resources into a provider of heat, flipping the narrative on environmental impact.
Conclusion: Sink or Swim
The tide has turned. The days of simply blasting cold air into a room are numbering. The AI wave is crashing down, and it demands a new approach to physics, engineering, and management.
The liquid cooling mandate is challenging, yes. It requires new skills, new hardware, and a new level of collaboration between IT and Facilities. But it also offers a path to unprecedented density, efficiency, and sustainability.
For data center operators, the choice is simple: grab a board and ride the wave or get crushed by it. By adopting a holistic approach to thermal management, one that integrates CDUs, tracks hydraulic capacity, and unifies the view of IT and Facilities, you can turn this mandate into your competitive advantage.
Surf’s up. Is your facility ready?