AI-Ready Data Centers: Blackwell GPUs, Cooling, and Aethir’s DePIN

Learn how AI-ready data centers work and find out why Aethir’s DePIN model is ready to support the future of enterprise AI compute demand.

Featured | 
Community
  |  
May 12, 2026

Key Takeaways

  1. Blackwell Changes the Thermal Equation: The NVIDIA GB200 NVL72 rack runs at 120 to 132 kW, far beyond what air cooling can handle at any density. 
  2. Direct-to-Chip Cooling Is the New Baseline: Direct-to-Chip (DTC) cooling now commands roughly 65 percent of the liquid-cooling market in 2026. 
  3. AI Factory CapEx Creates Structural Lock-In: At 120 kW or more per rack, operators face capital costs of $500K to $2M per megawatt for cooling infrastructure alone. This lock-in gives hyperscalers an advantage and creates real barriers to access for enterprises and developers. 
  4. DePIN Sidesteps the CapEx Wall: Aethir’s decentralized GPU network aggregates enterprise-grade hardware, including Blackwell-class GPUs across distributed nodes worldwide. Clientsyxcv access that compute without absorbing the infrastructure investment required for liquid cooling, CDUs, and high-density rack builds. 

The Blackwell Density Problem: Why Air Cooling Is Over

The NVIDIA Blackwell architecture has reset every assumption about data center thermal management. The GB200 NVL72 configuration packs 72 Blackwell GPUs and 36 Grace CPUs into a single rack, delivering a thermal load of 120-132 kW. That number ends the era of hot aisle containment as a primary cooling strategy. Air cannot remove heat at that density.

Thermal Design Power Has Redefined the Infrastructure Stack

Each Blackwell GPU generates up to 1,000 watts of thermal design power (TDP), more than three times the output of GPU generations from just seven years ago. At rack densities of 120 kW or higher, traditional hot-aisle containment fails to remove heat quickly enough, creating hotspots that throttle performance and damage hardware. The NVIDIA reference architectures for GB200 and GB300 deployments mandate direct-to-chip liquid cooling as a specification requirement.

Hot Aisle Containment Signals an Air-Only Past

Hot-air containment was the dominant thermal management approach for data centers running 15-30 kW racks. At Blackwell density, the same containment geometry that once managed heat now traps it. Facilities running air-only architectures cannot qualify as AI-ready data centers under any current technical definition, and any operator planning a Blackwell deployment without a liquid cooling plan is planning to fail.

GB300 NVL72 Raises the Thermal Bar Further

The GB300 NVL72, the next iteration of the Blackwell platform, supports up to 142 kW per rack in reference designs co-developed by Schneider Electric and NVIDIA. Each generation of the AI factory architecture pushes rack density higher and further tightens the thermal requirements. Data center operators who defer liquid-cooling infrastructure upgrades are not only falling behind on efficiency but also falling out of contention for Blackwell and post-Blackwell deployments entirely.

Direct-to-Chip Cooling: The AI-Ready Infrastructure Stack

The architecture of an AI-ready data center in 2026 is defined by its cooling stack. Direct-to-chip (DTC) cooling routes coolant directly to heat-generating components via a cold plate mounted on each chip. A coolant distribution unit (CDU) manages the loop, and a manifold distributes flow across the rack. This cold plate, CDU, and manifold stack has become the standard for any facility running high-density GPU compute at AI factory scale.

Cold Plate Technology Is the Core Thermal Interface

A cold plate is a metal block mounted directly on the GPU or CPU through which liquid coolant flows to absorb heat at the source. Cold plate cooling achieves PUE (power usage effectiveness) of 1.10 to 1.25, compared to 1.50 to 1.80 for traditional air-cooled facilities, and delivers proportional improvements in water usage effectiveness (WUE) for operators with sustainability targets. Direct-to-chip is now the dominant liquid-cooling approach, commanding roughly 65 percent of the liquid-cooling market in 2026.

Single-Phase vs Two-Phase Immersion Cooling

Beyond direct-to-chip, single-phase immersion cooling submerges entire servers in non-conductive liquid, achieving PUE as low as 1.02 to 1.10. Two-phase immersion, where coolant boils and recondenses in a closed loop, reaches PUE of 1.01 to 1.05 and is required for the most extreme rack densities above 140 kW. Both approaches require specialized facility design and represent larger capital investments than direct-to-chip, making them suitable for operators building net-new AI-ready data centers rather than upgrading existing infrastructure.

The Rear-Door Heat Exchanger as a Hybrid Transition Play

A rear-door heat exchanger (RDHx) fits onto the back of existing server racks and cools exhaust air before it re-enters the facility. RDHx serves as a bridge technology for operators transitioning from air-cooled to fully liquid-cooled infrastructure without a complete rebuild. At Blackwell densities, RDHx alone is insufficient, but it extends the viable operating range of partially air-cooled facilities running mixed workloads during a phased cooling migration.

The CapEx Reality of Building AI-Ready Data Centers

Building an AI-ready data center is not a software problem. Liquid cooling infrastructure adds $500K to $2M per megawatt in capital cost. A 10 MW GPU cluster requires between $5M and $20M in cooling infrastructure before the first GPU is powered on. That expense compounds when grid connection timelines, hardware procurement lead times, and low utilization rates are taken into account.

Grid Connection Timelines Have Become Structural Barriers

Connecting a new data center to the power grid now takes up to seven years in some regions, before permitting, construction, or hardware procurement are considered. For any organization that is not a tier-1 hyperscaler, the lead time between the investment decision and the operational AI-ready infrastructure makes a centralized build-out unrealistic. Grid stress is the top challenge cited by 79 percent of data center executives in a 2025 Deloitte survey, with AI expected to drive spikes in power demand through 2035.

GPU Supply Chain Constraints Compound the CapEx Problem

NVIDIA H100 and H200 chips cost over $40,000 each and remain in critically short supply. Enterprises seeking access to Blackwell-class hardware face 18 to 24-month waiting periods for premium GPU instances from major cloud providers. Supply chain constraints mean that even organizations willing to absorb the capital expense of an AI-ready data center build cannot guarantee hardware availability on any predictable timeline.

Low GPU Utilization Undermines the Business Case

GPU utilization in traditional data centers hovers between 30 and 50 percent, meaning operators pay for hardware that sits idle more than half the time. This utilization gap reflects procurement practices, workload-scheduling inefficiencies, and a structural mismatch between centralized data center capacity and the burst-and-pause nature of AI inference demand. The financial model for centralized AI-ready data center construction is broken for any organization operating at a scale below that of hyperscalers.

DePIN Data Centers: The Distributed Alternative Pioneered by Aethir

Decentralized Physical Infrastructure Networks (DePIN) take a fundamentally different approach to AI-ready infrastructure, and Aethir is pioneering decentralized GPU cloud computing for enterprise use cases. Instead of constructing centralized liquid-cooled facilities, DePIN networks aggregate enterprise-grade GPU hardware across distributed nodes worldwide. 

The result is a compute layer that delivers access to high-density GPU infrastructure without requiring any single operator to absorb the full CapEx of a Blackwell deployment or an AI-ready data center build. Aethir’s DePIN stack has 430,000+ GPU Containers distributed globally, across 200+ locations in 94 countries, along with a roster of 150+ partners and enterprise customers leveraging decentralized cloud computing.

Aethir’s roster includes thousands of H100s, H200s, GB200s, B200s, along with an upcoming deployment of 2300+ B300s.

Distributed Nodes Bypass the Cooling CapEx Requirement

In a DePIN data center model, individual node operators contribute existing hardware to a shared network. Each operator manages the cooling requirements for their own hardware, which may already include direct-to-chip systems. The network aggregates that capacity and exposes it through a unified compute layer, eliminating the need for any single enterprise to fund CDUs, manifolds, or rack-density upgrades at data center scale.

Aethir Delivers Major Cost Reduction vs Hyperscalers

Aethir’s decentralized GPU network provides access to enterprise-grade compute, including Blackwell-class GPUs at major discounts compared to AWS rates for comparable inference workloads. This cost structure reflects the elimination of centralized data center overhead, not promotional pricing.

On-Demand Access Removes the 18-Month Procurement Lead Time

One of the defining constraints of centralized AI-ready data center access is procurement lead time. Enterprises waiting for H100 or Blackwell instances through major cloud providers face 18 to 24-month queues. Aethir provides on-demand access to distributed GPU infrastructure, enabling enterprises and developers to run inference workloads, fine-tune models, and perform AI agent tasks without entering a multi-year procurement cycle or absorbing the capital cost of an AI-ready data center.

The AI Factory Era and What It Means for Compute Access

NVIDIA coined the term "AI factory" to describe the shift from data centers serving as passive storage facilities to active compute production systems. In this framing, the data center becomes a factory floor where raw data goes in, and intelligence comes out. GPU cloud liquid cooling is not a feature upgrade but the prerequisite for operating at AI factory throughput. The question for most enterprises is not whether to adopt this paradigm but how to access it without building a liquid-cooled AI-ready data center from scratch.

Inference Demand Drives 70 Percent of Total GPU Load

Seventy percent of GPU demand in 2026 is driven by inference, not training. Inference is inherently parallelizable, short-duration, and latency-sensitive, a profile that favors distributed compute over centralized data center queues. DePIN networks are structurally optimized for inference workloads in ways that large centralized facilities, designed primarily for training runs, are not.

Aethir’s decentralized GPU cloud is purpose-built to support inference workloads for developer teams of all sizes, startups, and large-scale enterprises.

The AI Factory Output Advantage Flows Through Distributed Access

NVIDIA reports that the GB300 NVL72 delivers 50 times higher AI factory output than Hopper-generation platforms, combining a 10x latency reduction and 5x higher throughput per megawatt. These gains are only realizable if enterprises can actually access the hardware. Decentralized compute networks extend AI factory-class performance to organizations that cannot afford the infrastructure investment required for an AI-ready data center build.

Aethir Positions DePIN at the Intersection of All Three Trends

Aethir operates at the convergence of Blackwell-era GPU demand, the liquid cooling transition, and the DePIN infrastructure model. Aethir’s decentralized GPU cloud provides access to enterprise-grade GPU compute, supports AI inference workloads at scale, and requires no participant to build or maintain a liquid-cooled, AI-ready data center. 

For developers, enterprise teams, and AI practitioners, that convergence is the core value proposition of decentralized GPU cloud infrastructure.

FAQ

What makes a data center AI-ready in 2026?

An AI-ready data center in 2026 is defined by its ability to support high-density GPU compute at 60 kW per rack or higher, backed by direct-to-chip liquid cooling infrastructure. Facilities that rely exclusively on air cooling or hot aisle containment cannot meet the thermal requirements of current-generation hardware like the NVIDIA GB200 NVL72 or GB300. 

What is direct-to-chip cooling and why do Blackwell GPUs require it?

Direct-to-chip cooling routes liquid coolant through a cold plate mounted directly on the GPU or CPU, removing heat at the source rather than cooling ambient air. Blackwell GPUs generate up to 1,000 watts of thermal design power per chip, making air-based cooling physically insufficient for the rack densities Blackwell deployments require. The NVIDIA reference architectures for GB200 and GB300 systems include direct-to-chip cooling as a mandatory specification.

What is rack density and how does it affect data center design?

Rack density measures the power consumption per server rack, expressed in kilowatts (kW). A rack running at 15 kW can be managed with standard air cooling, while a Blackwell rack running at 120-132 kW requires a dedicated liquid-cooling loop with a CDU, manifold, and cold plates for each GPU. As AI workloads become more demanding, rack density has become the primary design constraint for any operator building or upgrading toward an AI-ready data center.

How does DePIN infrastructure compare to traditional AI-ready data centers?

Traditional AI-ready data centers require massive capital investment in cooling, power infrastructure, and real estate before a single GPU is deployed. DePIN infrastructure like Aethir aggregates distributed GPU hardware contributed by node operators worldwide, eliminating the need for any single organization to fund a centralized build. 

Can enterprises access Blackwell-class GPUs without building liquid-cooled data centers?

Yes. Aethir provides access to Blackwell-class GB200, B200, and B300 enterprise-grade compute without requiring any individual enterprise to build or maintain liquid-cooled infrastructure. Aethir’s DePIN model means that Cloud Host operators who already run Blackwell hardware contribute that capacity to the network, while enterprises and developers access it on demand. Aethir operates this model at scale, with distributed compute accessible through a unified platform, without multi-year procurement cycles or investment in cooling infrastructure.

 

Resources

Keep Reading