Why NVIDIA hosting is harder than regular colo
Most colocation facilities were built for 5–10 kW racks. Modern NVIDIA deployments break that envelope on day one. Less than 5% of data centers in the world can support even 50 kW per rack. Here’s what actually trips up deployments:
1. Power density wall
A single HGX H200 8-GPU server pulls ~10.2 kW under load. Stack 4 of those in a rack and you’re at ~40 kW — vastly exceeding the 10–12 kW/rack design point of typical retail colocation. B200 systems push that to 14+ kW per server, and Blackwell rack-scale (GB200 NVL72) lands at 120 kW per rack and rising.
2. Cooling and when liquid becomes mandatory
H100 and H200 (700W TDP per GPU) can be air-cooled if the facility has rear-door heat exchangers or aggressive containment. B200 (1,000W) and B300 (1,100–1,400W) require direct-to-chip liquid cooling. No exceptions. If a provider doesn’t already have CDUs (Coolant Distribution Units) plumbed, they’re not a candidate and standing one up takes months.
3. NVLink fabric proximity
Multi-node NVLink/NVSwitch deployments have hard cable-length limits. NVIDIA’s reference designs require leaf switches within ~30 m of GPU pods, and the NVLink scale-out switch caps at ~20 m. This means SuperPOD-class deployments need contiguous racks, not whatever space the colo has spare. Many providers say “we have 4 racks available”, but spread across 3 rows. That kills NVLink scale-out.
4. Network fabric
400G InfiniBand (NDR) or 100G Ethernet minimum for training. Storage and management on a separate fabric. If a colocation can’t deliver 400G to your cabinet at reasonable cross-connect pricing, the rest doesn’t matter.
5. NVIDIA DGX-Ready certification
For DGX systems specifically, NVIDIA maintains a DGX-Ready Colocation partner program. Equinix, Digital Realty, DataBank, and EcoDataCenter are in. Many regional providers aren’t, but can still host non-DGX HGX systems just fine. We help filter both.





