Data Center Server Rack Layer


The rack aggregates dozens of servers into a standardized enclosure, providing shared power, cooling, and networking. It is the fundamental deployment unit inside a data center, bridging individual servers to larger clusters. With AI workloads driving 40–80 kW per rack, design has shifted toward liquid cooling, prefabrication, and dense interconnects.


Architecture & Design Trends

  • Power Density: Traditional enterprise racks operated at 5–10 kW, but AI racks routinely exceed 40 kW and can reach 80–100 kW with liquid cooling.
  • Form Factors: 19” racks remain standard, though Open Rack (OCP) formats allow for higher density and front-access cabling.
  • Cooling Evolution: Rear-door heat exchangers and liquid manifolds are replacing air-only systems; immersion tanks at rack level are emerging for ultra-dense nodes.
  • Prefabrication: Hyperscalers increasingly procure fully populated and factory-tested racks, reducing onsite integration time.
  • Networking Integration: Top-of-rack (TOR) switches and structured cabling trays are consolidated to simplify scaling at the pod/cluster level.

AI Training vs General-Purpose Racks

AI training racks differ significantly from traditional enterprise racks in power, cooling, density, and integration. The table below highlights the key differences.

Dimension AI Training Racks General-Purpose Racks
Primary Use GPU-dense AI training servers (40–100 kW per rack) CPU-based IT/business servers (5–15 kW per rack)
Power Distribution Three-phase PDUs, high-current busbars, 48VDC options Single/three-phase PDUs, standard 208/230V
Cooling Rear-door heat exchangers, liquid manifolds, immersion tanks Air cooling with fans, limited liquid retrofit
Networking TOR switches with 400–800G Ethernet or InfiniBand Standard TOR/EOR switches with 10–100G Ethernet
Weight 1500–2000+ lbs fully loaded (GPU + liquid + power gear) 800–1200 lbs typical (CPU + storage)
Integration Factory-integrated with servers, cabling, and cooling Populated onsite with mixed workloads
Monitoring Dense sensors (temp, leak detection, power telemetry) Basic temp/humidity sensors, door locks
Vendors Schneider, Vertiv, Rittal, Supermicro, Inspur, ODMs APC, Dell, HPE, Lenovo, Cisco, Tripp Lite
Cost $250K–$1M+ per fully populated rack $25K–$100K per populated rack

Notable Vendors

Vendor Product Line Form Factor Key Features
Schneider Electric EcoStruxure Racks 19" & OCP racks Integrated PDUs, cooling options, prefabrication
Vertiv VRC-S / SmartRow 19" racks Rack+cooling+PDU pre-integrated solutions
Rittal TS IT / Liquid Cooling Packages 19" racks Rear-door HX, modular liquid distribution
HPE Apollo & OCP racks OCP sled racks High-density AI server integration
Supermicro GPU-optimized rack solutions 4U server racks Turnkey GPU rack-scale systems
Inspur Rack-scale AI clusters OCP & 19" racks Factory-integrated GPU racks, China market leader
ODM Integrators Quanta, Wiwynn, Foxconn Custom hyperscale racks Prefabricated at scale for cloud providers

Server Rack BOM

Domain Examples Role
Compute Rack-scale GPU/CPU servers, blade enclosures Aggregates compute resources
Memory CXL memory switches, pooled DIMM shelves Improves utilization across servers
Storage NVMe-oF arrays, JBOD/JBOF units Rack-local persistent storage
Networking Top-of-rack switches, patch panels, structured cabling Links servers to cluster fabric
Power Rack PDUs, busbars, DC-DC shelves, rack-level battery backup Distributes and conditions power
Cooling Rear-door heat exchangers, liquid manifolds, immersion tanks Removes rack-level heat loads
Monitoring & Security Rack sensors (temp, humidity, airflow), electronic locks Provides telemetry and access control
Prefabrication Factory-integrated racks with PDU, cooling, and cabling pre-installed Speeds deployment and reduces onsite labor

Key Challenges

  • Thermal Limits: Traditional air cooling cannot handle >40 kW; liquid distribution manifolds are mandatory in AI racks.
  • Power Delivery: Racks require three-phase PDUs, higher amperage busbars, and sometimes direct 48VDC distribution.
  • Weight & Floor Loading: Fully loaded racks can exceed 1500–2000 lbs, stressing raised-floor designs.
  • Integration Complexity: Cabling (fiber + copper) and liquid manifolds add significant integration complexity.

Future Outlook

  • Liquid Standardization: Cold plates and liquid manifolds will be universal in AI racks by 2026.
  • Immersion Adoption: Rack-level immersion tanks will expand beyond pilots into mainstream hyperscale sites.
  • 48VDC Power: Direct DC distribution at rack level will reduce conversion losses and simplify designs.
  • Smart Racks: Embedded sensors and AI-driven DCIM integration will make racks self-monitoring and semi-autonomous.

FAQ

  • How much power does an AI rack consume? Typical ranges are 40–80 kW, with next-gen racks designed for >100 kW.
  • Do racks come fully integrated? Hyperscalers increasingly purchase racks pre-populated with servers, PDUs, and cabling.
  • What is the role of TOR switches? They aggregate server NICs within the rack and link to the cluster fabric.
  • How are liquid-cooled racks different? They contain distribution manifolds, rear-door heat exchangers, and leak-detection sensors.
  • Can racks still be air cooled? Enterprise racks often are, but AI racks at >40 kW require liquid assistance.