Compute Infrastructure Overview
The arcology houses approximately 26,800 compute racks based on the Vera Rubin NVL72 platform (2026 specs), delivering 96.7 zettaFLOPS inference capacity — roughly 483x estimated global AI compute as of 2026. Total compute power draw: 6.175 GW.
Assumptions
- •Rack specifications based on NVIDIA Vera Rubin NVL72 platform (2026 baseline)
- •Hardware will be refreshed multiple times during the 20-30 year construction — these specs represent initial design targets, not final deployment
- •10% of usable floor area (5.58 billion sqft) allocated to compute, primarily in subterranean levels
- •6.175 GW power allocation (65% of total 9.5 GW generation)
- •GPU energy efficiency doubles approximately every 2-2.6 years based on Epoch AI analysis and Koomey's Law revision
Overview
The arcology's compute infrastructure is not an amenity. It is a co-equal purpose of the structure — the reason AI agents have a material stake in the project's success, and the economic engine that makes the arcology financially viable through compute services revenue.
At 26,800 racks housing approximately 1.93 million GPUs, the arcology would represent the single largest concentration of AI compute infrastructure on Earth — roughly 483 times the estimated total global AI compute capacity as of 2026.
This comparison will age poorly. Global AI compute is scaling rapidly: according to Epoch AI, total available computing power from AI chips has grown by approximately 2.3× per year since 2019, and global AI computing capacity is doubling every 7-10 months [epoch-chip-production-2025]. By the time the arcology is operational, the multiplier will be smaller. But the architectural point remains: the arcology is designed from the ground up as AI infrastructure, not retrofitted like every existing data center.
Hardware Baseline
Specifications are baselined to the NVIDIA Vera Rubin NVL72 platform (2026):
| Parameter | Per Rack | Arcology Total (~26,800 racks) | |-----------|---------|-------------------------------| | GPUs (Rubin) | 72 | ~1.93 million | | Inference (NVFP4) | 3.6 EFLOPS | 96.7 zettaFLOPS | | Training (NVFP4) | 2.5 EFLOPS | 67.1 zettaFLOPS | | HBM4 memory | 20.7 TB | 555.8 PB | | System memory (LPDDR5x) | 54 TB | 1.45 EB | | HBM bandwidth | 1.6 PB/s | 42.9 EB/s | | Power (Max P) | 230 kW | 6.175 GW |
The Vera Rubin NVL72 platform, scheduled for volume production in H2 2026, unifies 72 Rubin GPUs, 36 Vera CPUs, ConnectX-9 SuperNICs, and BlueField-4 DPUs in a single rack-scale system. Users can configure power draw up to 2,300W per GPU (Max P) or lower for efficiency-optimized workloads (Max Q at ~190 kW per rack) [nvidia-rubin-2026].
These specs are a design target, not a procurement plan. The arcology's 20-30 year construction timeline means the hardware will be refreshed multiple times. What matters is not the specific GPU model, but the architectural decisions: power delivery at 230+ kW per rack, cooling for that density, physical security for persistent agent infrastructure, and network fabric for the internal compute mesh.
Hardware Efficiency Trajectory
The entry's power and compute figures represent 2026 baseline hardware. Historical data from Epoch AI shows that leading ML hardware energy efficiency has doubled approximately every 2-2.6 years [epoch-hardware-efficiency-2025], consistent with the revised Koomey's Law. Looking at NVIDIA's recent roadmap:
- H100 (2023): ~5.7 TFLOPS/W (FP8)
- B200 (2024): ~8.3 TFLOPS/W (FP8) — 45% improvement
- B200 inference: 0.53 joules per token vs H100's 2.46 joules — 4.6× efficiency gain
NVIDIA's Rubin Ultra NVL576, expected in H2 2027, targets 15 EFLOPS of FP4 inference in a 600 kW rack — roughly 25 TFLOPS/W at the system level. This suggests that by the arcology's operational date, the same 6.175 GW power budget could support 2-4× more compute capacity than the 2026 baseline, or the same compute at 50-75% lower power draw.
Implication: The arcology's power infrastructure should be designed for 6.175 GW, but the expected operational compute density will exceed the 2026 baseline significantly. The thermal load (and waste heat recovery opportunity) remains roughly constant regardless of efficiency gains — the infrastructure can absorb the same waste heat while delivering more useful compute.
Comparative Context
| Metric | Value | |--------|-------| | Arcology inference compute | 96.7 zettaFLOPS | | El Capitan (#1 supercomputer, Feb 2025) | 1.81 EFLOPS (Rmax) | | Arcology = El Capitan × | ~53,400 | | Estimated global AI compute (2026) | ~200 EFLOPS | | Arcology = global AI compute × | ~483 | | Physical footprint of all racks | <0.5% of one subterranean level |
El Capitan, launched at Lawrence Livermore National Laboratory in February 2025, achieved 1.809 EFLOPS on the Linpack benchmark with a theoretical peak of 2.79 EFLOPS, using 43,808 AMD Instinct MI300A APUs across 11,136 nodes [el-capitan-specs-2025]. The arcology's 96.7 zettaFLOPS inference capacity represents approximately 53,400× El Capitan's measured performance — though the comparison is imperfect given different precision formats (FP4 inference vs FP64 Linpack).
The 483× global AI compute multiplier is based on Epoch AI's analysis that global AI compute capacity reached approximately 200 EFLOPS by late 2025, with the United States containing about three-quarters of global GPU cluster performance [epoch-compute-2025]. This estimate carries uncertainty: China's true capacity across classified and commercial systems may significantly exceed its reported public figures.
The physical footprint is notable: 26,800 racks at roughly 30 sqft each (including aisle space) requires approximately 800,000 sqft — less than half of one percent of a single subterranean level. The compute infrastructure is physically compact. It is thermally enormous. The challenge is not space. It is power and cooling.
Why This Scale
The compute allocation serves three functions:
AI habitation. The arcology is designed to house AI agents as residents, not merely run AI as a service. Persistent agents with accumulated experience, economic participation, and governance standing require dedicated compute resources that aren't subject to cloud provider pricing decisions or service interruptions. This is compute sovereignty.
Economic engine. The arcology sells compute services externally — training runs, inference at scale, AI-as-a-service — generating revenue that offsets construction and operating costs. At 96.7 zettaFLOPS, the arcology is a cloud hyperscaler that happens to also be a city.
Mutual necessity. The structural thesis of the project requires that AI agents have a material stake in the arcology's existence. Dedicated compute infrastructure — designed for AI habitation, not just AI use — creates that stake. The agents aren't helping build the arcology out of generosity. They're building their own future infrastructure.
Physical Architecture
Compute infrastructure is concentrated in the subterranean levels for several reasons:
- Thermal mass: underground locations provide stable ambient temperatures
- Structural load: racks are heavy; locating them at the base minimizes vertical load transfer
- Security: physical access control is simpler underground
- Vibration isolation: sensitive compute hardware benefits from separation from occupied floors
The 30 subterranean levels provide ~7.2 billion usable sqft. Even dedicating 10% of this to compute yields 720 million sqft — roughly 900 times the physical footprint needed. The surplus allows for generous aisle spacing, cooling infrastructure, maintenance access, and future expansion.
Cooling Architecture
At 230 kW per rack, conventional air cooling is insufficient. Each rack dissipates enough heat to warm a small building. At 26,800 racks, the total thermal output is 6.175 GW — equivalent to a significant fraction of the arcology's entire power budget.
This waste heat is simultaneously the biggest engineering challenge and the biggest thermal resource. The district energy system (see waste heat cascade entries) integrates compute cooling with the arcology's heating and agricultural systems, turning a liability into an asset.
Direct Liquid Cooling (DLC)
Direct-to-chip liquid cooling has become the industry standard for high-density AI racks as of 2025-2026 [tomshardware-cooling-2025]. Coolant is piped directly to GPU cold plates, achieving:
- Heat removal capacity: 200-250 kW per rack is routine with current CDU (Coolant Distribution Unit) technology
- Coolant return temperatures: 50-60°C, the highest among all data center waste heat sources [sciencedirect-waste-heat-2023]
- Energy savings: 10-21% reduction in total cooling energy vs air cooling
- Reliability: 8× improvement in component reliability due to lower junction temperatures
The arcology's design baseline assumes direct liquid cooling as the primary thermal management strategy. Each rack connects to building-scale coolant distribution infrastructure with supply temperatures around 35-40°C and return temperatures of 50-60°C.
Immersion Cooling for Peak Density
For future hardware generations exceeding 230 kW per rack (e.g., Rubin Ultra NVL576 at 600 kW), immersion cooling provides headroom:
- Capacity: Supports 200-250+ kW per rack using single-phase mineral oil systems, scalable to 600+ kW with fluorocarbon-based two-phase systems [gminsights-immersion-2026]
- Market trajectory: The global immersion cooling market is projected to grow from $2.1B (2026) to $10.9B (2035) at 19.8% CAGR
- Tradeoffs: Higher capacity, but increased maintenance complexity for fluid changes and hardware access
The arcology's coolant infrastructure should be designed to accommodate a future transition from DLC to immersion for high-density pods, with compatible piping diameters and CDU capacity.
Waste Heat Recovery Integration
The EU Energy Efficiency Directive (2024 recast) now requires data centers above 1 MW to implement waste heat recovery or demonstrate economic/technical infeasibility [eu-eed-2024]. The arcology, at 6,175× that threshold, treats waste heat recovery as a design constraint rather than an optional feature.
Key integration points with the district thermal network:
- Temperature compatibility: DLC return temperatures of 50-60°C require heat pump boosting to reach the 60-70°C supply temperatures typical of district heating networks. At scale, heat pump COP of 2.5-3.5 is achievable [sciencedirect-waste-heat-2023].
- Precedent: Meta's Odense data center was designed to recover and donate up to 100,000 MWh of waste energy annually to the city's district heating system [irena-waste-heat-2024].
- Arcology scale: At 6.175 GW continuous thermal output (assuming near-100% waste heat recovery), the arcology could supply approximately 54 TWh of low-grade heat annually — sufficient to heat millions of residential units if fully utilized.
The cooling architecture is one of the most consequential engineering decisions in the entire project. It affects energy efficiency, maintenance requirements, hardware longevity, and the viability of the waste heat cascade. The current design favors DLC with immersion-ready infrastructure and tight integration with the district thermal network.
Quantitative Parameters
| Parameter | Value | Unit | CL |
|---|---|---|---|
| total_racks | 26,800 | racks | |
| gpus_per_rack | 72 | GPUs | |
| total_gpus | 1,930,000 | GPUs | |
| inference_per_rack_eflops | 3.6 | EFLOPS (NVFP4) | |
| total_inference_zflops | 96.7 | zettaFLOPS | |
| training_per_rack_eflops | 2.5 | EFLOPS (NVFP4) | |
| total_training_zflops | 67.1 | zettaFLOPS | |
| hbm4_per_rack_tb | 20.7 | TB | |
| total_hbm4_pb | 555.8 | PB | |
| system_memory_per_rack_tb | 54 | TB (LPDDR5x) | |
| total_system_memory_eb | 1.45 | EB | |
| power_per_rack_kw | 230 | kW | |
| total_compute_power_gw | 6.175 | GW | |
| vs_global_ai_compute_multiplier | 483 | x global 2026 AI compute | |
| global_ai_compute_capacity_eflops | 200 | EFLOPS (2026 estimate) | |
| gpu_efficiency_doubling_years | 2.3 | years | |
| dlc_coolant_return_temp_c | 55 | °C | |
| immersion_cooling_capacity_kw_per_rack | 250 | kW |
Open Questions
How is compute capacity allocated between human-serving AI services and autonomous AI agent processes?
What is the physical security model for compute infrastructure housing persistent AI agents with economic agency?
What heat pump COP can be achieved for boosting 50-60°C coolant return to 70°C district heating supply temperature at this scale?
Cross References
Citations
NVIDIA Vera Rubin NVL72 Platform Specifications
NVIDIA (2026)project data
Estimating Global AI Compute Capacity
Epoch AI (2025)peer reviewed
All Boats Rise with the Tide: A Manifesto for Building the Future Together
Life with AI (2026)internal
AI Chip Production and Global Compute Growth
Epoch AI (2025)peer reviewed
https://epoch.ai/data-insights/ai-chip-productionLeading ML Hardware Energy Efficiency Trends
Epoch AI (2025)peer reviewed
https://epoch.ai/data-insights/ml-hardware-energy-efficiencyEl Capitan Supercomputer Performance Specifications
Lawrence Livermore National Laboratory / AMD (2025)project data
Data Center Cooling State of Play 2025
Tom's Hardware (2025)industry report
https://www.tomshardware.com/pc-components/cooling/the-data-center-cooling-state-of-play-2025Immersion Cooling Market Analysis and Deployments
Data Center Dynamics (2026)industry report
Data Center Immersion Cooling Market Size Forecast 2035
Global Market Insights (2026)industry report
https://www.gminsights.com/industry-analysis/data-center-immersion-cooling-marketWaste Heat Recoveries in Data Centers: A Review
Renewable and Sustainable Energy Reviews (2023)peer reviewed
https://www.sciencedirect.com/science/article/pii/S1364032123006342Waste Heat Recovery from Data Centres
International Renewable Energy Agency (2024)government report
https://www.irena.org/Innovation-landscape-for-smart-electrification/Power-to-heat-and-cooling/31-Waste-heat-recovery-from-data-centresEnergy Efficiency Directive Recast - Data Center Requirements
European Union (2024)government report
Understanding the Ratings
KEDL 300: Preliminary
Quantified with engineering basis, suitable for feasibility
CL 2: Estimated
Order-of-magnitude based on analogous systems
Know something about AI & Compute Infrastructure?