Skip to content
Published on

AI's Power Crisis: Why Data Centers Need Nuclear Plants (The Numbers Don't Lie)

Authors

Introduction

2024 was the year the AI industry hit a very real wall: power. A single ChatGPT query consumes as much electricity as 10 Google searches, and NVIDIA's latest GPU draws as much power as a household air conditioner. Big Tech companies began racing to sign contracts with nuclear power plants.

In this article, we examine the scale of the power crisis that AI has created through hard numbers, explore Big Tech's nuclear rush, the water crisis, the cooling revolution, and the energy literacy that every developer should have.


1. The AI Power Crisis in Numbers

Even before the AI boom, data centers were already massive power consumers. But after the emergence of generative AI, the growth curve changed completely.

YearGlobal DC Power (TWh)US DC Power (TWh)AI Server Power (TWh)
202441518393
2025506228143
2026600276198
2028775355320
2030980426432

Here are the key figures:

  • Global data centers: 415TWh (2024) to 980TWh (2030), a 2.4x increase
  • US data centers: 183TWh (2024) to 426TWh (2030), 133% growth
  • AI server power: 93TWh (2025) to 432TWh (2030), roughly 5x increase
  • AI-optimized server share: 21% of total DC power (2025) to 44% (2030)

What These Numbers Mean

It can be hard to grasp how massive 980TWh really is. Let's put it in perspective:

  • South Korea's total annual power consumption: approximately 550TWh
  • Japan's total annual power consumption: approximately 900TWh
  • France's total annual power consumption: approximately 450TWh

In other words, by 2030 global data center power consumption will exceed Japan's entire electricity consumption. A single industry sector surpassing a nation's total power is unprecedented in history.

America's Power Supply Crisis

The US faces a particularly serious situation:

  • Data center power will grow from 6% (2024) to 12% (2030) of total US electricity
  • Northern Virginia (Loudoun County): World's largest DC cluster, already hitting grid capacity limits
  • Texas: DC construction rush driving electricity prices up
  • Georgia: Power regulators considering restrictions on new DC grid connections

According to Goldman Sachs, the US will need 47GW of new generation capacity by 2030, equivalent to 47 nuclear plants.


2. The Power Story of a Single GPU

NVIDIA GPU Power Consumption by Generation

At the center of the AI power crisis sits the GPU. Let's look at how much power NVIDIA's latest GPUs consume.

GPU ModelTDP (Watts)Release YearGeneration
A100400W2020Ampere
H100700W2023Hopper
B2001,000W2024Blackwell
B3001,400W2025Blackwell Ultra
GB200 NVL72 (rack)120kW2024Blackwell

In just four years, a single GPU's power consumption has increased from 400W to 1,400W, a 3.5x jump.

DGX B200 System Power Scale

The NVIDIA DGX B200 is a server containing 8 B200 GPUs. A single unit consumes approximately 14.3kW.

To put this in everyday terms:

  • Equivalent to running about 10 household air conditioners simultaneously
  • Roughly equal to the total power consumption of 5 average homes
  • Enough power to slow-charge 2 electric vehicles per hour

xAI Colossus: The World's Largest AI Cluster

Elon Musk's xAI built the Colossus cluster in Memphis, representing the extreme end of AI power consumption.

  • Phase 1: 100,000 H100 GPUs, approximately 150MW
  • Phase 2: Expanded to 200,000 H100 GPUs, approximately 300MW
  • Ultimate target: 1GW+ (equivalent to one nuclear plant)
  • Initially powered by gas turbines for self-generated electricity, sparking environmental controversy

Training vs Inference Power Comparison

AI power consumption breaks down into two phases.

Training

  • GPT-4 training: approximately 50GWh (estimated) = annual power for 5,000 average US homes
  • Training happens once, but as models grow larger, training power increases exponentially
  • Llama 3 405B training: 16,384 H100 GPUs running for 54 days

Inference

  • Each individual query uses little power, but billions are processed 24/7 worldwide
  • As of 2025, approximately 60% of AI power goes to inference
  • One ChatGPT query: approximately 0.01kWh (roughly 10x a Google search)
  • Global daily ChatGPT queries: over 100 million, meaning 1GWh+ per day

GPU Power Consumption Calculator

As a developer, you should be able to estimate the power consumption of your AI workloads.

# GPU power consumption calculator
def calculate_gpu_power(
    num_gpus: int,
    gpu_tdp_watts: int,
    utilization: float,  # 0.0 to 1.0
    hours_per_day: float,
    pue: float = 1.3,  # Power Usage Effectiveness
    days: int = 365
) -> dict:
    """
    Calculate GPU cluster power consumption

    Parameters:
        num_gpus: Number of GPUs
        gpu_tdp_watts: TDP per GPU (watts)
        utilization: Average utilization rate (0.0 to 1.0)
        hours_per_day: Daily operating hours
        pue: Data center PUE (includes cooling/infrastructure overhead)
        days: Annual operating days
    """
    # IT equipment power (kW)
    it_power_kw = (num_gpus * gpu_tdp_watts * utilization) / 1000

    # Total DC power (with PUE)
    total_power_kw = it_power_kw * pue

    # Daily energy consumption (kWh)
    daily_kwh = total_power_kw * hours_per_day

    # Annual energy consumption (MWh)
    annual_mwh = daily_kwh * days / 1000

    # Annual cost at US average rate (~$0.10/kWh)
    annual_cost_usd = daily_kwh * days * 0.10

    return {
        "IT Power (kW)": round(it_power_kw, 1),
        "Total Power (kW, with PUE)": round(total_power_kw, 1),
        "Daily Consumption (kWh)": round(daily_kwh, 1),
        "Annual Consumption (MWh)": round(annual_mwh, 1),
        "Annual Electricity Cost (USD)": round(annual_cost_usd, 2),
        "Equivalent US Homes": round(annual_mwh * 1000 / 10500),  # US avg ~10,500kWh/year
    }


# Example 1: 1,000 H100 training cluster
training_cluster = calculate_gpu_power(
    num_gpus=1000,
    gpu_tdp_watts=700,
    utilization=0.85,
    hours_per_day=24,
    pue=1.3
)
print("=== 1,000x H100 Training Cluster ===")
for key, value in training_cluster.items():
    print(f"  {key}: {value}")

# Example 2: 10,000 B200 inference cluster (50% utilization)
inference_cluster = calculate_gpu_power(
    num_gpus=10000,
    gpu_tdp_watts=1000,
    utilization=0.5,
    hours_per_day=24,
    pue=1.2
)
print("\n=== 10,000x B200 Inference Cluster ===")
for key, value in inference_cluster.items():
    print(f"  {key}: {value}")

Sample output:

=== 1,000x H100 Training Cluster ===
  IT Power (kW): 595.0
  Total Power (kW, with PUE): 773.5
  Daily Consumption (kWh): 18564.0
  Annual Consumption (MWh): 6775.9
  Annual Electricity Cost (USD): 677586.0
  Equivalent US Homes: 645

=== 10,000x B200 Inference Cluster ===
  IT Power (kW): 5000.0
  Total Power (kW, with PUE): 6000.0
  Daily Consumption (kWh): 144000.0
  Annual Consumption (MWh): 52560.0
  Annual Electricity Cost (USD): 5256000.0
  Equivalent US Homes: 5006

3. Big Tech's Nuclear Rush

Why Nuclear?

The reasons Big Tech companies are suddenly turning to nuclear power are clear.

CriteriaNuclearSolarWindNatural Gas
Capacity Factor93%25%35%87%
Carbon EmissionsZeroZeroZeroHigh
Land Area (1GW)1 km240 km2100 km22 km2
24/7 ReliabilityVery HighIntermittentIntermittentHigh
Baseload SuitabilityOptimalUnsuitableUnsuitablePossible

Data centers require stable power 365 days a year, 24 hours a day. Solar and wind depend on weather, making them unsuitable as baseload power sources. Nuclear is the only large-scale power source that is both zero-carbon and capable of 24/7 operation.

Microsoft: Three Mile Island Restart ($16B)

Microsoft's nuclear project carries enormous symbolic weight.

  • Target: Three Mile Island Unit 1 (TMI-1)
    • The 1979 accident occurred at Unit 2; Unit 1 is a separate reactor
    • Shut down in 2019 for economic reasons
  • Capacity: 835MW (enough to power about 800,000 homes)
  • Investment: Approximately $16 billion
  • Restart target: 2028
  • Contract: 20-year exclusive power supply to Microsoft
  • Significance: First nuclear plant restart in US history

Operated by Constellation Energy, the facility has been renamed the "Crane Clean Energy Center." Microsoft plans to use this power for Azure data centers.

Amazon: Susquehanna Nuclear Campus ($20B+)

Amazon is pursuing an even more aggressive nuclear strategy.

  • Susquehanna Nuclear Plant (Pennsylvania): 960MW power purchase agreement
    • Direct supply contract with Talen Energy for data center use
    • 960MW data center campus being built adjacent to the nuclear plant
  • Additional investment: Over $20 billion total
  • SMR investments: Invested in Energy Northwest (Washington state) SMR project
  • X-energy: $500M investment in SMR developer
  • Strategy: Dual approach combining existing nuclear + next-gen SMRs

Google/Kairos Power: First Corporate SMR Deal in the US

Google is focusing on next-generation nuclear technology with SMRs.

  • Partner: Kairos Power (molten salt-cooled SMR developer)
  • Capacity: 500MW (completion target: 2030s)
  • Significance: First corporate SMR power purchase agreement (PPA) in the US
  • Technology: Fluoride salt-cooled reactor (uses TRISO fuel)
    • Higher safety than conventional light-water reactors
    • Atmospheric pressure operation eliminates explosion risk
  • Phased construction: First reactor by 2030, with additional units following sequentially

Meta: Large-Scale New Nuclear RFP

Meta (Facebook) has announced the most ambitious nuclear plan.

  • Scale: 1-4GW of new nuclear generation capacity
  • Approach: Issued an RFP (Request for Proposals) for new nuclear construction
  • Target timeline: Early 2030s
  • Key distinction: Pursuing entirely new construction, not purchasing existing plants
  • Driver: Surging power demand from Meta's AI training infrastructure expansion

Big Tech Nuclear Investment Summary

CompanyProjectCapacityInvestmentTimeline
MicrosoftTMI-1 Restart835MW$16B2028
AmazonSusquehanna + SMR960MW+$20B+2025-2030
GoogleKairos SMR500MWUndisclosed2030+
MetaNew Nuclear RFP1-4GWUndisclosed2030+
Oracle3 SMR Plan1GW+Undisclosed2030+

Combined: Big Tech is looking to secure more than 10GW of new nuclear capacity, equivalent to over 10 large nuclear plants.


4. The Water Crisis: AI's Hidden Cost

AI's Water Consumption

Power is not the only resource problem for AI. Data center cooling requires enormous amounts of water.

  • AI-related water usage: 312.5-764.6 billion liters per year (estimated)
  • This is comparable to global bottled water consumption
  • US data centers alone consume approximately 66 billion liters annually

GPT-4 Training Water Footprint

The water consumed by a single GPT-4 training run is staggering.

  • GPT-4 training: Approximately 700,000 liters of water
  • This is roughly 0.3 Olympic swimming pools
  • Caused by evaporative cooling systems needed to dissipate training heat

Everyday AI Usage Water Costs

The AI services we use daily also consume water.

  • 25-50 ChatGPT conversations: Approximately one 500ml water bottle
  • Image generation AI (DALL-E, Midjourney): About 3.3 liters per image
  • AI code generation (Copilot): About 0.01 liters per code suggestion

Data Centers in Water-Stressed Regions

The problem is that many data centers are located in regions already facing water scarcity.

  • Western US: Large-scale DC clusters in desert areas like Arizona and Nevada
  • Chile: Local residents protesting Google DC construction
  • Uruguay: Google DC project raising regional water shortage concerns
  • Saudi Arabia/UAE: Expanding AI investment vs. severe water scarcity

According to WRI (World Resources Institute), approximately 30% of global data centers are located in high water-stress regions.


5. The Cooling Revolution

The Limits of Air Cooling

Traditional data center cooling used air-based systems similar to air conditioning. But as GPU heat output has surged, the limitations of air cooling have become apparent.

  • A100 era: About 10-15kW per server rack, air cooling was sufficient
  • H100 era: About 40-70kW per rack, air cooling reaching its limits
  • B200/B300 era: Over 100kW per rack, air cooling is impossible

NVIDIA has effectively made liquid cooling mandatory starting with the Blackwell architecture (B200/B300). The GB200 NVL72 rack is designed exclusively for liquid cooling.

Types of Liquid Cooling Technology

There are three main liquid cooling technologies currently used in data centers.

1. Direct-to-Chip (DTC) Liquid Cooling

  • Cold plate method where coolant directly contacts the GPU/CPU chip
  • Most common and efficient approach
  • AWS: Achieved 46% cooling energy reduction with DTC liquid cooling
  • Can be retrofitted to existing data centers

2. Immersion Cooling

  • Entire server submerged in non-conductive coolant
  • Both single-phase and two-phase variants exist
  • Highest cooling efficiency but complex maintenance
  • Microsoft experimenting with deployment

3. Rear-Door Heat Exchanger

  • Water-circulating heat exchanger installed on the back door of server racks
  • Can be added to existing air-cooling infrastructure
  • Suitable for mid-level heat dissipation

Cooling Technology Comparison

TechnologyCooling EfficiencyInstallation CostMaintenanceSuitable Workload
Air CoolingLowLowEasyGeneral servers
Rear-DoorMediumMediumModerateMixed workloads
DTC LiquidHighMedium-HighModerateAI/HPC
ImmersionHighestHighComplexUltra-dense AI

Microsoft's Liquid Cooling Results

Microsoft deployed liquid cooling at scale across Azure data centers with significant results.

  • Azure data center carbon emissions reduced by 12%
  • PUE improved from 1.3 to 1.12 (approaching the ideal of 1.0)
  • Water consumption also decreased (compared to evaporative cooling)
  • Plan to standardize liquid cooling in all new DCs by 2026

The Rise of Zero-Water Data Centers

In response to the water crisis, data center designs that use no water at all are emerging.

  • Microsoft: Declared "Water Positive by 2030"
    • Goal to replenish more water than consumed
  • Meta: Researching waste heat recovery systems that use no water
  • Nordic DC Model: Natural cooling in cold regions like Finland and Sweden
    • Achieving zero water usage by cooling with outside air
    • Meta's Lulea (Sweden) DC is the leading example

6. The Sustainability Dilemma

AI's Carbon Footprint

The AI industry's carbon emissions are growing rapidly.

  • AI-related carbon emissions: 32.6-79.7 million tons of CO2 per year (estimated)
  • This is equivalent to the total emissions of mid-sized countries like Belgium or Czech Republic
  • One ChatGPT query: About 4.32g CO2 (roughly 6-10x a Google search)
  • One GPT-4 training run: Approximately 12,500 tons of CO2

Big Tech's Net Zero Goals vs Reality

Big Tech companies have pledged carbon neutrality, but the gap between goals and reality is widening due to AI demand growth.

Google

  • Goal: Net zero by 2030
  • Reality: 2023 carbon emissions 48% higher than 2019
  • Cause: AI training/inference infrastructure expansion

Microsoft

  • Goal: Carbon negative by 2030
  • Reality: 2023 emissions 29% higher than 2020
  • Cause: Explosive Azure AI service demand

Amazon

  • Goal: Net zero by 2040 (Climate Pledge)
  • Reality: Emissions rising with AWS expansion
  • Response: Maintaining position as world's largest renewable energy buyer

Renewable Energy PPA Landscape

Big Tech is signing massive renewable energy Power Purchase Agreements (PPAs) to meet carbon neutrality goals.

  • Data center industry total: Over 27GW of clean energy PPAs signed
  • Amazon: World's largest single corporate renewable energy buyer (25GW+)
  • Microsoft: 10GW+ renewable energy PPAs
  • Google: 7GW+ renewable energy PPAs + nuclear contracts

Efficiency Improvements vs Demand Growth

In the AI industry, the speed of energy efficiency improvements and the speed of demand growth are in constant competition.

Efficiency improvement factors:

  • Generational GPU performance-per-watt gains (H100 to B200: 4x training efficiency)
  • Quantization reducing model size and power
  • Inference optimization technologies (vLLM, TensorRT-LLM, etc.)
  • PUE improvements (1.5 to 1.1)

Demand growth factors:

  • Exponential growth in AI users
  • Continuous expansion of model sizes (scaling laws)
  • New AI use cases (agents, multimodal, etc.)
  • More devices shipping with built-in AI

Based on current trends, demand growth is outpacing efficiency improvements. This is precisely why Big Tech is turning to nuclear as a fundamental solution.


7. South Korea and Japan's AI Power Situation

South Korea: Surging Data Center Power Demand

South Korea faces rapidly growing AI data center power demand.

Current situation:

  • Domestic DC power demand: About 4GW (2024), projected to reach 8GW (2030)
  • Expected to grow from about 5% to over 10% of total electricity
  • Seoul metropolitan area (Pangyo, Anyang, Goyang) DC clusters: Hitting grid capacity limits
  • Naver, Kakao, KT, SK competing to expand AI data centers

South Korea's nuclear status:

  • World's 5th largest nuclear operator (25 active reactors)
  • Total nuclear capacity: About 25.8GW (approximately 30% of generation)
  • Shin-Hanul Units 3 and 4 construction resumed
  • APR1400: Korean reactor design exported globally (UAE Barakah)

AI power response:

  • KEPCO: Considering dedicated data center electricity rates
  • Government: Announced special AI infrastructure power supply plan
  • SK hynix/Samsung: Researching AI semiconductor power efficiency improvements
  • KHNP: Pursuing SMR development for data center power supply

Japan: Post-Fukushima Nuclear Restarts Meet AI Demand

Japan faces a unique situation. After shutting down most nuclear plants following the 2011 Fukushima disaster, restarts are accelerating, driven partly by AI demand.

Current situation:

  • Pre-Fukushima: 54 nuclear reactors operating (30% of total generation)
  • Post-Fukushima: Nearly all shut down
  • As of 2024: 12 reactors restarted, more restarts being pursued
  • AI data center power demand: Growing rapidly

Where AI meets nuclear:

  • SoftBank/NVIDIA: Plans for AI supercomputer construction in Japan (thousands of GPUs)
  • Microsoft: Announced $2.9 billion investment in Japan AI infrastructure
  • Amazon: Expanding Tokyo/Osaka regions
  • NTT/KDDI: Expanding proprietary AI data center construction

Energy policy shifts:

  • Japanese government: Targeting 20-22% nuclear share (2030)
  • Pursuing next-generation innovative reactor development
  • Mixed strategy of renewables + nuclear
  • Expanding power infrastructure investment to attract data centers

South Korea vs Japan Comparison

CategorySouth KoreaJapan
Active Reactors2512 (restarting)
Nuclear Share (Generation)~30%~7% (target: 20%)
DC Power Demand Growth15-20% annually12-18% annually
AI Semiconductor StrengthMemory (HBM) world No. 1Equipment/materials
SMR DevelopmentKHNP i-SMRMitsubishi/Hitachi

8. Power Literacy for Developers

Choosing a Model = Choosing Power Consumption

The AI model a developer selects directly determines power consumption.

# Model inference power comparison (rough estimates)
model_power_comparison = {
    "GPT-4 (API)": {
        "params": "~1.8T (estimated)",
        "power_per_query_wh": 0.01,  # ~10Wh = 0.01kWh
        "latency_ms": 2000,
        "quality": "Best"
    },
    "GPT-3.5 (API)": {
        "params": "175B",
        "power_per_query_wh": 0.002,
        "latency_ms": 500,
        "quality": "Good"
    },
    "Llama 3 8B (local)": {
        "params": "8B",
        "power_per_query_wh": 0.0005,
        "latency_ms": 200,
        "quality": "Fair"
    },
    "Phi-3 Mini (edge)": {
        "params": "3.8B",
        "power_per_query_wh": 0.0001,
        "latency_ms": 100,
        "quality": "Basic"
    },
}

# Annual power comparison for 100K daily queries
daily_queries = 100_000

print("=== Annual Power for 100K Daily Queries ===\n")
for model, specs in model_power_comparison.items():
    annual_kwh = specs["power_per_query_wh"] * daily_queries * 365
    annual_cost_usd = annual_kwh * 0.10  # US average electricity rate
    print(f"{model}:")
    print(f"  Parameters: {specs['params']}")
    print(f"  Power per query: {specs['power_per_query_wh']} kWh")
    print(f"  Annual power: {annual_kwh:,.0f} kWh")
    print(f"  Annual cost: ${annual_cost_usd:,.0f}")
    print(f"  Quality: {specs['quality']}")
    print()

Key takeaway: Not every task requires the largest model. Selecting an appropriately sized model for the task benefits both your budget and the environment.

Inference Optimization = Cost + Environmental Optimization

Optimization at the inference stage directly reduces power consumption.

# Power reduction by inference optimization technique
optimization_techniques = {
    "Baseline (no optimization)": {
        "throughput_multiplier": 1.0,
        "power_reduction": 0,
        "description": "Default PyTorch inference"
    },
    "TensorRT-LLM": {
        "throughput_multiplier": 2.5,
        "power_reduction": 0.30,
        "description": "NVIDIA optimized inference engine"
    },
    "vLLM (PagedAttention)": {
        "throughput_multiplier": 2.0,
        "power_reduction": 0.25,
        "description": "Efficient memory management for higher throughput"
    },
    "INT8 Quantization": {
        "throughput_multiplier": 1.8,
        "power_reduction": 0.35,
        "description": "FP16 -> INT8 reduces compute/memory"
    },
    "INT4 Quantization (GPTQ/AWQ)": {
        "throughput_multiplier": 2.5,
        "power_reduction": 0.50,
        "description": "Aggressive quantization for maximum savings"
    },
    "Knowledge Distillation": {
        "throughput_multiplier": 3.0,
        "power_reduction": 0.60,
        "description": "Large model -> small model knowledge transfer"
    },
    "Speculative Decoding": {
        "throughput_multiplier": 2.0,
        "power_reduction": 0.20,
        "description": "Draft model generates quickly, main model verifies"
    },
}

base_power_kwh = 100_000  # Baseline annual power (kWh)
electricity_rate = 0.10  # USD/kWh

print("=== Power/Cost Savings by Optimization Technique ===")
print(f"Baseline: {base_power_kwh:,} kWh/year\n")

for technique, specs in optimization_techniques.items():
    saved_kwh = base_power_kwh * specs["power_reduction"]
    saved_cost = saved_kwh * electricity_rate
    co2_saved = saved_kwh * 0.4  # kg CO2 per kWh (US average)
    print(f"{technique}:")
    print(f"  Throughput multiplier: {specs['throughput_multiplier']}x")
    print(f"  Power reduction: {specs['power_reduction']*100:.0f}%")
    print(f"  Annual savings: {saved_kwh:,.0f} kWh (${saved_cost:,.0f})")
    print(f"  CO2 saved: {co2_saved:,.0f} kg")
    print(f"  Description: {specs['description']}")
    print()

How to Cut Power with Quantization and Distillation

Here are practical power-saving techniques you can apply today.

Quantization Practical Guide:

# GPTQ quantization example (AutoGPTQ library)
from auto_gptq import AutoGPTQForCausalLM
from transformers import AutoTokenizer

model_name = "meta-llama/Llama-3-8B"
quantized_model_name = "TheBloke/Llama-3-8B-GPTQ"

# Load quantized model (INT4)
tokenizer = AutoTokenizer.from_pretrained(quantized_model_name)
model = AutoGPTQForCausalLM.from_quantized(
    quantized_model_name,
    device="cuda:0",
    use_safetensors=True,
)

# Memory usage comparison
# FP16 original: ~16GB VRAM
# INT4 GPTQ: ~4GB VRAM (75% reduction)
# Power consumption: ~50% reduction (can use smaller GPU)

Knowledge Distillation Overview:

# Knowledge distillation concept (pseudocode)
# Teacher model: Llama 3 70B (large, high-quality, high-power)
# Student model: Custom 7B (small, specialized, low-power)

# 1. Generate large synthetic dataset using teacher model
# 2. Train student model on synthetic data
# 3. Achieve 90%+ of teacher performance on specific tasks
# 4. Inference power is 10-20% of teacher

# Benefits:
# - 80-90% inference cost reduction
# - 3-5x latency improvement
# - Massive CO2 emission reduction
# - Deployable on edge devices

The Green AI Movement

As awareness of AI's environmental impact grows, the "Green AI" movement is gaining momentum.

Core principles:

  1. Efficiency first: Don't spend 10x power for a 0.1% accuracy improvement
  2. Transparent reporting: Disclose training power and carbon emissions in papers and model releases
  3. Right-sized models: Use the smallest model that meets the task requirements
  4. Inference optimization: Apply quantization, pruning, and distillation before deployment
  5. Infrastructure choices: Select cloud regions powered by clean energy

Practical steps:

  • Calculate training carbon emissions with the ML CO2 Impact tool
  • Prioritize models with Hugging Face Energy Star badges
  • Prototype with small models first, scale up only when needed
  • Deploy inference servers in regions with high renewable energy percentages
# Track training carbon emissions with codecarbon
# pip install codecarbon

from codecarbon import EmissionsTracker

tracker = EmissionsTracker(
    project_name="my_ai_project",
    measure_power_secs=30,
    tracking_mode="process",
)

tracker.start()

# ... AI training or inference code ...
# model.train()
# model.predict()

emissions = tracker.stop()

print(f"Power consumed: {tracker.final_emissions_data.energy_consumed:.4f} kWh")
print(f"CO2 emitted: {tracker.final_emissions_data.emissions:.4f} kg")
print(f"Duration: {tracker.final_emissions_data.duration:.0f} seconds")

Quiz

Q1. Data Center Power Scale

What is the projected global data center power consumption for 2030?

Answer: Approximately 980TWh

From 415TWh in 2024 to approximately 980TWh in 2030, a roughly 2.4x increase. This exceeds Japan's total annual power consumption (approximately 900TWh). AI servers are projected to account for about 44% of this total.

Q2. GPU Power Consumption

What is the TDP (Thermal Design Power) of a single NVIDIA B300 GPU?

Answer: 1,400W

The B300 is a Blackwell Ultra generation GPU with a TDP of 1,400W. This is 3.5 times the 2020 A100 (400W). A single DGX B200 system (8 GPUs) consumes about 14.3kW, equivalent to roughly 10 household air conditioners.

Q3. Big Tech Nuclear Investment

What are the capacity and investment scale for Microsoft's Three Mile Island Unit 1 restart?

Answer: 835MW, approximately $16 billion

Microsoft is restarting Unit 1, not Unit 2 where the 1979 accident occurred. Unit 1 was shut down in 2019 for economic reasons and targets a 2028 restart. This is the first nuclear plant restart in US history, operated by Constellation Energy.

Q4. AI Water Usage

How much water does approximately 25-50 ChatGPT conversations consume?

Answer: About 500ml (one water bottle)

This water is consumed by data center evaporative cooling systems. A single GPT-4 training run uses about 700,000 liters, and the AI industry as a whole consumes 312.5-764.6 billion liters per year, comparable to global bottled water consumption.

Q5. Developer Power Reduction

What power reduction can you expect from applying INT4 quantization (GPTQ/AWQ)?

Answer: Approximately 50%

INT4 quantization converts FP16 models to 4-bit integers, reducing memory usage by about 75% and power consumption by about 50%. Throughput improves by approximately 2.5x. However, some quality loss is possible, so task-specific benchmarking is recommended.


References

  1. IEA (International Energy Agency) - "Electricity 2024: Analysis and Forecast to 2026" - Global data center power consumption projections
  2. Goldman Sachs - "AI, Data Centers, and the Coming US Power Demand Surge" (2024) - 47GW US power demand analysis
  3. NVIDIA - Blackwell Architecture Technical Brief - B200/B300 GPU power specifications
  4. Constellation Energy - Three Mile Island Unit 1 restart official announcement
  5. Amazon - Susquehanna Nuclear Data Center Campus project announcement
  6. Google/Kairos Power - SMR Power Purchase Agreement (PPA) official announcement
  7. Meta - Nuclear Energy RFP official announcement (2024)
  8. Shaolei Ren et al. - "Making AI Less Thirsty" (2024) - AI water consumption study (University of California, Riverside)
  9. EPRI (Electric Power Research Institute) - "Powering Intelligence" (2024) - Comprehensive data center power demand report
  10. Uptime Institute - Global Data Center Survey 2024 - PUE and cooling technology trends
  11. AWS - Direct-to-Chip Liquid Cooling technical whitepaper - 46% cooling energy reduction
  12. WRI (World Resources Institute) - Global water stress map and data center location analysis
  13. codecarbon - ML training carbon emission tracking library documentation
  14. Hugging Face - Energy Efficiency Leaderboard - Model energy efficiency comparison
  15. KEPCO (Korea Electric Power Corporation) - Domestic data center power demand report
  16. METI (Japan Ministry of Economy, Trade and Industry) - Basic Energy Plan and nuclear restart status
  17. xAI - Colossus Memphis Supercomputer official announcement