Your AI Training Cluster Thirsty? Let's Talk Water.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
Sounds crazy. It's not.
Grid carbon intensity varies wildly throughout the day:
Same compute. Same results. 70% less carbon.
Grid carbon intensity changes dramatically based on:
Current grid breakdown in typical US markets:
That's a 3x difference in carbon intensity. Why waste it?
To build a carbon-aware scheduler, you need four components:
You're adding 6-12 hours of latency to training jobs. This isn't suitable for time-critical work. But for research training, batch jobs, and experimentation? Nobody needs results at 2 PM instead of 8 AM.
Production inference has to run when users request it. This is for training and batch processing only.
Works best in regions with variable renewable penetration. In hydro-dominated regions (like Quebec), the grid is already green 24/7.
It's often cheaper too.
Off-peak electricity rates in many markets are 30-50% lower than peak rates. By time-shifting to low-carbon periods, you're often also time-shifting to low-cost periods.
Free carbon reduction. Lower costs. Same results.
if carbon_intensity < threshold:
start_training()
elif time_until_deadline < max_wait:
start_training() # Can't wait forever
else:
queue_for_later(predicted_low_carbon_window)For a typical 10,000 GPU training run lasting 30 days:
Scale that across all training runs at a hyperscaler, and you're talking about meaningful impact.
Most training jobs don't have real time pressure. Research experiments, hyperparameter sweeps, model iterations - they can wait a few hours.
The best sustainability tech is the tech that makes green choices automatic. Carbon-aware scheduling is exactly that: set it once, save carbon forever.
If you're running AI training at scale and not considering carbon-aware scheduling, you're leaving money and carbon on the table.
The implementation is straightforward. The savings are real. The planet thanks you.
GreenCIO's Cost Prediction Agent includes carbon-aware scheduling recommendations. Request a demo to see how much carbon (and money) you could save.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
Traditional SaaS is too slow for energy markets. We pivoted to 'Autonomous Organization as a Service'—software that works while you sleep.
Giving an agent 30 tools costs $0.45 per run. We implemented a 'Code-First Skills' pattern to drop that to $0.003.
Grid interconnection is the #1 bottleneck for AI. Google X's Tapestry project is trying to virtualize the grid to fix it.
News tells you what happened yesterday. Markets tell you what will happen tomorrow. We built an agent to trade on the difference.
Starting August 2025, mandatory environmental reporting kicks in for AI models. Most CTOs are completely unprepared.
We forced our AI agents to fight. The 'Bull' vs. The 'Bear'. The result was better decisions than any single model could produce.
Installed capacity is a vanity metric. LCOE is the only number that levels the playing field between solar, gas, and nuclear.
We didn't want to pay for a Bloomberg terminal, so we wrote a 950-line TypeScript scraper that builds our own intelligence feed.