Your AI Training Cluster Thirsty? Let's Talk Water.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
We were spending $0.45 per request just on tool definitions.
Not on actual AI work. Just telling the model what tools exist.
With 30 tools in context, you're burning 150,000 tokens before the conversation even starts. At $3/million input tokens for Claude, that adds up fast.
So we rewrote everything.
Instead of loading tool schemas into the context window, we:
| Metric | Before | After | Reduction |
|---|---|---|---|
| 30 tools token count | 84,000 | 600 | 99.3% |
| Cost per request | $0.45 | $0.003 | 99.3% |
Treat skills like files on disk, not system prompts.
Agents discover what they need. They don't carry everything everywhere.
Think about how you work: you don't read the entire manual before starting a task. You look up the specific section when you need it.
Each skill has a skill.md file with YAML frontmatter. At build time, we extract just the metadata:
{
"lcoe-calculator": {
"name": "LCOE Calculator",
"trigger": "lcoe|levelized cost|energy economics",
"agents": ["INVESTMENT_INTELLIGENCE", "COST_PREDICTION"]
}
}That's it. ~50 bytes per skill instead of 5,000.
When a query comes in, we do a simple regex match against triggers. If we hit, we load the full skill instructions.
Instead of 30 different tool schemas, we have one:
execute_skill(skill_id: string, script: string, input: object)The agent knows how to call this. The skill instructions tell it what to put in the input.
skills/
└── lcoe-calculator/
├── skill.md # Entry point with metadata
├── scripts/ # Executable Python scripts
│ ├── calculate_lcoe.py
│ └── sensitivity.py
├── schemas/ # JSON schemas for validation
│ ├── input.schema.json
│ └── output.schema.json
└── examples/ # Sample inputs/outputsWe based this on Anthropic's "Code Execution with MCP" pattern but pushed it further. Their insight: agents should write and execute code, not just call predefined tools.
Our extension: predefine the code patterns as skills, so agents get the best of both worlds - reliable, tested scripts with the flexibility of dynamic invocation.
If you're building multi-agent systems and hitting token limits or cost walls, look at your tool definitions first. That's probably where the waste is.
The solution isn't fewer tools. It's smarter tool loading.
Full implementation details in our GitHub. Happy to share if anyone's hitting the same wall.
We ran the numbers: A 10k H100 cluster can consume 2 million gallons of water a month. Here is the math and the engineering fix.
Traditional SaaS is too slow for energy markets. We pivoted to 'Autonomous Organization as a Service'—software that works while you sleep.
Grid interconnection is the #1 bottleneck for AI. Google X's Tapestry project is trying to virtualize the grid to fix it.
News tells you what happened yesterday. Markets tell you what will happen tomorrow. We built an agent to trade on the difference.
Starting August 2025, mandatory environmental reporting kicks in for AI models. Most CTOs are completely unprepared.
We forced our AI agents to fight. The 'Bull' vs. The 'Bear'. The result was better decisions than any single model could produce.
Installed capacity is a vanity metric. LCOE is the only number that levels the playing field between solar, gas, and nuclear.
Grid carbon intensity varies by 3x throughout the day. We built a scheduler that pauses AI training when the grid is dirty.
We didn't want to pay for a Bloomberg terminal, so we wrote a 950-line TypeScript scraper that builds our own intelligence feed.