Is your AI training cluster thirsty? Let's talk water.
A practical look at AI cooling water demand, where the risk concentrates, and how teams can mitigate it.
"Add this article to our feed."
npm run url-news "https://techcrunch.com/article-url"That's it. Article extracted, analyzed, tagged, and added.
We initially tried two extremes:
The solution was a hybrid architecture.
Before we waste a single token, we process the content locally.
We built a robust extractor that runs right on the machine. It handles:
og:title, authors, dates, and site names using standard meta tags.Once we have clean, high-signal text, then we bring in the heavy guns. We pass the cleaned JSON to an LLM (Claude or OpenAI) for the tasks that actually require intelligence:
1. Cost Control
By cleaning the HTML locally, we reduce the token count by 60–80% before the API call. We pay to process information, not <div> tags.
2. Speed
Local relevance scoring means we can discard low-value URLs in milliseconds without network latency.
3. Reliability
If the LLM is down or hallucinates, we still have the locally extracted title, date, and raw content. The system degrades gracefully.
We use a "waterfall" strategy. Try the most reliable method (Open Graph tags); if missing, fall back to semantic HTML (<article>); if missing, use heuristics (largest text block). This ensures we get something usable from almost any site.
We maintain a weighted dictionary of domain-specific terms ("PPA", "interconnection", "H100"). An article must cross a point threshold to be considered "intelligence." This simple filter saves us from filling our database with generic tech news.
You don't need AI for everything.
The most effective AI systems are often 20% AI and 80% solid engineering. By letting code do what code does best (scraping, filtering, formatting), we free up the AI to do what it does best (reasoning and synthesis).
See our Intelligence Feed in action at /intelligence-feed. The source code for our extractor is available in our repository.
A practical look at AI cooling water demand, where the risk concentrates, and how teams can mitigate it.
Why we moved from traditional SaaS patterns to a multi-agent operating model for infrastructure intelligence.
How code-first skills and tighter context routing drove major cost reductions without quality loss.
Why grid-visibility tooling may become the limiting factor for AI data center expansion.
Where market-implied probabilities beat headlines for timing-sensitive energy and infrastructure decisions.
What the EU AI Act means for AI energy reporting, compliance timelines, and exposure management.
How structured disagreement between specialist agents produced better portfolio decisions.
Why LCOE remains a core metric for comparing technologies and underwriting long-horizon energy risk.
How carbon-aware workload scheduling reduces both emissions and compute cost volatility.
A portfolio-level briefing on grid constraints, power costs, and capital-allocation implications.
Who is funding hyperscale buildout, where structures are changing, and what risk shifts to lenders.
A practical playbook for lowering AI energy intensity without sacrificing delivery speed.