Private equity poured $108B into data center deals in 2024 — triple the prior year — with global data center investment exceeding $61B in the first 11 months of 2025. Yet no technical due diligence framework exists for evaluating AI data center investments from an infrastructure engineering perspective.
Two emergent metrics have no authoritative champion: “tokens per watt” — the AI efficiency metric connecting inference economics to power infrastructure — and “resilience-adjusted TCO,” a concept nobody has formalized that factors downtime costs, failure recovery time, and checkpoint overhead into GPU cluster total cost of ownership. An estimated 30–65% of data center power capacity is stranded, creating a natural bridge between infrastructure engineering and investment thesis development.
URE covers infrastructure economics from the facility up: how data center design drives cost per token, how placement decisions shape inference latency, how sovereign AI economics differ from what the map suggests, and how the phases of technology adoption create both risk and opportunity for infrastructure operators.
A few years ago, I was having dinner with the Americas VP of a European energy supermajor — one of those companies that extracts oil from war zones, negotiates with regimes that don’t appear on polite lists, and operates in places where “political risk” means your assets might get nationalized or your personnel kidnapped.
Seventy-plus countries. Active operations in Libya, Nigeria, Angola, Myanmar, Yemen. The kinds of places where security briefings come before breakfast.
...
I see people everywhere anxious about whether AI will disrupt their jobs, their industries, their lives. I’ve always approached this with calm. Not indifference—calm.
The future rarely sends advance notice, but it is always arriving. This isn’t news. It’s the human condition.
A few years ago, I attended a keynote by Michio Kaku where he framed—perfectly, for me—the relationship between humanity and technological change. What follows is my version. I can’t claim novelty, and I’m not a domain expert in sociology or economics. I’m an infrastructure builder observing the same pattern from the inside.
...
Atlas is a single pane of glass for multi-cloud cost visibility. This post documents the pipeline: ingestion, streaming, storage, query, forecasting, and visualization.
Traditional internet architecture solved latency with caching. Static content, images, JavaScript bundles—all pushed to edge nodes milliseconds from users. CDNs achieve 95-99% cache hit rates. The compute stays centralized; the content moves to the edge.
AI breaks this model completely.
Every inference requires real GPU cycles. You can’t cache a conversation. You can’t pre-compute a response to a question that hasn’t been asked. The token that completes a sentence depends on every token before it.
...