The AI Data Centers That Fit on a Truck
What Happened
A traditional data center protects the expensive hardware inside it with a “shell” constructed from steel and concrete. Constructing a data center’s shell is inexpensive compared to the cost of the hardware and infrastructure inside it, but it’s not trivial. It takes time for engineers to consider p
Our Take
Modular AI data centers — containerized GPU clusters on flatbed trucks — are moving from concept to production. Microsoft and CoreWeave are delivering pre-configured H100 racks that go live in days, no construction permits required.
For edge inference — factory floor vision, real-time video — latency to cloud GPUs is the real constraint. These units cut round-trip inference from 80ms to under 5ms. Defaulting to cloud APIs for everything is the wrong call when data can't legally leave the facility.
Teams running regulated inference in healthcare or defense should spec this now. Standard SaaS with no compliance requirements: ignore it.
What To Do
Spec a truck-deployable H100 cluster instead of routing regulated inference through cloud APIs because data residency laws will block your deployment before latency does.
Builder's Brief
What Skeptics Say
Modular portable data centers hit hard thermal density and power delivery ceilings that make them uneconomical for sustained high-throughput inference; the use case is narrower than the hype and doesn't threaten hyperscaler economics.
Cited By
React
Get the weekly AI digest
The stories that matter, with a builder's perspective. Every Thursday.