Infrastructure Bottlenecks Are Eclipsing Silicon as AI Scaling Constraints
While semiconductor export controls dominate geopolitical attention, memory supply and electrical power are emerging as the binding physical limits on AI deployment. Micron's projection of persistent shortages through 2030 signals that high-bandwidth memory availability now gates model training and inference capacity regardless of GPU access. Simultaneously, Google's willingness to fund complete power generation infrastructure—rather than wait for utility investment—reveals that electricity supply cannot keep pace with hyperscale buildout timelines. NTT's plan to double data centre capacity to 4 gigawatts occurs as operators accept they must act as quasi-utility providers to secure sites.
These infrastructure constraints compound differently across geographies. China's H200 licensing breakthrough matters less if memory shortages limit what can be built with those chips, while the Iran conflict's helium supply disruption threatens South Korean and Taiwanese fabs independent of U.S.-China technology competition. The result is a fracturing of advantage: access to frontier chips no longer guarantees AI capability if supporting infrastructure—memory, power, cooling, rare earths—cannot scale in parallel. This favours actors with integrated control over full technology stacks and supply chains rather than those optimising single components.