Memory and Power Emerge as Near-Term AI Infrastructure Limits
Memory pricing dynamics are forcing a fundamental recalculation of AI infrastructure economics. DRAM prices increased approximately 53% in 2024 and continue rising in 2026, with server vendors now issuing quote estimates rather than firm prices due to unavailability. Microsoft is publicly addressing Windows 11 memory usage concerns, and Sony has suspended orders for compact flash and SD cards entirely due to unavailable memory chips. The crunch reflects a structural mismatch: AI training and inference workloads require exponentially more high-bandwidth memory and DRAM than traditional computing, but semiconductor fabrication capacity takes 18-24 months to scale. Google researchers' TurboQuant technique for reducing AI memory usage has not prevented memory-maker share prices from declining, suggesting markets expect demand destruction from high prices before new fab capacity comes online.
This creates immediate strategic implications beyond simple cost pressure. Mistral AI's $830 million debt financing for a Paris data centre represents a bet that owning compute infrastructure outweighs cloud rental flexibility, but the company must now manage power procurement, cooling, and hardware refresh cycles while competing on model development. The willingness of lenders to finance AI-specific data centres at this scale indicates credit markets view compute capacity itself as valuable collateral, independent of any single model's success. Meanwhile, the shift from Moore's Law economics to customisation economics — with further process shrinks below 2nm delivering better performance per watt but becoming exponentially harder and more expensive — advantages companies with chip design expertise and long-term volume commitments rather than those waiting for the next node to automatically solve constraints.