Compute & Infrastructure
Top Line
Microsoft announced $10 billion in AI infrastructure investment in Japan and $5.5 billion in Singapore, part of a broader Asia-Pacific buildout to capture regional AI demand amid rising geopolitical fragmentation of compute capacity.
Chinese domestic chip suppliers including Huawei and Cambricon now hold 41% of China's data center accelerator market as U.S. export controls continue to restrict Nvidia access, accelerating the bifurcation of global AI hardware ecosystems.
The global semiconductor foundry market reached $320 billion in 2025 with 16% year-over-year growth, driven by AI chip demand, as TSMC extends its manufacturing dominance amid rising capacity constraints.
Nvidia invested $2 billion in Marvell Technologies despite Marvell's primary clients developing custom ASICs to compete with Nvidia, signaling a strategic bet on ecosystem lock-in through NVLink interconnect standards rather than chip monopoly.
Researchers disclosed new Rowhammer attacks targeting Nvidia GPU memory that can compromise host CPU systems, exposing security vulnerabilities in AI infrastructure as GPU deployment accelerates without adequate hardening.
Key Developments
Microsoft executes multi-billion dollar Asia-Pacific infrastructure expansion
Microsoft committed $10 billion over four years to AI and cloud infrastructure in Japan and $5.5 billion to Singapore through 2029, according to Bloomberg and Data Center Dynamics. The Japan investment represents Microsoft's largest commitment to the country and includes data center capacity, AI skilling programs, and research partnerships. Singapore has hosted Microsoft cloud infrastructure since 2010, making this an expansion of existing operations rather than greenfield deployment.
These investments follow a pattern of hyperscalers building sovereign compute capacity partnerships with allied governments concerned about data residency and supply chain security. Japan has explicitly prioritized domestic AI infrastructure through government policy, while Singapore serves as a strategic hub for Southeast Asian cloud services despite land and power constraints that make per-megawatt buildout costs significantly higher than U.S. markets.
Chinese domestic chip suppliers capture 41% market share as export controls reshape ecosystem
Nvidia's share of China's data center accelerator market has contracted as domestic suppliers including Huawei, Cambricon, and others have collectively reached 41% market share, according to Tom's Hardware. The shift accelerated following tightening U.S. export restrictions that continue to evolve, with Tom's Hardware reporting that frequent regulatory changes are creating planning uncertainty for both chip buyers and suppliers globally.
Huawei's Ascend series and Cambricon's MLU accelerators have gained traction among Chinese cloud providers and enterprises that previously relied on Nvidia hardware. While these chips generally lag Nvidia's latest generations in raw performance and software maturity, they provide sufficient capability for many inference workloads and some training applications. The market consolidation around domestic suppliers represents the most significant forced technology bifurcation in the AI era, creating parallel ecosystems with limited interoperability.
Foundry market hits $320 billion as TSMC extends dominance amid capacity constraints
Global semiconductor foundry revenue reached $320 billion in 2025, growing 16% year-over-year, driven primarily by AI accelerator demand, according to Counterpoint Research data reported by Tom's Hardware. TSMC's market position strengthened further during the period, reflecting its monopoly on advanced process nodes below 5nm that are critical for leading AI chips. The growth rate, while substantial, represents a deceleration from earlier pandemic-era peaks, suggesting foundry expansion is beginning to catch up with demand after years of severe shortages.
Industry observers note that physical and energy constraints are increasingly limiting further buildout velocity. Data Center Dynamics reported that AI infrastructure is hitting physical limits, with efficiency becoming critical as data center power requirements outpace grid capacity expansion in key markets. This creates a paradox where foundries can manufacture chips faster than data centers can deploy the power infrastructure to run them at scale.
Nvidia invests $2 billion in Marvell despite custom ASIC competition
Nvidia invested $2 billion in Marvell Technologies and announced a partnership focused on NVLink Fusion interconnect technology, according to Tom's Hardware. The investment is notable because Marvell's largest customers, including major cloud providers, are actively developing custom ASICs designed to reduce dependence on Nvidia GPUs. Rather than competing directly, Nvidia is positioning NVLink as infrastructure that custom chip designers must adopt to achieve competitive performance in multi-chip systems.
The strategy represents a shift from hardware monopoly to ecosystem control. By making NVLink the de facto standard for high-speed chip-to-chip communication in AI clusters, Nvidia maintains influence even as customers design their own silicon. Marvell's expertise in custom ASIC development and its relationships with hyperscalers make it an ideal partner to embed NVLink across next-generation designs, effectively turning potential displacement into soft lock-in through interconnect dependency.
New GPU memory attacks expose security vulnerabilities in AI infrastructure
Security researchers disclosed GDDRHammer and GeForge, new variants of Rowhammer attacks that exploit Nvidia GPU memory to gain complete control of host systems, according to Ars Technica. The attacks manipulate GPU memory bit flips to compromise CPU security boundaries, demonstrating that AI accelerators introduce attack surfaces beyond traditional CPU vulnerabilities. Unlike previous Rowhammer variants targeting system RAM, these attacks leverage the massive memory bandwidth and capacity of modern GPUs used for AI workloads.
The disclosure comes as GPU deployment accelerates across cloud infrastructure and enterprise data centers without corresponding hardening of security architectures. Multi-tenant cloud environments where GPUs are shared among customers are particularly vulnerable, as one tenant could potentially exploit these vulnerabilities to access another tenant's data or compromise the host system. Nvidia has not yet released mitigation guidance, and it remains unclear whether fixes will require firmware updates, driver changes, or hardware revisions in future GPU generations.
Signals & Trends
Optical interconnect suppliers becoming infrastructure beneficiaries as chip-to-chip bandwidth demands explode
Bloomberg reported that optical component makers are experiencing momentum as investors recognize them as AI infrastructure plays. As GPU clusters scale to hundreds of thousands of accelerators, copper interconnects face physical bandwidth and power efficiency limits. Optical interconnects offer higher bandwidth over longer distances with lower power consumption, making them essential for next-generation AI cluster architectures. This represents a deepening of the AI infrastructure investment thesis beyond chips and data centers into previously overlooked enabling technologies that become critical at hyperscale.
IBM betting on dual-architecture mainframes to remain relevant for AI workloads
IBM announced partnerships with Arm to enable Arm workloads on IBM Z mainframes and LinuxONE systems, targeting AI and data-intensive applications, according to multiple sources including The Register and Data Center Dynamics. The move reflects IBM's recognition that its proprietary mainframe architecture risks irrelevance as AI workloads overwhelmingly target x86 and Arm ecosystems. By creating dual-architecture systems that can run both traditional mainframe workloads and modern Arm-based AI frameworks, IBM is attempting to preserve its installed base while adapting to new compute paradigms. The strategy faces significant execution risk, as running Arm workloads on mainframe-class systems may offer questionable economics compared to purpose-built AI infrastructure.
Infrastructure deals collapsing as AI funding environment tightens beyond pure software plays
Data Center Dynamics reported that AI coding startup Poolside's data center partnership with CoreWeave fell apart, and a planned funding round from Nvidia also collapsed. This follows a pattern of infrastructure deals facing increased scrutiny as capital becomes more selective about AI investments with long buildout timelines and uncertain utilization. While AI software companies continue to raise capital readily, infrastructure plays requiring billions in upfront investment and multi-year development cycles are encountering higher skepticism about demand sustainability and path to profitability. The divergence suggests investors are differentiating between application-layer AI opportunities and infrastructure bets that require correctly timing capacity buildout with uncertain future demand.
Explore Other Categories
Read detailed analysis in other strategic domains