Public Policy & Governance
Top Line
The US Department of Justice has formally intervened in xAI's challenge to Colorado's AI regulation law, signalling the Trump administration's active intent to establish federal pre-emption over state-level AI governance — a direct conflict with the growing patchwork of state legislative activity.
London's Metropolitan Police deployed Palantir's AI surveillance tool against its own officers without apparent public scrutiny, triggering political pushback from Mayor Sadiq Khan and raising acute questions about the governance of AI-enabled internal policing in democratic institutions.
China blocked Meta's $2bn acquisition of AI agent developer Manus, formalising a requirement for explicit government approval before domestic tech firms accept US investment — a concrete cross-border regulatory intervention with immediate M&A implications.
EU AI Omnibus negotiations remain unresolved on core scope and safeguards questions as of April 2026, with Anthropic's Mythos incident intensifying pressure on EU institutions to close governance gaps before the framework is finalised.
Google has reportedly signed a classified AI agreement with the Pentagon covering 'any lawful government purpose', continuing a pattern of Silicon Valley firms deepening defence integration with limited public accountability mechanisms.
Key Developments
DOJ Federal Pre-emption Play on Colorado AI Law Redraws State-Federal Fault Line
The US Department of Justice's intervention in xAI's lawsuit against Colorado's AI regulation law — on 14th Amendment equal protection grounds — is not simply a legal filing; it is an explicit statement of executive branch preference for federal over state AI governance. The Trump administration's argument that Colorado's law discriminates against certain companies is legally narrow, but the political signal is broad: Washington intends to assert supremacy over the emerging state-level regulatory patchwork before it consolidates. Colorado's law, which imposes obligations on developers of high-risk AI systems, was one of the more substantive enacted state frameworks in the US. As reported by The Guardian, the DOJ intervention creates a direct conflict between the administration and state authority.
Cross-jurisdictionally, this mirrors the EU's centralised AI Act model, where member-state divergence is constrained by a single framework — but the US dynamic is adversarial rather than cooperative, and the federal alternative to Colorado's law does not yet exist in enacted form. The DOJ is effectively arguing against state regulation without a federal replacement on the table. Industry players backing federal pre-emption include major AI developers who prefer a single compliance regime; civil liberties groups and state AGs are the primary opposition, arguing that federal inaction creates a vacuum that harms citizens.
Metropolitan Police Palantir Deployment Exposes Governance Vacuum in Public Sector AI Surveillance
The Metropolitan Police's week-long deployment of Palantir's AI tool to surveil its own officers — uncovering alleged misconduct ranging from work-from-home violations to suspected corruption — raises fundamental governance questions that extend beyond the specific use case. As reported by The Guardian, the deployment appears to have occurred without public consultation or disclosed legal basis under UK data protection frameworks. The scale of the resulting investigations — hundreds of officers — suggests the tool's output is being treated as actionable intelligence, which raises due process concerns about AI-generated evidence thresholds in disciplinary proceedings.
The political escalation is significant: Mayor Sadiq Khan's office signalled he may seek to block Scotland Yard from formalising a Palantir contract, citing concerns about 'using public money to support firms who act contrary to London's values' — a direct reference to Palantir's role in US immigration enforcement and Israeli military operations, as reported by The Guardian. This creates an unusual governance tension: the Mayor of London has oversight powers over the Met but limited authority over specific procurement decisions once in process. The UK's existing AI governance framework — based on sector-specific regulators and voluntary norms — provides no clear mechanism to adjudicate this dispute.
China Formalises Investment Veto Power Over Domestic AI Sector, Blocking Meta-Manus Deal
China's blocking of Meta's $2bn acquisition of Manus — confirmed as Beijing requiring explicit government approval for domestic tech companies accepting US investment — represents a concrete regulatory action, not a policy proposal. As reported by The Guardian, the measure transforms what was previously a de facto constraint into a formalised prior-approval regime for inbound US capital in AI. This is a reciprocal governance move: it mirrors US CFIUS restrictions on Chinese investment in American AI companies, but applies at the domestic firm level rather than requiring the acquirer to seek approval.
The Manus case is particularly notable because the target is an AI agent developer — a category that both the US and China regard as strategically sensitive. The blocking formalises the bifurcation of the global AI supply chain at the application layer, not just semiconductors. For multinational AI governance, this means the EU, UK, and other jurisdictions face increasing pressure to define their own investment screening approaches for AI acquisitions before being caught between US and Chinese regulatory camps.
EU AI Omnibus Negotiations Stall as Mythos Incident Tests Governance Readiness
The EU AI Omnibus — intended to streamline and partially roll back compliance obligations from the original AI Act — remains in active negotiation with unresolved disagreements on scope and safeguards, according to CDT Europe's April 2026 AI Bulletin. The Anthropic Mythos incident, in which a model withheld from public release on cybersecurity grounds was allegedly accessed by unauthorised parties, is being cited within EU policy discussions as evidence that the Omnibus must not weaken frontier model oversight provisions. The tension is between member states seeking to reduce compliance burdens on European AI developers and civil society organisations — including CDT Europe — arguing that the original Act's safeguards are already insufficient for the pace of frontier development.
The Omnibus negotiations are a direct regression risk: if concluded in a form that narrows the definition of general-purpose AI models subject to systemic risk obligations, the EU framework would be less stringent at precisely the moment when frontier model incidents are multiplying. The Mythos case illustrates the gap between announcement-based governance — where developers self-report risks and voluntarily restrict access — and enforceable obligations to notify regulators and implement access controls. No current EU or US regulatory framework would have required Anthropic to notify a regulator before the alleged unauthorised access occurred.
Google Pentagon Deal and UK's AI Strategy Rhetoric Expose Public Accountability Deficit in Government AI Adoption
Google's reported classified AI agreement with the Pentagon — allowing use of its models for 'any lawful government purpose' — continues the pattern of major AI developers entering defence contracts with minimal disclosed governance terms, as reported by The Guardian. The classified nature of the deal means standard public procurement accountability mechanisms do not apply, and the breadth of the 'any lawful government purpose' formulation provides no meaningful constraint. This follows similar agreements by OpenAI and others, suggesting a deliberate standardisation of terms that insulates Pentagon AI deployments from oversight.
In the UK, Technology Secretary Liz Kendall's statement that Britain must 'seize the initiative on AI or be left at the mercy and whim' of a future shaped by the technology — reported by The Guardian — is political rhetoric, not a regulatory action. It reflects genuine strategic anxiety about the UK's position given US companies' dominance of AI compute, but the government has not announced new legislative measures or enforcement mechanisms. The concurrent revelation that UK government departments are presenting conflicting forecasts on AI data centre energy demands, as The Guardian reports, suggests the AI strategy lacks basic inter-departmental coordination.
Signals & Trends
The Federal Pre-emption Strategy Is Becoming the Primary Battleground for US AI Governance Architecture
The DOJ's Colorado intervention is not an isolated legal filing — it is part of a coherent executive strategy to prevent the emergence of a state-level regulatory mosaic that would constrain AI developers differently across jurisdictions. With no federal AI legislation in the pipeline and an administration ideologically opposed to prescriptive regulation, federal pre-emption through litigation and executive action serves as a de facto deregulatory instrument. Policy professionals should track parallel developments: whether federal agencies issue AI guidance that pre-empts state action by occupying regulatory space, and whether the administration moves to codify pre-emption through executive order. The trajectory suggests the US will reach a critical juncture within 12-18 months where either a federal framework emerges or the DOJ's litigation strategy forces courts to rule on the constitutional limits of state AI regulation — with outcomes that will reshape the global regulatory reference point.
AI Governance Is Fracturing Along Investment and Acquisition Lines, Not Just Technology Lines
China's Manus blocking decision and the US CFIUS regime are creating a de facto global AI governance architecture built around capital flows rather than technical standards or use-case rules. This is significant because investment screening is more immediately enforceable than technology regulation — it operates through transaction approval rather than ongoing compliance monitoring. The implication for the EU, UK, and other jurisdictions is that their AI governance frameworks need an investment screening dimension that is currently either absent or not AI-specific. The UK's National Security and Investment Act has AI-adjacent provisions but was not designed for the current environment where AI agent developers represent strategic national assets. Jurisdictions that do not develop explicit AI-sector investment screening risk becoming conduits for capital flows that their technology governance frameworks were not designed to handle.
The Gap Between AI Deployment in Public Institutions and Available Accountability Mechanisms Is Widening Rapidly
The Metropolitan Police Palantir deployment, the Pentagon's classified AI contracts, and the UK government's inter-departmental incoherence on AI infrastructure all reflect the same structural problem: public institutions are deploying AI systems at operational scale faster than legislative bodies or oversight agencies can establish enforceable accountability frameworks. In the UK, this is particularly acute because the government's AI governance model relies on existing sector regulators adapting their frameworks — a process that is demonstrably slower than procurement and deployment cycles. The Palantir case is an early but significant example of what happens when high-impact AI deployments outpace governance: political accountability substitutes for regulatory accountability, producing inconsistent and personality-dependent outcomes rather than systematic institutional safeguards. The pattern is likely to intensify as public sector AI adoption accelerates under political pressure to demonstrate AI productivity gains.
Explore Other Categories
Read detailed analysis in other strategic domains