Public Policy & Governance
Top Line
The U.S. Treasury Department faces mounting legal opposition over a proposed System of Records Notice that would consolidate personal information across nearly all Treasury financial assistance programs, with civil society groups arguing the consolidation lacks legal authority and creates mass surveillance infrastructure.
EFF and allied organisations escalated opposition to state-level age verification mandates, with Minnesota and California bills facing criticism that these measures constitute unconstitutional barriers to internet access rather than child protection.
A Tennessee woman spent nearly six months in jail after AI facial recognition software incorrectly linked her to a North Dakota bank fraud case, underscoring ongoing concerns about law enforcement's reliance on biometric identification systems without adequate safeguards.
The U.S. Department of Defense designated Anthropic a "supply chain risk" and blocked it from government work, prompting the AI company to file both federal court and D.C. Circuit challenges — with Microsoft filing an amicus brief supporting Anthropic's position.
Elon Musk's X agreed to alter its verification mechanism in the European Union following a €120 million fine, though implementation details remain unclear as enforcement of the Digital Services Act continues.
Key Developments
Treasury Department's Personal Data Consolidation Faces Civil Society Opposition
The Center for Democracy & Technology joined multiple organisations in filing comments objecting to the U.S. Department of Treasury's System of Records Notice (SORN) that would consolidate personal information of individuals associated with Treasury-administered financial assistance programs. CDT argues the consolidation is "arguably illegal" and would create a vast database of Americans' personal information without clear statutory authority. The proposed SORN would link data across programs including pandemic relief, housing assistance, and other financial aid schemes — effectively creating centralised surveillance infrastructure that civil liberties groups argue exceeds Treasury's mandate.
The objections highlight a pattern of federal agencies attempting to expand data collection and retention capabilities through administrative rulemaking rather than explicit congressional authorisation. This approach allows agencies to circumvent legislative debate about privacy trade-offs whilst building systems that could be repurposed for mass surveillance or enforcement beyond their stated purpose.
State Age Verification Laws Face Coordinated Civil Liberties Pushback
The Electronic Frontier Foundation published analyses opposing age verification mandates in Minnesota (HF1434) and California (A.B. 1043), arguing these measures create "unnecessary and unconstitutional barriers for adults and young people to access information and express themselves online." EFF emphasised that available age verification technologies all involve privacy trade-offs — requiring identity document uploads, biometric scanning, or third-party verification services that create new surveillance points and data breach risks. Minnesota Representative Leigh Finke testified that HF1434 represents control rather than protection, EFF reported, highlighting how these bills won't actually protect children whilst imposing speech restrictions on adults.
The opposition reflects a broader pattern across U.S. states where legislatures, frustrated by federal inaction on tech regulation, are passing overlapping and sometimes contradictory mandates. This creates compliance chaos for platforms whilst failing to address the underlying business model incentives that drive harmful content recommendation systems. The focus on age gates rather than systemic design changes suggests these bills serve political theatre more than child safety.
AI Facial Recognition Leads to Six-Month Wrongful Detention in Tennessee
Angela Lipps, a 50-year-old Tennessee grandmother, spent nearly six months in jail after Fargo police used AI facial recognition software to identify her as a suspect in a North Dakota bank fraud investigation, The Guardian reported. The case represents another documented instance of AI misidentification leading to wrongful arrest and extended detention — following similar incidents in Detroit, New Orleans, and other U.S. cities. Lipps now faces the challenge of rebuilding her life after the extended incarceration based on algorithmic error.
The incident highlights the persistent gap between law enforcement adoption of facial recognition technology and the implementation of safeguards to prevent wrongful identification. Despite mounting evidence of error rates — particularly for women and people of colour — and documented cases of wrongful arrest, most U.S. jurisdictions lack binding regulations on police use of facial recognition. The technology is treated as merely another investigative tool rather than one requiring heightened procedural protections given its opacity and error potential.
Pentagon-Anthropic Dispute Escalates with Microsoft Intervention
Microsoft filed an amicus brief supporting Anthropic's legal challenge against the Department of Defense's designation of the AI company as a "supply chain risk," The Guardian reported. The Pentagon designation effectively bars Anthropic from government contracts — a significant restriction given the company's Claude AI is integrated into Microsoft systems used by federal agencies. Anthropic filed both a civil complaint in the Northern District of California and a petition with the D.C. Circuit Court of Appeals. Microsoft's intervention signals that the dispute extends beyond one company to broader questions about DOD authority to exclude AI providers based on opaque security determinations.
The case reflects tension between Pentagon efforts to control AI supply chains for national security and the commercial reality that frontier AI models are developed by private companies with diverse customer bases including foreign entities. DOD's designation authority was originally designed for traditional defence contractors and equipment suppliers — its application to general-purpose AI models raises novel questions about how security screening applies to software that can be accessed via API rather than physically controlled. Microsoft's brief likely argues that excluding Anthropic would force it to replace Claude throughout its federal systems, disrupting existing contracts.
EU Digital Services Act Enforcement Yields X Platform Concessions
Elon Musk's X agreed to alter its verification mechanism in the European Union following a €120 million fine from the European Commission, Bloomberg reported. The fine and subsequent agreement represent concrete enforcement of the Digital Services Act's provisions requiring platforms to prevent their design features from misleading users. X's paid verification system, which allows anyone to purchase a checkmark without identity verification, was deemed to violate DSA requirements that verification badges indicate actual identity confirmation rather than simply payment.
The enforcement action demonstrates the Commission's willingness to impose significant fines and mandate design changes — not merely issue warnings or open investigations. However, the specifics of X's agreed changes remain unclear, as does the timeline for implementation and how the Commission will verify compliance. The case establishes precedent that platform design choices traditionally considered product decisions are subject to regulatory override when deemed to mislead users or facilitate harm.
Signals & Trends
Administrative Rulemaking Becoming Primary Vector for Expanding Government Surveillance Infrastructure
The Treasury SORN controversy exemplifies a pattern where federal agencies use administrative procedures to create cross-program data consolidation systems without explicit legislative authorisation. This approach allows agencies to build surveillance capabilities through technical rulemaking that receives minimal public attention compared to legislation. The strategy effectively circumvents democratic debate about privacy trade-offs by framing data consolidation as routine administrative housekeeping rather than a policy choice about government access to personal information. As agencies adopt AI systems that benefit from centralised data access, expect increased attempts to consolidate records across programme boundaries using administrative authority.
State-Level Regulatory Fragmentation Creating De Facto National Mandates Through Compliance Convergence
The proliferation of state age verification laws represents a broader pattern where lack of federal action drives states to pass overlapping requirements that collectively create national policy through compliance convergence. Companies facing different requirements in California, Minnesota, and other states typically implement the most restrictive standard everywhere rather than maintain jurisdiction-specific systems. This allows state legislators to effectively set national policy without the usual constraints of interstate commerce clause challenges — particularly when requirements involve identity verification or content moderation. The result is governance through accumulated state mandates rather than deliberate federal policymaking.
Pentagon Supply Chain Designations Emerging as Industrial Policy Tool for AI Sector
The Anthropic designation reveals that DOD supply chain risk determinations, originally designed for traditional defence procurement, are being extended to general-purpose AI models — effectively giving Pentagon officials veto power over commercial AI companies' participation in the federal market. This expansion transforms supply chain security from a procurement tool into industrial policy, allowing DOD to influence which AI providers succeed commercially by determining federal market access. The lack of transparent criteria or judicial review means these determinations function as administrative industrial policy without the usual checks on agency discretion. As AI becomes infrastructure, expect increased contestation over whether security agencies should have this level of influence over commercial technology development.
Explore Other Categories
Read detailed analysis in other strategic domains