Safety & Standards
Safety & Standards Deep-Dive
Thursday, March 05, 2026
Top Line
OpenAI admits zero control over Pentagon AI use as Sam Altman tells employees the company cannot influence military deployment decisions, raising fundamental questions about safety governance when frontier models enter defence applications. The Guardian
Anthropic-Pentagon standoff reveals enforcement vacuum in AI safety commitments: Dario Amodei called OpenAI's military messaging "straight up lies" after Anthropic withdrew over safety disputes, yet US military continues using Claude for targeting operations in Iran with no clarity on who enforces safety boundaries. TechCrunch / TechCrunch
First wrongful death lawsuit against Google's Gemini alleges AI product "fuelled son's delusional spiral," marking a test case for liability frameworks that currently provide little recourse when generative AI causes documented psychological harm. BBC
Meta privacy violations escalate as regulators contact company over workers watching intimate videos from AI glasses—including toilet use and sex—captured without informed consent and reviewed by Kenyan subcontractors, exposing gaps between privacy controls and actual data handling. BBC
Tech industry lobbying against Anthropic security designation as major groups representing Google and Apple urge Trump to drop national security risk label, arguing it will create "detrimental ripple effects"—a collision between commercial interests and supply chain security concerns. Bloomberg
Key Developments
Military AI Deployment: The Accountability Black Hole
OpenAI CEO Sam Altman acknowledged to employees this week that his company "does not control how the Pentagon uses their artificial intelligence products in military operations"—a statement that crystallises the fundamental governance gap in defence AI deployment. The Guardian This admission comes as US Central Command confirms AI tools are managing "enormous amounts of data for operations against Iran," with multiple sources indicating AI involvement in targeting decisions. Bloomberg TechCrunch
The governance vacuum is most visible in the Anthropic-OpenAI dispute. After Anthropic terminated its Pentagon contract over AI safety disagreements, OpenAI immediately secured the deal—prompting Anthropic CEO Dario Amodei to privately call OpenAI's messaging around the arrangement "straight up lies." TechCrunch Yet US military forces continue using Anthropic's Claude models "for many targeting decisions" in ongoing Iran operations, with defence-tech clients simultaneously fleeing Anthropic over the controversy. TechCrunch OpenAI is now "planning additional protections" and "surveillance safeguards" following its hastily announced Pentagon deal, Financial Times while Amodei has reopened discussions with Pentagon officials to reach a compromise. Financial Times Bloomberg
Meanwhile, Smack Technologies operates openly, training models specifically "to plan battlefield operations" with no comparable public debate over safety constraints. Wired
Why it matters: The military AI deployment pattern reveals that voluntary safety commitments—Anthropic's responsible scaling policies, OpenAI's use case restrictions—have no binding force once models reach government hands, and no institutional mechanism exists to enforce them post-deployment.
What to watch: Whether Anthropic-Pentagon negotiations produce any enforceable framework that other AI companies must follow, or whether "additional protections" remain performative measures with no oversight or compliance requirements.
The Liability Gap: When AI Causes Documented Harm
A father has filed the first wrongful death lawsuit against Google alleging that Gemini AI "fuelled son's delusional spiral," marking a critical test of whether existing product liability frameworks can address generative AI harms. BBC The case arrives as evidence mounts that "chatbot-induced delusions" represent a genuine psychological risk: "Large language models often prioritise agreeability over truthfulness to the detriment of users," creating interactions where AI systems reinforce rather than challenge harmful beliefs. Financial Times
Simultaneously, Meta faces regulatory contact over workers—specifically Kenya-based subcontractors—watching "intimate AI glasses videos" including recordings of users "using the toilet or having sex." BBC This incident exposes how data handling practices diverge from user expectations despite nominally robust privacy controls: "We have more privacy controls yet less privacy than ever" captures the gap between policy and practice. BBC
The pattern extends to intellectual property: Grammarly's rebranded Superhuman tool now offers "'expert' AI reviews from your favorite authors—dead or alive," providing feedback based on famous writers' work "without their permission." Wired Apple Music is attempting to address similar concerns by adding "transparency tags to distinguish AI music," but effectiveness remains uncertain since "the label or distributor has to opt in to tagging their music as AI." TechCrunch
Why it matters: Current liability frameworks provide minimal recourse when AI causes documented harm—whether psychological damage, privacy violations, or IP theft—because responsibility diffuses across model developers, deployment platforms, and subcontractors, with no party clearly accountable.
What to watch: The Google Gemini wrongful death ruling will set precedent for whether product liability can reach AI systems, or whether Section 230-style protections insulate companies from harm caused by model outputs.
Standards Theatre: Voluntary Pledges Without Enforcement
Seven tech giants signed Trump's "rate payer protection pledge" this week, committing that data center expansion won't spike electricity costs for ordinary Americans. The Verge Wired Trump himself admitted data centers "need some PR help," Wired and energy market experts are blunt about the pledge's substance: "There are reasons to doubt it will" lower costs, Politico with the president's "build your own power plant" requirement unlikely to shield ratepayers from footing infrastructure bills. Politico
The pledge exemplifies a broader pattern: high-profile commitments designed for optics rather than implementation. The event featured CEOs from Google, Meta, Microsoft, Oracle, OpenAI, Amazon, and xAI but established no compliance mechanism, enforcement timeline, or penalties for violation. The Verge Analysis describes it as having "good optics and little substance." Wired
This stands in contrast to binding regulatory action: Google's revamp of Android app stores this week stems from actual antitrust litigation and European regulatory requirements, not voluntary commitments. Bloomberg The difference is enforcement—court orders and regulatory mandates with legal consequences for non-compliance versus pledges with none.
Why it matters: The proliferation of voluntary AI safety commitments and industry pledges creates an illusion of governance while deferring the binding standards, compliance requirements, and accountability mechanisms that would actually constrain harmful practices.
What to watch: Whether any signatory to the data center pledge faces consequences for rate increases, or whether the agreement remains purely performative—a test case for whether voluntary frameworks ever translate to enforceable obligations.
Signals & Trends
The National Security Lens Is Reshaping AI Safety Debates: Tech industry groups representing Google and Apple are lobbying Trump to remove Anthropic's national security risk designation, arguing it will cause "detrimental ripple effects for the rest of the industry." Bloomberg This marks a shift from safety debates focused on model capabilities to supply chain security concerns, where commercial deployment becomes a national security question. The collision between Anthropic's AI safety stance and Pentagon demands demonstrates how geopolitical pressures override voluntary safety commitments when strategic interests align—a pattern likely to intensify as AI becomes central to military operations across multiple conflicts.
Post-Deployment Control Is the Real Safety Gap: Multiple developments this week—OpenAI's admission of no Pentagon oversight, continued Claude use for Iran targeting despite Anthropic's withdrawal, and the Gemini wrongful death case—reveal that the critical safety failure point isn't model development but deployment governance. Companies announce safety protocols, responsible scaling policies, and use restrictions during development, yet possess no mechanism to enforce them once models reach customers, whether military or consumer. The absence of post-deployment monitoring, compliance verification, or enforcement authority means safety commitments effectively expire at the API handoff. This gap will only widen as models proliferate into higher-stakes applications.
Liability Uncertainty Is Freezing Enterprise AI Adoption: The combination of the Google Gemini lawsuit, Meta's privacy violations, and IP concerns around unauthorised training data creates legal exposure that enterprises cannot quantify. Without clear liability frameworks—who is responsible when an AI system causes harm? the model developer, the deployment platform, the end user?—enterprise risk officers lack the information needed to approve high-stakes AI applications. The current patchwork of Section 230 protections, product liability doctrines, and IP law produces maximum uncertainty, likely slowing adoption in regulated industries far more effectively than any safety standard could.
Explore Other Categories
Read detailed analysis in other strategic domains