Google, Microsoft, and xAI Volunteered to Let the Government Watch Them
What happened
On May 5, Google, Microsoft, and Elon Musk's xAI agreed to give the US government early access to new AI models for national security testing through the Center for AI Standards and Innovation (CAISI). The move comes one day after the White House confirmed it is weighing mandatory pre-release vetting rules triggered by Anthropic's Mythos model, which alarmed officials with its autonomous capabilities. CAISI has already conducted 40 evaluations for OpenAI and Anthropic. The three new companies are voluntarily joining a review architecture that may soon become compulsory.
These three companies didn't volunteer out of principle. They volunteered because Anthropic just showed them what non-cooperation costs.
Prediction Markets
Prices as of 2026-05-05 — the analysis was written against these odds
The Hidden Bet
Voluntary compliance signals support for AI safety.
Voluntary compliance signals a calculation that mandatory rules are coming and that being early gives you more say in how they're written. Google, Microsoft, and xAI are positioning for regulatory capture, not oversight.
Government review will catch dangerous models before release.
CAISI has 40 prior evaluations and no published methodology, criteria, or enforcement power. A review process with no defined pass/fail is an approval theater, not a safety gate.
The Mythos crisis is what triggered this.
The DoD exclusion of Anthropic from classified contracts was announced the same week. Companies that refused to cooperate with military use terms got shut out. The 'voluntary' agreements may be less about Mythos and more about the Pentagon making clear that access requires compliance.
The Real Disagreement
The genuine fork is between two views of what just happened. One: companies are accepting legitimate security review and the government gets to know what is in frontier models before the public does. Two: the companies that agreed to review will now shape what 'safe' means, locking in definitions that advantage incumbents and slow competitors. Both outcomes follow from the same action. The first view requires believing CAISI will act as a neutral evaluator. The second requires believing Google and Microsoft's legal teams will be in every standards meeting. The second is more structurally likely. I would lean toward concern about regulatory capture, but the cost of that concern is that without any oversight, Mythos-level models go straight to the public with no review at all.
What No One Is Saying
The companies with the most to lose from a mandatory vetting regime just volunteered for it. That is a tell. When dominant incumbents embrace regulation, they are usually writing it.
Who Pays
Open-source AI developers and smaller labs
Immediate if rules pass before end of 2026
If voluntary pre-release review becomes mandatory, it creates a compliance cost that only large companies can absorb. A two-person lab releasing a capable open-weight model cannot run a CAISI evaluation.
Anthropic
Already in effect
Already excluded from DoD classified contracts for refusing military use terms. If it is now also outside the voluntary review coalition, it is doubly isolated from the government relationships that determine where AI gets deployed at scale.
The public
From the first mandatory review cycle onward
Pre-release government review with no public transparency means models may be modified based on national security feedback the public never sees. The reviewed model and the released model may not be the same.
Scenarios
Soft Mandate
The White House issues an executive order requiring pre-release evaluations for models above a capability threshold. The threshold is defined in a process where CAISI consults with the companies already inside the review coalition. Open-source labs are exempted or given a reduced regime.
Signal CAISI publishes draft evaluation criteria in the next 30 days and the comment period includes industry representatives from exactly the companies that already signed agreements.
Voluntary Holds
The arrangement remains informal. Companies submit models to CAISI as a gesture, CAISI produces classified summaries for the NSC, and no model is ever actually blocked. The program becomes a paper compliance track that companies use in congressional testimony.
Signal No model is delayed or modified as a result of CAISI review within six months of the program launching.
Capability Trigger
A CAISI evaluation flags a model as presenting unacceptable autonomous action risk. The White House blocks its release. This is the first real test of whether the review has teeth. The company challenges the decision in court and the legal basis for the block proves thin.
Signal A major model release is delayed by more than two weeks following a CAISI review.
What Would Change This
If CAISI publishes clear, quantitative criteria for what constitutes a failed evaluation and commits to public disclosure of review outcomes without national security redactions, the regulatory capture concern weakens significantly. That would look like actual oversight rather than industry co-management.