The White House That Killed AI Regulation Is Now Considering It
What happened
The Trump administration is considering an executive order that would require US government officials to review AI models before their public release, the New York Times reported Monday. The proposed order would create a working group of tech company executives and government officials to develop oversight procedures. The reversal comes after the administration rescinded Biden's 2023 AI executive order as one of its first deregulatory acts. Anthropic's Mythos model, which demonstrated the ability to identify and exploit cybersecurity vulnerabilities at a level that alarmed officials, is reported as the proximate trigger. Separately, Senator Banks introduced the bipartisan AI OVERWATCH Act on May 4, which would expand chip export controls to cover AI model weights and access.
The administration is discovering that 'move fast' works fine until a model ships that can attack critical infrastructure, and now it needs a framework it explicitly dismantled.
Prediction Markets
Prices as of 2026-05-05 — the analysis was written against these odds
The Hidden Bet
Pre-release government vetting can catch dangerous capabilities without slowing development.
The proposed structure gives government 'early access' without delaying releases. But early access without authority to require changes is not oversight; it is notification. A government working group that cannot block a release has theatrical power, not real power.
Anthropic's Mythos is the exceptional case; most models don't require this level of scrutiny.
Mythos is exceptional because Anthropic disclosed its capabilities. Other companies may have models with similar or greater risks that they have not disclosed. The review regime is only as good as the disclosure incentive, and there is currently no legal requirement to disclose dangerous capabilities.
The tech industry will cooperate with a working group that includes company executives.
Including company executives in the review body creates an obvious conflict of interest: a company reviewing a competitor's pre-release model has every incentive to leak, delay, or disadvantage it. The UK's AI Security Institute solved this by keeping reviews inside government with independent technical staff.
The Real Disagreement
The actual fork is whether the US government can design an AI oversight process that is genuinely independent of the companies it regulates, given that those companies are simultaneously the government's allies in the China competition. The Biden approach tried to do this through NIST and independent safety evaluation. The Trump administration dismantled that, then invited tech executives into the regulatory structure itself. The problem is real: Mythos-class capabilities are genuinely dangerous. But putting the companies being regulated on the review committee is not a solution to regulatory capture; it is a formalization of it. I'd lean toward the Biden-era NIST model being structurally sounder, but the current administration will not rebuild what it destroyed.
What No One Is Saying
Anthropic disclosed Mythos's cybersecurity capabilities voluntarily, in a responsible disclosure model it developed itself. The administration is now building a regulatory framework in response to a problem that only became visible because one company chose transparency. If the framework punishes disclosure by adding compliance burden, the next company will not be transparent.
Who Pays
Smaller AI labs and open-source developers
If the executive order is signed, carve-out definitions will be negotiated in the 90 days following.
A pre-release review regime with a government working group and tech-executive membership will inevitably be calibrated to the scale of major frontier labs. Smaller labs and open-source projects will face disproportionate compliance costs or be carved out entirely, freezing the market structure in favor of incumbents.
National security agencies dependent on AI tools
Ongoing tension; visible whenever a classified deployment is delayed pending civilian review.
The DoD, CIA, and NSA depend on commercial AI capability. A pre-release review regime that slows frontier model deployment, even marginally, delays the military AI capabilities the administration says it wants to accelerate.
Anthropic
Immediate; the AI OVERWATCH Act and the proposed executive order both implicate Anthropic's product pipeline.
Anthropic is simultaneously the company that caused this regulatory response and the company already excluded from Pentagon AI contracts for refusing military use terms. If pre-release vetting becomes law, Anthropic faces review from a working group that includes the companies that benefited from its exclusion.
Scenarios
Executive order signed as described
Trump signs an order creating a working group with early-access rights but no blocking authority. Companies comply by sharing models weeks before release. No model is ever actually delayed. The oversight is real on paper and nominal in practice.
Signal White House publishes the EO text; it includes language about 'non-binding review' or 'notification requirements' rather than 'approval requirements'.
Tech industry lobbying kills it
Company executives on the working group design a process so procedurally complex that it is never operationalized. The order stays on the books; no model is ever reviewed. The competitive framing ('this will let China win') dominates the internal White House debate.
Signal No model reviews are completed within 6 months of the order being signed.
Congress moves faster
The AI OVERWATCH Act passes with bipartisan support, creating a statutory review regime with actual teeth. The executive order becomes redundant and is folded into the legislative framework.
Signal AI OVERWATCH Act advances to a floor vote in either chamber within 60 days.
What Would Change This
If a company other than Anthropic voluntarily discloses a dangerous capability in a future model, and that disclosure is treated as a compliance liability rather than a public good, the argument that self-regulation is insufficient becomes undeniable even to the administration.