US Government's CAISI to Pre-Test Google, Microsoft and xAI Frontier AI Models Before Public Release
The Center for AI Standards and Innovation (CAISI), housed within NIST at the U.S. Department of Commerce, announced on May 5, 2026 that it has signed pre-deployment evaluation agreements with Google DeepMind, Microsoft, and xAI. The deals require the companies to provide access to frontier AI models before public release for government safety and capability assessments.
Scope of Evaluations
CAISI will conduct "pre-deployment evaluations and targeted research to better assess frontier AI capabilities and advance the state of AI security," including testing for cybersecurity, biosecurity, and chemical weapons risks. Evaluations occur in classified environments. Developers sometimes provide versions of models with reduced or removed safeguards to allow more thorough national security assessments.
Building on Prior Agreements
CAISI previously signed similar agreements with OpenAI and Anthropic in 2024 and has completed more than 40 evaluations to date, including models that remain unreleased publicly. The new agreements extend this framework to cover every major frontier AI developer operating in the United States.
Significance
The development is notable given that the Trump administration came into office opposing AI oversight. The willingness to expand pre-release model testing suggests that national security considerations create bipartisan support for AI governance even when broader regulatory frameworks remain contested. Coverage: CNBC, HPCwire.
Related Articles
The European Parliament and Council agreed on May 7 to simplify the AI Act, pushing high-risk compliance deadlines to December 2027 and August 2028 while adding a new ban on AI-generated non-consensual intimate imagery.
EU Parliament and Council reached a provisional agreement on the Digital Omnibus regulation on May 7, 2026, extending high-risk AI compliance deadlines by up to 24 months and adding a new prohibition on non-consensual sexual AI content.
The Center for AI Standards and Innovation (CASI) secured agreements with Google DeepMind, Microsoft, and xAI to review frontier AI models for national security risks before launch. The policy shift follows alarm over Anthropic's Claude Mythos autonomous cybersecurity capabilities.
Comments (0)
No comments yet. Be the first to comment!