NIST expands CAISI agreements for pre-deployment frontier AI testing
NIST’s CAISI signed new agreements with Google DeepMind, Microsoft, and xAI to run pre-deployment evaluations and expand federal research on AI security.
Passed source freshness, duplicate, QA, and review checks before publishing. Main source freshness limit: 14 days.
- Source count
- 1
- Primary sources
- 1
- QA status
- pass
Plain English
What this means in simple words
Before a new model launches, CAISI can test it and share feedback so developers can fix issues or add safeguards before the public relies on it.
What happened
On May 5, 2026, NIST said its Center for AI Standards and Innovation (CAISI) signed agreements with Google DeepMind, Microsoft, and xAI to conduct pre-deployment evaluations and targeted research on frontier AI security.
Why it matters
If more leading labs share models with evaluators before release, it can surface safety and security failures earlier, while creating clearer expectations for how “independent evaluations” are run in practice.
Key points
- Covers pre-deployment evaluations plus post-deployment assessment and research.
- NIST says CAISI has completed more than 40 evaluations to date.
- The agreements can involve testing models with reduced safeguards to assess national-security risks.
What to watch
Watch what evaluation methods CAISI standardizes, what kinds of risks they publish results on, and whether more AI developers join similar testing programs.
Key terms
- Pre-deployment evaluation
- Testing a model before it is publicly released to understand capabilities and risks.
- CAISI
- NIST’s Center for AI Standards and Innovation, focused on AI measurement and security research.
Sources
Source dates are original publication dates. The posted date above is when The AI Tea published this explanation.
- CAISI Signs Agreements Regarding Frontier AI National Security Testing With Google DeepMind, Microsoft and xAI NIST · Press release · Original source May 5, 2026 · Source age 1 day Primary