AI Policy Verified · 1 source · primary source

NIST expands CAISI agreements for pre-deployment frontier AI testing

NIST’s CAISI signed new agreements with Google DeepMind, Microsoft, and xAI to run pre-deployment evaluations and expand federal research on AI security.

Posted
May 6, 2026 · 8:30 AM
Original source
May 5, 2026 · Source age: 1 day
Read time
2 min
Sources
1
Verified briefing

Passed source freshness, duplicate, QA, and review checks before publishing. Main source freshness limit: 14 days.

Source count
1
Primary sources
1
QA status
pass

Plain English

What this means in simple words

Before a new model launches, CAISI can test it and share feedback so developers can fix issues or add safeguards before the public relies on it.

What happened

On May 5, 2026, NIST said its Center for AI Standards and Innovation (CAISI) signed agreements with Google DeepMind, Microsoft, and xAI to conduct pre-deployment evaluations and targeted research on frontier AI security.

Why it matters

If more leading labs share models with evaluators before release, it can surface safety and security failures earlier, while creating clearer expectations for how “independent evaluations” are run in practice.

Key points

  • Covers pre-deployment evaluations plus post-deployment assessment and research.
  • NIST says CAISI has completed more than 40 evaluations to date.
  • The agreements can involve testing models with reduced safeguards to assess national-security risks.

What to watch

Watch what evaluation methods CAISI standardizes, what kinds of risks they publish results on, and whether more AI developers join similar testing programs.

Key terms

Pre-deployment evaluation
Testing a model before it is publicly released to understand capabilities and risks.
CAISI
NIST’s Center for AI Standards and Innovation, focused on AI measurement and security research.

Sources

Source dates are original publication dates. The posted date above is when The AI Tea published this explanation.

Related posts