Frontier Models

Last reviewed: 2026-05-11

The most capable AI systems — “frontier models” — are governed by a distinct set of frameworks that overlap with, but go beyond, general AI governance. The EU AI Act’s GPAI obligations, the GPAI Code of Practice, the CAISI testing agreements, California SB 53, Korea’s frontier-safety track, and the International Network of AI Safety Institutes all target this category. This chapter consolidates frontier-model governance in one place because the obligations are otherwise scattered across multiple jurisdictional chapters.

Frontier model governance architecture

Figure: Frontier-model governance is layered — binding regulation at the top, standards and benchmarks operationalising it, voluntary government frameworks bridging to industry, and developer commitments at the base. The AI Safety Institute network coordinates evaluation across all layers.

What counts as a “frontier model”?

There is no universal definition, but converging criteria include:

As of May 2026, models commonly understood as frontier include OpenAI’s GPT-5 (released 7 August 2025), Google’s Gemini 3 (released 18 November 2025), and Anthropic’s Claude Opus 4.5 (released 24 November 2025), along with subsequent Q1-Q2 2026 revisions of each.

EU GPAI obligations (in force August 2025)

Under Articles 53-55 of the EU AI Act, providers of general-purpose AI models must:

The GPAI Code of Practice (10 July 2025) provides a structured way to demonstrate compliance — see EU AI Act for governance structure. Code signatories as of August 2025: Google, Microsoft, OpenAI, Anthropic. Meta declined.

CAISI testing agreements (US, 2025-2026)

The Center for AI Standards and Innovation (formerly NIST AI Safety Institute) has signed bilateral pre-deployment and post-deployment evaluation agreements with frontier developers:[1]

Testing covers cybersecurity, CBRN risk, and a small number of designated dual-use capability categories. The agreements are voluntary; CAISI does not have statutory authority to compel testing of frontier models, but agreements provide structured access for evaluation.

California SB 53 — Transparency in Frontier AI Act

Signed 29 September 2025 and discussed in detail under US State Laws, SB 53 requires large frontier developers to:

SB 53’s substantive structure mirrors the GPAI Code of Practice’s Safety & Security chapter, simplifying dual compliance.

Korea AI Basic Act — frontier safety track

Korea’s AI Basic Act (effective 22 January 2026) includes a frontier-safety track applicable to “high-impact AI” — broadly aligned with frontier-model concepts. Obligations include safety frameworks, pre-deployment evaluation, and incident reporting to Korean authorities. Extraterritorial application means providers serving Korean users are within scope regardless of where models are developed.

Voluntary frameworks and the AI Safety Institute network

Several voluntary mechanisms supplement statutory obligations:

The International Network of AI Safety Institutes — including UK AISI, US CAISI, Singapore, Japan AISI, Korea, France INESIA, and others — conducts coordinated evaluations and shares findings through agreed protocols. India is hosting the next major summit in the series.

Common frontier-safety architecture

Across the EU GPAI Code, SB 53, the CAISI agreements, and most RSPs, a common architecture is converging:

  1. Capability evaluations at defined thresholds — before initial deployment, before scaling, after substantial updates.
  2. Risk identification and mitigation — with documented thresholds at which mitigation actions trigger.
  3. Safety case — structured argument that residual risk is acceptable, addressed to a defined audience (internal governance, regulator, public).
  4. Pre-deployment testing — either internal or in cooperation with AI Safety Institutes.
  5. Post-deployment monitoring — for misuse, incidents, capability change.
  6. Incident reporting — to regulators (EU AI Office, Cal OES, Korean MSIT) and, where applicable, to the Frontier Model Forum or peer institutions.
  7. Transparency — published framework, safety case, model card.

If you are developing frontier models, the minimum credible posture includes:

For deployers and downstream integrators

If you integrate frontier models rather than develop them, the governance focus is:

  1. Verify GPAI compliance of the upstream model (Code signatory, technical documentation available).
  2. Receive and act on downstream documentation provided under Article 53(1)(b) EU AI Act.
  3. Implement use-case-specific risk management — the frontier-safety framework of the upstream developer does not substitute for your own risk assessment of the deployed application.
  4. Maintain provenance — document which model version is in production, what changed when, and your validation of those changes.
  5. Monitor for incidents in your deployment and feed back to the upstream developer.

  1. NIST. Center for AI Standards and Innovation (CAISI). ↩︎