Back to Resources

AI Governance Frameworks

Comprehensive overview of leading AI governance frameworks, standards, and regulatory requirements from around the world.

NIST AI Risk Management Framework

National Institute of Standards and Technology (USA)

VoluntaryPublished

A voluntary framework to help organizations manage risks associated with AI systems. Provides a structured approach to identifying, assessing, and mitigating AI risks.

Key Principles

  • Valid and Reliable - AI systems perform consistently and accurately
  • Safe - AI systems do not pose unreasonable risks
  • Secure and Resilient - AI systems are protected from threats
  • Accountable and Transparent - AI systems are explainable and traceable
  • Explainable and Interpretable - AI decisions can be understood
  • Privacy-Enhanced - AI systems protect individual privacy
  • Fair with Harmful Bias Managed - AI systems are equitable

Core Components

Govern

Establish policies, processes, and procedures for responsible AI development and use

Map

Understand the context, categorize risks, and document the AI system

Measure

Assess, analyze, and track identified AI risks and impacts

Manage

Allocate resources and take actions to address identified risks

Version

1.0 (January 2023)

Applicability

All organizations developing, deploying, or using AI systems

EU Artificial Intelligence Act

European Union

Mandatory (phased: 6-36 months)Enacted (phased implementation)

World's first comprehensive AI regulation. Risk-based approach categorizing AI systems by risk level (unacceptable, high, limited, minimal) with corresponding obligations.

Key Principles

  • Risk-Based Classification - AI systems categorized by risk level
  • Prohibited AI Practices - Certain high-risk applications banned
  • High-Risk Requirements - Strict obligations for high-risk AI
  • Transparency Obligations - Users must know they're interacting with AI
  • General-Purpose AI Rules - Special requirements for foundation models
  • Governance and Enforcement - EU AI Office and national authorities
  • Innovation Support - Regulatory sandboxes and SME support

Core Components

Prohibited AI Systems

AI systems that pose unacceptable risk (e.g., social scoring, real-time biometric surveillance)

High-Risk AI Systems

AI in critical areas (employment, education, law enforcement) with strict requirements

Limited Risk AI

Transparency obligations (e.g., chatbots must disclose they are AI)

Minimal Risk AI

No specific obligations beyond general law

Version

Final Text (March 2024)

Applicability

All AI systems placed on EU market or affecting EU citizens

ISO/IEC 42001:2023

International Organization for Standardization

Voluntary (but enables certification)Published

International standard for AI Management Systems (AIMS). Provides requirements and guidance for establishing, implementing, maintaining, and continually improving AI systems.

Key Principles

  • Context of the Organization - Understand stakeholders and requirements
  • Leadership - Top management commitment and policy
  • Planning - Risk assessment and objectives
  • Support - Resources, competence, awareness, communication
  • Operation - Planning, development, and control of AI systems
  • Performance Evaluation - Monitoring, measurement, analysis
  • Improvement - Nonconformity, corrective action, continual improvement

Core Components

AI Management System

Framework for managing AI development, deployment, and operation

Risk Management

Systematic approach to identifying and mitigating AI risks

Impact Assessment

Evaluation of AI system effects on individuals and society

Data Governance

Controls for data quality, provenance, and lifecycle management

Version

1.0 (December 2023)

Applicability

Organizations of any size developing or using AI systems

OECD AI Principles

Organisation for Economic Co-operation and Development

VoluntaryPublished

First intergovernmental standard on AI, adopted by 42 countries. Five values-based principles for responsible stewardship of trustworthy AI.

Key Principles

  • Inclusive Growth, Sustainable Development and Well-being
  • Human-Centered Values and Fairness
  • Transparency and Explainability
  • Robustness, Security and Safety
  • Accountability

Core Components

Values-Based Principles

Five high-level principles for trustworthy AI

National Policy Recommendations

Guidance for governments on AI policy and investment

International Cooperation

Framework for cross-border collaboration on AI governance

Version

Adopted May 2019

Applicability

Governments and organizations globally

Singapore Model AI Governance Framework

Infocomm Media Development Authority (Singapore)

VoluntaryPublished

Practical framework with detailed implementation guidance. Focus on operationalizing ethical AI principles through concrete practices.

Key Principles

  • Internal Governance Structures and Measures
  • Human Involvement in AI-Augmented Decision-Making
  • Operations Management
  • Stakeholder Interaction and Communication

Core Components

Implementation Guide

Detailed practices for each principle with concrete examples

Companion Guides

Industry-specific guidance (e.g., financial services, healthcare)

AI Verify

Open-source testing toolkit for AI governance validation

Version

2.0 (January 2020)

Applicability

Organizations deploying AI, especially in regulated industries

Canada Algorithmic Impact Assessment

Treasury Board of Canada Secretariat

Mandatory for Canadian federal governmentMandatory for federal government

Required assessment tool for Canadian federal government automated decision systems. Risk-based approach determining oversight requirements.

Key Principles

  • Transparency and Accountability
  • Legal and Policy Compliance
  • Data Quality and Relevance
  • System Security and Robustness
  • Human-in-the-Loop Requirements

Core Components

Impact Assessment Tool

Questionnaire-based risk scoring system (Level I-IV)

Mitigation Requirements

Risk-based requirements for peer review, testing, monitoring

Transparency Obligations

Public disclosure requirements based on impact level

Version

Version 4.0

Applicability

Canadian federal government (reference for other jurisdictions)

Framework Selection Guidance

For U.S. Organizations:

Start with NIST AI RMF as your foundation. It provides comprehensive risk management guidance and aligns with U.S. regulatory expectations.

For EU Market Participants:

EU AI Act compliance is mandatory. Use ISO 42001 for systematic implementation and NIST AI RMF for risk management practices.

For Global Organizations:

Combine OECD AI Principles (high-level values), ISO 42001 (management system), and region-specific regulations where you operate.

For Practical Implementation:

Singapore Model Framework provides the most detailed implementation guidance with concrete practices and industry-specific examples.