Back to AI lexicon
Emerging & Specialized Topics

AI assurance

AI assurance

Quick definition

AI assurance is the process of measuring, evaluating, and communicating how trustworthy an AI system is. It means verifying that AI systems work reliably, treat people fairly, stay secure, and meet ethical and legal standards. The practice borrows from established assurance disciplines in finance, engineering, and cybersecurity, and gives users, regulators, and other stakeholders a reason to trust the system.

The UK government describes AI assurance as a three-part toolkit: Measure (gathering qualitative and quantitative performance data), Evaluate (assessing risks against established benchmarks), and Communicate (reporting findings via dashboards, reports, or certification).

How AI assurance differs from AI audit and AI governance

These terms are often conflated, but they mean different things:

  • AI Governance is the set of structures, policies, and processes that guide AI deployment. It usually sits with a cross-functional leadership group spanning Legal, Technology, Risk, HR, and Business Operations.
  • AI Assurance acts as the independent verification arm. It supplies evidence to the governing body that the controls and policies in place are actually working. It covers the full range of tools, processes, standards, certifications, and provider services that give ongoing or periodic confidence about AI system behavior.
  • AI Audit is a point-in-time examination of an AI system's lawfulness, ethics, and technical soundness. Auditing is one activity within the broader assurance practice.

Put simply: governance sets the rules, assurance verifies they work, and auditing checks compliance at a specific moment.

Why it matters

Any company deploying AI in sensitive or high-risk areas needs assurance. Without it, AI systems can produce biased, unsafe, or non-compliant outcomes, and the company faces legal penalties, reputational damage, and operational risk.

Assurance activities help demonstrate accountability under regulations like the EU AI Act, ISO 42001, and NIST AI RMF. The UK AI assurance market alone had over 524 companies operating in it as of 2024, with an estimated value of roughly one billion pounds, and is projected to grow as regulatory requirements expand.

Core assurance techniques

The UK government identifies six core AI assurance techniques, split between proactive upstream guardrails and reactive downstream guardrails:

Risk assessment

Identifies organizational, reputational, legal, and technical risks before and during deployment. Risk assessment forms the foundation of any assurance program by mapping potential failure modes to their likely impact.

Impact assessment

Looks at broader societal effects on equality, human rights, data protection, and the environment. Where risk assessment focuses on the system itself, impact assessment considers downstream consequences for affected populations.

Bias audit

Examines inputs, outputs, and decision patterns for unlawful discrimination or systematic unfairness. Auditors test across protected attributes, measure fairness metrics, and document disparities along with remediation actions.

Compliance audit

Checks adherence to internal policies, regulations, and contractual obligations. While bias audits zero in on fairness, compliance audits cover the full range of applicable rules and standards.

Conformity assessment

Uses testing, inspection, and certification to show that an AI product meets specified requirements before it enters the market. Under the EU AI Act, conformity assessment is especially important for high-risk systems.

Formal verification

Applies mathematical methods to prove that a system satisfies specified requirements. More rigorous than empirical testing, but currently practical only for certain types of AI system properties.

Standards for AI assurance

ISO/IEC 42001

The first international AI management system standard, published in December 2023. It sets out requirements for establishing, implementing, maintaining, and improving an AI Management System (AIMS). Accredited third-party bodies can certify against it, and it addresses ethical considerations, transparency, and continuous learning. The number of certified companies grew throughout 2024 as AI adoption accelerated.

NIST AI Risk Management Framework

A voluntary, flexible framework built around four functions: Govern (policies and culture), Map (identifying and framing risks), Measure (analyzing risks), and Manage (implementing responses). NIST released a Generative AI Profile in July 2024 that addresses risks specific to LLMs and generative AI systems.

IEEE 7000 series

Ethics-focused standards for AI and autonomous system design. IEEE 7000-2021 covers ethical concerns during system design; IEEE 7001-2021 addresses transparency of autonomous systems. IEEE's CertifAIEd program provides independent assessments, and as of late 2024, 167 authorized assessors were active across 28 countries.

AI assurance in the EU AI Act

The EU AI Act creates a tiered conformity assessment regime based on risk classification.

For most high-risk AI systems (covering areas like HR, education, credit scoring, and law enforcement), providers can follow an internal control procedure, essentially a self-assessment. But for systems in safety-critical sectors such as medical devices, machinery, and vehicles, conformity assessment must involve a notified body: an independent third-party organization officially designated by EU Member State authorities.

Assessment areas under the Act include risk management systems, data governance practices, technical documentation, transparency obligations, human oversight mechanisms, accuracy and robustness requirements, and post-market monitoring plans.

Third-party assurance providers

The AI assurance market has several tiers of providers:

Certification bodies like BSI, DNV, SGS, and PECB carry out ISO/IEC 42001 audits and certification. Accreditation bodies (UKAS in the UK, ANAB in the US) attest the competence of these providers.

Specialized AI assurance platforms such as Holistic AI, Credo AI, and Monitaur offer commercial assurance tooling. HITRUST launched the first certifiable AI security assessment aligned with ISO, NIST, and OWASP standards. The Cloud Security Alliance launched STAR for AI in late 2025, and Anthropic, Sierra, and Zendesk were among the first to post ISO/IEC 42001 certificates in the public STAR Registry.

National initiatives are also emerging. Singapore's AI Verify Foundation launched a Global AI Assurance Pilot to codify norms around technical testing of generative AI applications, pairing assurance providers with enterprise deployers.

The UK AI assurance ecosystem

The UK has built the most structured national approach to growing an AI assurance market. The Centre for Data Ethics and Innovation (now the Responsible Technology Adoption Unit within DSIT) published roadmaps identifying six priority areas: demand building, market development, standards creation, professionalization, regulatory enablement, and research partnership.

DSIT also ran the Fairness Innovation Challenge in 2024, funding teams that developed new bias auditing approaches for healthcare, HR, financial services, and higher education. The UK government maintains a portfolio of AI assurance techniques with real-world case studies across sectors, covering technical, procedural, and educational methods.

Key challenges

Lack of standardized metrics

No universal benchmark exists for measuring AI trustworthiness. Metrics for fairness, explainability, and robustness are contested, domain-specific, and sometimes contradictory, which has led companies to adopt inconsistent approaches.

The generative AI assurance gap

Most current assurance work focuses on foundation model safety during pre-deployment, not end-to-end application reliability in enterprise settings. LLMs exhibit emergent behaviors that are hard to characterize through traditional input-output testing, and those behaviors can shift unpredictably with prompt variation, fine-tuning, or model updates.

Talent and skills shortage

Third-party AI assurance providers struggle to find qualified people. Few professionals combine the necessary domain expertise, AI/ML technical depth, and governance or audit methodology knowledge.

The self-assessment problem

Under the EU AI Act, internal self-assessment rather than third-party review is the default conformity pathway for most high-risk AI systems. The result is a structural tension: the same company develops, deploys, and verifies its own system.

Real-world example

A national healthcare provider develops an AI system to prioritize patient treatment. Before deployment, the team runs bias tests across demographic groups, validates the model against clinical benchmarks, and brings in third-party auditors to verify that predictions are fair. After deployment, continuous monitoring tracks performance metrics and detects drift, while regular assurance reviews confirm ongoing compliance with internal policies and regulatory requirements. The layered approach reduces the risk of discrimination claims and supports regulatory approval.

Best practices

  • Establish governance first. Define clear accountability and ownership for AI risk. Keep the development and deployment function separate from assurance and oversight.

  • Match effort to risk. High-stakes systems in healthcare, finance, hiring, and law enforcement demand rigorous third-party assessment. Lower-risk systems can often rely on internal review.

  • Adopt a recognized framework. ISO/IEC 42001 covers management system requirements; NIST AI RMF handles risk identification. Pursue certification where it signals credibility to customers, regulators, or partners.

  • Don't stop at launch. Assurance is not a one-time certification event. Set up continuous monitoring for model drift, bias emergence, and performance degradation throughout the system's life.

  • Keep thorough records. Document training data provenance, model architecture choices, evaluation results, known limitations, and incident history.

  • Invest in internal skills. Train staff in AI risk, bias evaluation, and explainability. Build internal red teaming capacity and acquire tooling for monitoring model behavior in production.

  • Treat generative AI differently. LLM-based systems need additional measures: adversarial prompt testing, output toxicity and hallucination evaluation, and supply chain transparency for the underlying model providers.

FAQ

What is the goal of AI assurance?

To make sure AI systems perform reliably, ethically, and within the law. Assurance gives users, regulators, and partners verifiable evidence that systems are operating as intended.

Who is responsible for AI assurance?

It depends on the company's structure and the system's risk level, but responsibility usually falls on AI governance teams, risk management professionals, compliance officers, and sometimes external auditors. Internal audit provides independent assurance to the governing body that AI risk is being managed effectively.

Is AI assurance mandatory?

In some sectors and jurisdictions, yes. The EU AI Act requires conformity assessment for high-risk systems before they can enter the market. Even where it is not legally required, assurance is widely considered a best practice for managing AI risk.

How is AI assurance different from AI auditing?

Auditing is one activity within the broader assurance practice. An audit assesses a system against a checklist or framework at a specific point in time. Assurance takes a wider and more continuous view, covering ongoing risk management, monitoring, testing, documentation, and communication of trustworthiness across the full AI lifecycle.

What standards support AI assurance?

The primary standards are ISO/IEC 42001 (AI management systems), the NIST AI Risk Management Framework, and the IEEE 7000 series (ethics in AI design). Other relevant standards include ISO/IEC 23894 for AI risk management, ISO/IEC 24027 for bias in AI systems, and ISO/IEC 25059 for AI quality models.

How should companies approach AI assurance for generative AI?

Generative AI needs additional assurance measures beyond what traditional ML models require. Key areas include adversarial prompt testing, output toxicity and hallucination evaluation, content safety monitoring, and supply chain transparency for the underlying model providers. The three most common enterprise concerns are accuracy and robustness, use-case-specific regulatory compliance, and content safety.

Implement AI assurance in your organization

Get hands-on with VerifyWise's open-source AI governance platform

AI assurance | AI Governance Lexicon | VerifyWise