MIT's AI Incident Tracker Harm Taxonomy represents the most systematic effort to date for cataloging and categorizing real-world AI failures. Unlike theoretical risk frameworks, this dataset captures actual incidents that have already occurred, organizing them through a multi-dimensional lens that examines harm levels, threat characteristics (novelty, autonomy, imminence), and national security implications. The resource provides both a searchable incident database and a structured taxonomy that enables pattern recognition across time, geography, and AI application domains.
While most AI risk taxonomies focus on hypothetical scenarios or high-level categories, MIT's approach grounds risk assessment in documented reality. The taxonomy goes beyond simple harm categorization by introducing temporal analysis—tracking how incident patterns evolve over time—and incorporating national security impact assessments for each documented case. The three-dimensional threat characterization (novelty, autonomy, imminence) provides a nuanced framework that distinguishes between different types of AI risks, from novel but predictable failures to autonomous systems causing immediate harm.
The classification system operates across multiple layers:
Harm Levels: Incidents are categorized by severity and scope of impact, from individual privacy violations to systemic societal harms.
Threat Characteristics: Each incident receives scores across three dimensions:
National Security Assessment: Every incident includes evaluation of potential implications for national security, infrastructure, or governance systems.
Temporal Tracking: The database maintains chronological records enabling analysis of how AI incident patterns change over time, revealing emerging threat vectors and evolving risk landscapes.
AI safety researchers will find the empirical foundation invaluable for validating theoretical risk models against real-world data. Policy makers and regulators can use the taxonomy to understand which types of AI harms are actually materializing and at what frequency. Risk management professionals in organizations deploying AI systems can benchmark their risk assessments against documented incident patterns. Insurance companies and auditors will find the categorization system useful for developing AI-specific risk evaluation methodologies. Academic researchers studying AI governance can leverage the dataset for longitudinal studies of AI safety trends.
Start by exploring incidents within your specific domain or use case to understand historical failure patterns. Use the three-dimensional threat scoring to calibrate your own risk assessment processes—incidents scoring high on autonomy and imminence deserve different mitigation strategies than those scoring high on novelty alone. The national security assessments can inform enterprise risk discussions, particularly for organizations in critical infrastructure sectors.
For policy applications, the temporal analysis features enable evidence-based discussions about regulatory priorities. Instead of debating theoretical AI risks, stakeholders can examine which harms are actually increasing in frequency or severity over time.
The taxonomy captures only incidents that have been publicly reported and documented, creating potential bias toward high-profile failures while missing routine or unreported harms. The national security impact assessments reflect analysis at the time of documentation and may not account for evolving geopolitical contexts. Additionally, the three-dimensional threat scoring system, while systematic, involves subjective judgments that may not translate perfectly across different organizational contexts or regulatory frameworks.
The dataset's focus on documented incidents also means it cannot predict entirely novel categories of AI harm that haven't yet occurred, limiting its utility for forward-looking risk planning in rapidly evolving AI capabilities.
Published
2024
Jurisdiction
Global
Category
Risk taxonomies
Access
Public access
US Executive Order on Safe, Secure, and Trustworthy AI
Regulations and laws • White House
Safe, Secure, and Trustworthy Development and Use of Artificial Intelligence
Regulations and laws • U.S. Government
Highlights of the 2023 Executive Order on Artificial Intelligence
Regulations and laws • Congressional Research Service
VerifyWise helps you implement AI governance frameworks, track compliance, and manage risk across your AI systems.