As AI systems increasingly operate with autonomy—scheduling meetings, writing code, browsing the web, and executing multi-step tasks—the question of governance becomes critical. OpenAI's interdisciplinary research team has produced this practical framework addressing a fundamental question: how do you govern AI systems that can act independently in the world? This paper doesn't just theorize about risks; it provides concrete practices that developers, deployers, and users can implement today to ensure agentic AI systems remain safe and accountable.
Agentic AI systems are distinguished by their ability to pursue complex goals with limited direct supervision. Unlike chatbots that respond to individual prompts, these systems can:
This autonomy creates new governance challenges. Traditional AI oversight models assume humans review outputs before they have effects. Agentic systems may take consequential actions before any human sees them.
The paper organizes recommendations around seven core practices:
At least one human entity should be accountable for every uncompensated direct harm caused by an agentic AI system. This creates incentives to reduce the likelihood and severity of harms efficiently.
System deployers should provide users with a ledger of actions taken by the agent. This lighter-touch method gives users visibility into agent operations without substantially slowing them down.
Significant decisions by autonomous systems should be reviewed by a human first. The paper provides guidance on which actions warrant approval versus logging.
Agentic systems should operate within clearly defined capability boundaries that limit their potential impact, especially for early deployments.
New agentic capabilities should be rolled out gradually, with monitoring at each stage to catch unexpected behaviors before widespread deployment.
Where possible, agentic actions should be reversible, allowing recovery from errors without permanent harm.
All agentic systems should have reliable shutdown mechanisms that can halt operations when necessary.
The paper identifies distinct parties in the agentic AI lifecycle, each with different responsibilities:
Clear role definitions help distribute responsibility appropriately and ensure no governance gaps exist.
The paper acknowledges the tension between safety and utility. Requiring human approval for every action would negate the benefits of autonomous operation. The framework provides guidance on calibrating oversight levels based on:
This risk-based approach allows agentic systems to operate efficiently while maintaining appropriate safeguards for high-stakes decisions.
Published
2025
Jurisdiction
Global
Category
Governance frameworks
Access
Public access
VerifyWise helps you implement AI governance frameworks, track compliance, and manage risk across your AI systems.