Agentic AI is dominating industry conversations. While it may seem like a new buzzword, it’s rooted in a rich historical field of AI research: autonomous agent systems.
Today, AI technology has enabled these agents to reach unprecedented levels of autonomy, enabling them to plan and interact with unstructured data. In an enterprise context, AI agents are systems capable of managing certain business operations with minimal or no human intervention.
These agents integrate with an organization's data infrastructure and possess an interface to receive user inputs and deliver outputs. Consider a chatbot for financial services customer support, which can answer questions about the institution's offerings, retrieve customer information or documents and even execute financial transactions.
AI agents display various levels of autonomy based on their ability to interact with their environment (accessing data and executing actions), their planning capabilities and the degree of human control. As it stands today, AI Agents in the enterprise often have limited autonomy, but with technology advancing rapidly, fully autonomous agents likely in the future.
As the technology evolves, it’s clear that we need a common terminology for discussing autonomy. A shared language and understanding is crucial in enabling leaders to effectively assess the maturity of agentic AI systems, as well as the associated benefits, risks and trade-offs at each level.
A good starting point for this discussion is the widely adopted — although not formally standardized — framework for autonomous driving capability. Adapting this scale to reflect the levels of autonomy displayed by an AI agent is a useful way to begin benchmarking maturity.
“We're announcing today that we are going to have fully autonomous vehicles in commercial operation by 2021”
2016, Mark Fields (Ford CEO)
Enterprise is currently having its self-driving car moment. There is a huge buzz about what agentic AI will be able to do, but, like cars, enterprises are not going to change from no autonomy to full autonomy one day to the next. It’s worth noting, some AI automation already existed in both cases - Advanced Driver Assistance Systems (ADAS)for cars and Optical Character Recognition (OCR) for enterprise. Like cars, enterprises will change one step at a time, iteratively
Below is an adaptation of the driving-automation framework for AI agents. Each level captures an agent’s task scope, planning sophistication and required human oversight.
AI agent autonomy is assessed by examining how systems coordinate tasks, sequence their operations and involve humans in decision making. We use three dimensions to define levels of maturity:
● Automation: The scope of tasks a system can coordinate
● Planning: How tasks are sequenced, either programmatic (human-designed) or AI-driven
● Human control: The level of oversight required for any action on the environment
By mapping these criteria across to the proposed levels of agent autonomy, we can see how each level delivers against these key areas.
This proposed structure doesn’t just provide a clearer way to talk about and assess agentic automation. It’s crucial in establishing an objective and tailored AI governance framework.
By linking each autonomy level to specific governance requirements, we can calibrate controls to match actual capabilities, balance innovation with oversight and build trust with stakeholders. This maturity-based approach creates a living framework that evolves alongside AI agents, ensuring compliance and strategic alignment.
Uncover fresh perspectives with our handpicked blog posts on AI advancements.