In recent years, the term “AI agent” has become increasingly prevalent in discussions surrounding artificial intelligence. However, the ambiguity surrounding what constitutes an AI agent often leads to confusion. On one hand, we have simple chatbots that can summarize emails, while on the other, there are sophisticated AI systems capable of autonomously analyzing market trends, drafting strategies, and scheduling meetings. Both are labeled as “AI agents,” yet they differ significantly in terms of intelligence, autonomy, and the level of trust we place in them. This disparity raises critical questions about the nature of AI agents, their capabilities, and the implications of their use in various domains.
To understand what an AI agent truly is, we must first delve into the foundational definitions provided by experts in the field. According to Stuart Russell and Peter Norvig, authors of the seminal textbook “Artificial Intelligence: A Modern Approach,” an agent can be defined as anything that perceives its environment through sensors and acts upon that environment through actuators. This definition provides a solid foundation for understanding modern AI agents, which can be broken down into four key components:
1. **Perception (the senses)**: This refers to how an agent takes in information about its digital or physical environment. It encompasses the input stream that allows the agent to comprehend the current state of the world relevant to its tasks.
2. **Reasoning engine (the brain)**: The reasoning engine is the core logic that processes perceptions and determines the next course of action. In contemporary AI agents, this is typically powered by large language models (LLMs). The reasoning engine is responsible for planning, breaking down large goals into manageable steps, handling errors, and selecting the appropriate tools for the task at hand.
3. **Action (the hands)**: This component describes how an agent affects its environment to move closer to its goal. The ability to take action through various tools is what empowers an agent.
4. **Goal/objective**: The overarching task or purpose that guides all of the agent’s actions is encapsulated in this component. It represents the “why” that transforms a collection of tools into a purposeful system. Goals can range from simple tasks, such as finding the best price for a book, to complex objectives like launching a marketing campaign for a new product.
When these components are integrated, a true AI agent emerges as a full-body system. The reasoning engine serves as the brain, but it is rendered ineffective without the senses (perception) to understand the world and the hands (actions) to change it. This complete system, all directed by a central goal, is what creates genuine agency.
However, distinguishing between a standard chatbot and a true AI agent becomes apparent when considering these components. A typical chatbot may perceive a user’s question and respond with an answer, but it lacks an overarching goal and the ability to utilize external tools to achieve it. In contrast, an AI agent possesses the capacity to act independently and dynamically toward a goal, making discussions about levels of autonomy crucial.
The rapid advancement of AI technology can create a sense of navigating uncharted territory. Yet, when it comes to classifying autonomy, we are not starting from scratch. Other industries have grappled with similar challenges for decades, offering valuable insights for the realm of AI agents. The core challenge remains consistent: how do we establish a clear, shared language for the gradual transfer of responsibility from humans to machines?
One of the most successful frameworks for classifying autonomy comes from the automotive industry. The SAE J3016 standard defines six levels of driving automation, ranging from Level 0 (fully manual) to Level 5 (fully autonomous). This model emphasizes two fundamental concepts:
1. **Dynamic driving task (DDT)**: This encompasses everything involved in the real-time act of driving, including steering, braking, accelerating, and monitoring the road.
2. **Operational design domain (ODD)**: These are the specific conditions under which the system is designed to operate, such as “only on divided highways” or “only in clear weather during the daytime.”
For each level, the key question is straightforward: Who is responsible for the DDT, and what is the ODD? At Level 2, the human must supervise at all times. At Level 3, the vehicle can handle the DDT within its ODD, but the human must be prepared to take over if necessary. At Level 4, the vehicle can manage everything within its ODD and can safely pull over if it encounters a problem.
The key insight for AI agents derived from this framework is that a robust classification system is not solely about the sophistication of the AI’s “brain.” Instead, it focuses on clearly defining the division of responsibility between humans and machines under specific, well-defined conditions.
Aviation offers another valuable perspective with its detailed 10-level spectrum of automation, proposed by Parasuraman, Sheridan, and Wickens. This model is particularly useful for systems designed for close human-machine collaboration. For instance, at Level 3, the computer narrows the selection down to a few options for the human to choose from. At Level 6, the computer allows the human a restricted time to veto before executing an action. At Level 9, the computer informs the human only if it decides to do so.
This aviation model is ideal for describing the collaborative “centaur” systems emerging today. Most AI agents will not achieve full autonomy (Level 10) but will exist somewhere along this spectrum, acting as co-pilots that suggest actions, execute with approval, or operate with a veto window.
In the realm of robotics, the National Institute of Standards and Technology (NIST) has developed the Autonomy Levels for Unmanned Systems (ALFUS) framework, which introduces another critical dimension: context. This framework assesses autonomy along three axes:
1. **Human independence**: How much human supervision is required?
2. **Mission complexity**: How difficult or unstructured is the task?
3. **Environmental complexity**: How predictable and stable is the environment in which the agent operates?
The key insight for AI agents from this framework is that autonomy is not a singular metric. An agent performing a simple task in a stable, predictable digital environment (such as sorting files in a single folder) is fundamentally less autonomous than an agent tackling a complex task across the chaotic, unpredictable landscape of the open internet, even if the level of human supervision remains constant.
As we explore the emerging frameworks for AI agents, it becomes evident that no single model suffices. Most proposals fall into three distinct, yet often overlapping, categories based on the primary question they seek to address:
1. **Capability-focused frameworks**: These classify agents based on their underlying technical architecture and what they can achieve. They provide a roadmap for developers, outlining a progression of increasingly sophisticated technical milestones that often correspond directly to code patterns. For example, Hugging Face has proposed a framework that uses a star rating system to illustrate the gradual shift in control from human to AI. This ranges from zero stars (simple processors) to four stars (fully autonomous agents capable of generating and executing entirely new code).
2. **Interaction-focused frameworks**: This category defines autonomy not by the agent’s internal skills but by the nature of its relationship with the human user. The central question here is: Who is in control, and how do we collaborate? For instance, a framework detailed in a paper on Levels of Autonomy for AI Agents defines levels based on the user’s role, ranging from the user as an operator (direct control) to the user as an observer (full autonomy with progress reporting).
3. **Governance-focused frameworks**: These frameworks are less concerned with how an agent operates and more with what happens when it fails. They aim to answer crucial questions about law, safety, and ethics. For example, think tanks like Germany’s Stiftung Neue Verantwortung analyze AI agents through the lens of legal liability, helping regulators determine who is responsible for an agent’s actions: the user who deployed it, the developer who built it, or the company that owns the platform it runs on.
A comprehensive understanding of AI agents requires examining all three questions simultaneously: an agent’s capabilities, how we interact with it, and who bears responsibility for its outcomes.
Despite the advancements in defining and categorizing AI agents, significant gaps and challenges remain. One of the most pressing issues is determining the “road” for a digital agent. The SAE framework for self-driving cars introduced the concept of an operational design domain (ODD), which specifies the conditions under which a system can operate safely. However, what constitutes a safe operational boundary for a digital agent operating across the chaotic expanse of the internet?
The “road” for an AI agent is essentially the entire internet—an infinite, ever-changing environment where websites can be redesigned overnight, APIs can become deprecated, and social norms in online communities can shift rapidly. Defining a “safe” operational boundary for an agent that can browse websites, access databases, and interact with third-party services is one of the most significant unsolved problems in the field. Without a clear digital ODD, we cannot guarantee the same safety standards that are becoming commonplace in the automotive industry.
Currently, the most effective and reliable AI agents tend to operate within well-defined, closed-world scenarios. As previously discussed, focusing on “bounded problems” rather than open-world fantasies is essential for achieving real-world success. This approach involves establishing a clear, limited set of tools, data sources, and potential actions for the agent.
Moreover, while today’s agents excel at executing straightforward plans—such as finding the price of an item using one tool and then booking a meeting with another—they face considerable challenges when confronted with tasks requiring more complex capabilities. These include:
1. **Long-term reasoning and planning**: Many agents struggle to create and adapt intricate, multi-step plans in the face of uncertainty. While they can follow a recipe, they cannot yet invent one from scratch when unexpected situations arise.
2. **Robust self-correction**: What happens when an API call fails, or a website returns an unexpected error? A truly autonomous agent must possess the resilience to diagnose the problem
