AI Agents in Action Foundations for Evaluation and Governance 2025

Page 14 of 34 · WEF_AI_Agents_in_Action_Foundations_for_Evaluation_and_Governance_2025.pdf

agents capable of planning and executing actions independently across authorized environments. Autonomy in this context refers to an agent’s capacity to decide when and how to act toward a goal, adapting to changing conditions without human guidance. Automation, on the other hand, refers to systems that execute predefined functions reliably under specified conditions without human intervention. The key distinction is that autonomy entails decision- making flexibility (i.e. choosing what to do), whereas automation emphasizes execution reliability (i.e. doing what the system is programmed to do). In the automotive sector, SAE International’s Taxonomy and Definitions for Terms Related to Driving Automation Systems for On-Road Motor Vehicles16 framework defines driving automation from Level 0 (no automation) to Level 5 (full automation). A similar spectrum can be applied to AI agents. This spectrum can be conceived of as moving from no autonomy (for example, a simple chatbot that only answers user queries) to full autonomy (for example, a customer service agent that automates interactions, resolves queries and personalizes responses using a company’s knowledge base). Establishing levels of autonomy can help organizations set clear expectations for functionality and implement proportionate governance mechanisms. Authority defines the actions an agent is permitted to take. It sets the boundaries of system access, such as permissions to use tools, interact with databases or execute transactions. Like autonomy, authority exists on a sliding scale, from read-only access to full administrative control. Autonomy and authority can be combined in different ways, depending on an agent’s purpose and design. They are not inherent system properties but design choices that can be made based on the agents’ intended functions, risk considerations and oversight requirements. They can also be calibrated during assessment or adjusted in real time. Operational context refers to the use case and environment in which the agent operates. The environment is especially critical, as it determines observability, predictability of outcomes, interaction with other agents and how conditions evolve over time.17 Use case defines the domain and environment where the agent performs its distinct function for stakeholders. For example, an autonomous cleaning agent in the residential sector performs household vacuuming and floor cleaning as part of routine home maintenance. Environment represents the operating conditions the agent functions under, ranging from simple and predictable settings to complex, uncertain and dynamic contexts. A complex environment is one where the agent navigates and acts under uncertainty, with incomplete or noisy information, unpredictable outcomes, changing conditions over time, continuous ranges of possible actions or states, and interactions with other agents whose behaviour also affects results. By contrast, a simple environment is one where the agent operates with complete information, predictable and static outcomes, independent episodes, a finite set of states or actions, and no need to consider other actors. Establishing levels of autonomy can help organizations set clear expectations for functionality and implement proportionate governance mechanisms. Classification dimensions FIGURE 6 Agent characteristics 1. Function 3. Predictability 4. Autonomy 5. AuthorityWhat does the agent do? 2. Role Deterministic Non-deterministic Low High Low HighOperational context 6. Use case 7. Environment Specialist Generalist Simple ComplexApplication domain and environment where the agent performs its function AI Agents in Action: Foundations for Evaluation and Governance 14
Ask AI what this page says about a topic: