AI Agents in Action Foundations for Evaluation and Governance 2025

Page 26 of 34 · WEF_AI_Agents_in_Action_Foundations_for_Evaluation_and_Governance_2025.pdf

Baseline governance mechanisms for AI agents TABLE 2 Governance area Foundational mechanism Purpose Access controlEnforce least-privilege access; define task boundaries.Prevent each agent from accessing unnecessary data, systems, or tools; reduce risk of misuse or accidental harm. Legal and complianceConduct a data protection impact assessment (DPIA); perform privacy and regulation compliance checks, such as General Data Protection Regulation or the California Consumer Privacy Act (CCPA).Ensure data handling and processing complies with relevant laws and regulations. Testing and validationPerform sandbox runs or controlled pilots with non-production data; install input-output filters; perform third-party audits.Validate expected behaviour, detect errors and prevent untested code from affecting live systems, conduct audits (code, red teaming, etc.). Monitoring and loggingImplement logging for all agent actions; set up anomaly alerts or dashboards.Maintain traceability for accountability; enable early detection, incident response and post- incident analysis. Human oversightDefine and assign oversight models, including HITL/HOTL. Require policy review before deployment and set supervisory triggers for exceptions.Ensure accountable human control for material decisions, keep behaviour aligned with organizational policies and provide escalation paths when the agent acts unexpectedly. Traceability and identityAssign unique agent identifiers; tag outputs to the responsible agent instance.Attribute actions and outcomes to specific agents; enable forensic review and performance tracking. Long-term management Establish protocols for ongoing monitoring, updates and eventual decommissioning.Ensure continued alignment, performance and relevance throughout the agent’s life cycle. Trustworthiness and explainability Implement explainability tools; establish trust metrics.Ensure agent behaviour is interpretable and measurable; build user confidence. Manual redundancyEstablish manual redundancy procedures to ensure the sustained continuity of critical business use cases.Preserve data integrity and plan for human resources to take over.The example illustrates that an agent’s overall impact emerges from the interaction of multiple dimensions across function, role, predictability, autonomy, authority and context. As these dimensions shift, so does the risk profile, reinforcing the need for governance frameworks that are both progressive and adaptive. Effective governance requires maintaining an appropriate level of human oversight in relation to the agent’s autonomy, authority and operational context. In high-risk or less predictable settings, a human-in-the-loop (HITL) configuration ensures that agents can suggest or prepare actions, but final decisions remain subject to explicit human approval. In more stable or clearly defined environments, a human-on-the-loop (HOTL) configuration allows agents to act within defined boundaries, while humans monitor behaviour, receive alerts and retain the ability to intervene or override when necessary. Integrating these oversight models into governance structures helps maintain accountability and human judgment as agents operate with greater independence and scale. AI Agents in Action: Foundations for Evaluation and Governance 26
Ask AI what this page says about a topic: