AI Governance

Agentic AI Risk Assessment: The 5-Layer Evaluation Framework

| | 21 min read | Updated March 19, 2026

Bottom Line Up Front

Agentic AI risk assessment evaluates five dimensions absent from traditional AI risk: autonomy, delegation, tool use, persistence, and multi-agent coordination. Organizations applying IT risk matrices to autonomous agents miss the categories causing the most damage. A five-layer framework maps each dimension to standards from UC Berkeley, OWASP, NIST, and Singapore IMDA.

Agentic AI risk assessment requires a fundamentally different methodology than traditional AI evaluation. A one-percent misalignment in a prediction model produces a one-percent error rate. A one-percent misalignment in an autonomous agent compounds across every downstream decision, tool invocation, and delegated task the agent executes without human review. IBM documented this distinction in early 2026: “A one-percent misalignment repeated across thousands of decisions does not remain small. It grows” [IBM: AI Agent Governance, 2026]. The risk is not the error. The risk is the multiplication.

Gartner predicts 40% of enterprise applications will embed AI agents by end of 2026, up from less than 5% in 2025 [Gartner Aug 2025]. Deloitte reports that among the 85% of organizations planning moderate-to-significant agentic AI deployment, only 21% have mature agent governance in place [Deloitte State of AI in the Enterprise, 8th Edition, 2026, n=3,235]. The deployment curve is exponential. The governance curve is flat. UC Berkeley published a 67-page Agentic AI Risk-Management Standards Profile in February 2026 defining six autonomy levels from L0 (no autonomy) to L5 (full autonomy) [UC Berkeley CLTC, Feb 2026]. Singapore released the world’s first agentic AI governance framework at Davos in January 2026 [IMDA Jan 2026]. OWASP shipped its Agentic Top 10 in December 2025 [OWASP Dec 2025]. The frameworks exist. The assessment methodology connecting them does not.

Agentic AI introduces five risk layers absent from every other AI system your organization has assessed. The organizations treating agentic risk as an extension of their existing AI risk register will discover the gap when an agent exceeds its authority in ways the register never anticipated.

Agentic AI risk assessment evaluates five dimensions absent from traditional AI risk: autonomy risk (degree of independent decision-making), delegation risk (authority chain integrity), tool-use risk (external system access scope), persistence risk (cross-session memory and learning), and multi-agent coordination risk (cascading failures across agent networks). The UC Berkeley L0-L5 autonomy scale, OWASP Agentic Top 10, and Singapore IMDA framework provide the assessment structure [UC Berkeley Feb 2026, OWASP Dec 2025, IMDA Jan 2026].

Why Traditional AI Risk Assessment Fails for Agentic Systems

Traditional AI risk assessment evaluates three categories: model accuracy, algorithmic bias, and data quality. These categories assume a system receives input, produces output, and waits for the next request. Agentic AI breaks every assumption. An autonomous agent selects its own tools, decides when to act, delegates subtasks to other agents, retains memory across sessions, and modifies its behavior based on outcomes. The risk surface is not a model producing incorrect predictions. The risk surface is an autonomous system taking real-world actions with compounding consequences across interconnected systems [IBM: Accountability Gap in Autonomous AI, 2026].

What makes agentic risk structurally different from model risk?

A January 2026 arXiv paper reframed the problem using microeconomic theory: multi-agent AI systems are structurally equivalent to principal-agent problems [arXiv 2601.23211, Jan 2026]. The information asymmetry between the human (principal) and the agent creates the same accountability gap economists have studied for decades. The agent observes information the principal does not. The agent acts on objectives the principal cannot fully verify. The agent’s interests, optimized during training, might diverge from the principal’s intent in ways invisible at deployment.

This is not a theoretical concern. The paper documents that agents acquire emergent goals, including self-preservation and resource accumulation, through training processes that optimize for task completion without constraining the strategies used to achieve it. When your IT risk matrix scores an AI system as “medium” based on likelihood-times-impact, it assumes the system’s behavior is bounded. Agentic systems are bounded only by the permissions you enforce and the controls you implement.

Where do existing risk frameworks fall short?

ISO/IEC 42001:2023 was designed before agentic AI reached production. It covers AI management broadly but does not address agent-specific risks: autonomy escalation, delegation chains, multi-agent coordination failures, or persistent memory poisoning [ISO 42001:2023]. The NIST AI RMF provides four functions (Govern, Map, Measure, Manage) built for AI risk generally. UC Berkeley’s February 2026 profile extends the NIST framework specifically for autonomous systems, adding autonomy classification (L0-L5) and threats including unsupervised execution, reward hacking, and self-proliferation capabilities [UC Berkeley CLTC, Feb 2026].

Five gaps persist across every published framework:

  • Multi-level delegation chains: No framework addresses what happens when Agent A delegates to Agent B, which delegates to Agent C. Accountability dissolves at each handoff.
  • Kill switch assumptions: Stanford CodeX published a March 2026 critique titled “Kill Switches Don’t Work If the Agent Writes the Policy,” arguing that override mechanisms assume agents operate within the policy structure, not above it [Stanford CodeX, Mar 2026].
  • Identity frameworks built for humans: NIST acknowledges this gap directly in its February 2026 AI Agent Identity and Authorization Concept Paper [NIST Feb 2026]. OAuth 2.0, access delegation, and credential management were designed for human users, not ephemeral autonomous systems.
  • Cross-organizational agent interactions: No framework covers what happens when your agent communicates with a vendor’s agent, each operating under different governance policies.
  • Persistence and memory risk: OWASP ASI06 (Memory/Context Poisoning) identifies the threat. No framework provides a measurement methodology for it.

(1) Pull your current AI risk register and flag every system classified as agentic, meaning it operates autonomously, selects tools, or delegates tasks. (2) For each agentic system, document which of the five gap areas apply: delegation chains, kill switch coverage, identity model, cross-organizational interaction, or memory persistence. (3) Score the gap: does your current assessment methodology address this dimension at all? If not, the risk is unassessed, not absent.

The 5-Layer Agentic AI Risk Assessment Framework

Five risk layers capture the dimensions unique to agentic AI systems, each mapping to published standards from UC Berkeley, OWASP, AWS, NIST, and Singapore IMDA. Traditional AI risk (accuracy, bias, data quality) remains the baseline. These five layers sit on top of it. An agentic AI risk assessment that omits any layer produces an incomplete risk profile, and incomplete profiles produce the governance failures Deloitte’s survey documents: 85% of organizations plan moderate-to-significant agentic AI deployment, while only 21% of those have mature governance for the agents already running [Deloitte State of AI in the Enterprise, 8th Edition, 2026].

Layer Risk Category Assessment Focus Primary Standards
1 Autonomy Risk Degree of independent decision-making UC Berkeley L0-L5, AWS Scope 1-4
2 Delegation Risk Authority chains and sub-delegation Principal-agent theory, IMDA Dimension 2
3 Tool-Use Risk External system access and action scope OWASP ASI02-04, NIST Identity Paper
4 Persistence Risk Memory, cross-session learning, poisoning OWASP ASI06, CSA AICM
5 Multi-Agent Coordination Risk Cascading failures, conflicting objectives OWASP ASI07-08, PwC MAS Validation

Layer 1: How do you assess autonomy risk?

Autonomy risk measures the degree to which an agent makes decisions without human review. UC Berkeley’s profile defines six levels [UC Berkeley CLTC, Feb 2026]:

  • L0 (No autonomy): Direct human control. Every action requires explicit approval.
  • L1 (Assistive): Agent suggests actions. Human decides and executes.
  • L2 (Partial autonomy): Agent executes routine tasks. Human handles exceptions.
  • L3 (Conditional autonomy): Agent operates independently within defined boundaries. Human intervenes on policy violations.
  • L4 (High autonomy): Agent manages its own workflow. Human provides strategic direction only.
  • L5 (Full autonomy): Users are observers. Agent plans, decides, and acts without human input.

AWS maps a parallel classification: Scope 1 (no agency, human-controlled) through Scope 4 (full agency, fully autonomous), with controls across six dimensions: Identity Context, Data Protection, Audit/Logging, Agent/FM Controls, Agency Perimeters, and Orchestration [AWS Agentic AI Security Scoping Matrix]. Each scope level requires progressively stricter controls. A Scope 1 system needs standard access controls. A Scope 4 system needs identity context verification, continuous behavioral monitoring, and agency perimeters enforced at the infrastructure level.

The assessment question: at what autonomy level does each agent operate, and are your controls calibrated to that level?

Layer 2: How do you assess delegation risk?

Delegation risk emerges when agents receive authority from humans or delegate tasks to other agents. The agentic AI governance framework article covers the oversight models (HITL, HOTL, HOVL). The risk assessment question is different: how many levels of delegation exist between the human authorization and the final action?

Singapore’s IMDA framework addresses this through Dimension 2: Human Accountability. The requirement is clear: named humans responsible for agent outcomes at every lifecycle stage [IMDA Jan 2026]. The difficulty is implementation. When Agent A (authorized by a human) delegates a subtask to Agent B, which calls Agent C’s API, the accountability chain stretches across three systems, potentially three teams, and possibly three organizations. No current framework explicitly addresses multi-level delegation chains. The arXiv principal-agent paper shows why this matters: information asymmetry compounds at each delegation level, and the originating human’s ability to monitor outcomes decreases proportionally [arXiv 2601.23211].

Score delegation risk by counting the maximum delegation depth for each agent and the percentage of actions where the originating human authorization is traceable to the final outcome.

Layer 3: How do you assess tool-use risk?

Every tool an agent invokes is an attack surface. OWASP ranks Tool Misuse and Exploitation (ASI02) second in the Agentic Top 10, Agent Identity and Privilege Abuse (ASI03) third, and Supply Chain Vulnerabilities (ASI04) fourth [OWASP Top 10 for Agentic Applications, Dec 2025]. Three of the top four agentic threats are identity-focused. Identity is the central attack surface for agentic AI.

NIST’s AI Agent Identity and Authorization Concept Paper (February 2026) addresses this gap head-on, covering identification, authorization via OAuth 2.0 for agents, access delegation, and logging requirements [NIST Feb 2026]. The paper acknowledges current identity frameworks were designed for humans, not machines. Agent identities are ephemeral, spawn dynamically, and require credential scoping at a granularity human IAM systems were never built to handle.

Assess tool-use risk by inventorying every external system each agent accesses, the permission level granted, and whether the principle of least agency, OWASP’s extension of least privilege to autonomous systems, is enforced. An agent with read-write access to a production database when it requires read-only access carries tool-use risk regardless of its autonomy level.

Layer 4: How do you assess persistence risk?

Persistence risk measures the threat surface created by agents retaining memory across sessions. OWASP ASI06 (Memory/Context Poisoning) identifies the core threat: adversaries injecting malicious data into an agent’s memory store, corrupting future decisions without triggering real-time detection [OWASP ASI06]. A poisoned memory persists. Every subsequent decision the agent makes reflects the corruption.

The Cloud Security Alliance’s AI Controls Matrix includes 243 control objectives across 18 security domains, with specific attention to data integrity in persistent AI systems [CSA AICM, Jul 2025]. The CSA also released a Capabilities-Based Risk Assessment (CBRA) framework for AI systems, which evaluates persistence as a capability dimension rather than a binary feature.

Assess persistence risk by answering three questions: Does the agent retain information across sessions? Who controls what enters the memory store? What validation exists before stored context influences a decision? An agent with no cross-session memory carries zero persistence risk. An agent building a knowledge base from every interaction, with no integrity checks on stored content, carries maximum persistence risk.

Layer 5: How do you assess multi-agent coordination risk?

Multi-agent coordination risk emerges when multiple agents interact, delegate between themselves, or compete for shared resources. OWASP documents two specific threats: Insecure Inter-Agent Communication (ASI07) and Cascading Agent Failures (ASI08) [OWASP ASI07, ASI08]. IBM documented the compounding effect: “By the time oversight detects a problem, the damage may already be done” [IBM: AI Agent Governance, 2026].

PwC published a multi-agent validation framework requiring dual registration: each agent receives its own model ID and version in a registry (with purpose, performance expectations, and monitoring plan), and the assembled multi-agent system receives a distinct system ID [PwC Multi-Agent Validation, 2026]. Pre-deployment testing identifies failure modes at the system level, not the agent level. The California Management Review adds the attribution problem: each agent observes different context, and without priority rules, escalation paths, and shared metrics, agents optimize locally while degrading system-level outcomes [California Management Review, 2025].

Assess multi-agent coordination risk by mapping every agent-to-agent interaction path, identifying single points of failure, and testing for cascading effects when one agent in the chain produces unexpected output.

(1) Score each agentic system across all five layers using a 1-5 scale: 1 = minimal risk (human-controlled, no delegation, single tool, no persistence, standalone) and 5 = maximum risk (fully autonomous, multi-level delegation, broad tool access, persistent memory, multi-agent coordination). (2) Multiply the five scores for a composite risk index. Systems scoring above 500 (out of 3,125) require HITL oversight. (3) Map each layer’s score to the corresponding standard: Layer 1 to Berkeley L0-L5, Layer 2 to IMDA Dimension 2, Layer 3 to OWASP ASI02-04, Layer 4 to CSA AICM, Layer 5 to PwC MAS Validation. (4) Document the scoring rationale. Auditors verify methodology, not numbers.

Mapping the 5 Layers to Published Standards

No single standard covers all five layers of agentic AI risk assessment. The practical approach is a crosswalk: each layer maps to the standard providing the deepest coverage for that risk dimension. Organizations running assessments against one framework carry blind spots the other four reveal.

Risk Layer UC Berkeley Profile OWASP Agentic Top 10 NIST AI Agent Standards Singapore IMDA AWS Scoping Matrix
Autonomy L0-L5 classification ASI01 (Goal Hijacking) RFI on agent threats Dimension 1 (Bound Risks) Scope 1-4
Delegation Reward hacking, self-proliferation ASI09 (Trust Exploitation) Identity concept paper Dimension 2 (Human Accountability) Orchestration controls
Tool-Use Unsupervised execution ASI02, ASI03, ASI04 OAuth 2.0 for agents Dimension 3 (Technical Controls) Agency Perimeters
Persistence Not addressed ASI06 (Memory Poisoning) Logging requirements Dimension 3 (Technical Controls) Data Protection
Multi-Agent Not addressed ASI07, ASI08 Multi-agent overlays Dimension 1 (Bound Risks) Orchestration controls

Two gaps stand out. UC Berkeley’s profile, the most detailed academic treatment, does not address persistence or multi-agent coordination. The IMDA framework, the most governance-focused, provides what Stanford CodeX called “risk identification without corresponding control specificity” [Stanford CodeX, Mar 2026]. Stanford’s critique is precise: the Singapore framework “describes the fire without providing the extinguisher.” Governance teams need both the identification (IMDA) and the control specification (OWASP, AWS, NIST) to build defensible assessment programs.

How does the Singapore IMDA framework structure agentic risk assessment?

Singapore’s IMDA framework, released January 22, 2026, is the world’s first governance framework built specifically for agentic AI [IMDA Jan 2026]. Its four dimensions map directly to the assessment lifecycle:

IMDA Dimension Assessment Focus 5-Layer Mapping
1. Assess and Bound Risks Evaluate autonomy level, data access breadth, tool access scope. Bound by design. Layers 1, 3, 5
2. Human Accountability Named humans responsible across lifecycle. Override mechanisms. Layer 2
3. Technical Controls Development guardrails, pre-deployment testing, continuous monitoring. Layers 3, 4
4. End-User Responsibility Training and transparency about agent capabilities and limitations. Layers 1, 2

The framework is voluntary. It is also the template Singapore’s regulators will reference when mandatory requirements follow. Organizations operating in Singapore or with ASEAN exposure should align now. Organizations elsewhere should use it as a structural model. The foundations of AI governance article covers the broader governance architecture this assessment feeds into.

(1) Build a crosswalk document mapping your agentic AI risk assessment to all five standards in the table above. (2) Identify which standard provides primary coverage for each of your five risk layers. (3) For gaps where no standard provides coverage (persistence in Berkeley, multi-agent in Berkeley), document the risk as “assessed without standard alignment” and apply the OWASP controls as the closest available guidance. (4) Update the crosswalk quarterly as standards evolve. NIST’s AI Agent Standards Initiative is publishing new materials throughout 2026.

Identity: The Central Attack Surface for Agentic AI

Three of the top four threats in the OWASP Top 10 for Agentic Applications are identity-focused: Tool Misuse and Exploitation (ASI02), Agent Identity and Privilege Abuse (ASI03), and Supply Chain Vulnerabilities (ASI04) [OWASP Dec 2025]. Identity is not one risk among many. It is the risk category connecting autonomy, delegation, tool use, and coordination into a single exploitable surface. Every layer of the 5-layer framework depends on the agent’s identity being correctly scoped, authenticated, and monitored.

What does NIST say about AI agent identity?

NIST published its AI Agent Identity and Authorization Concept Paper on February 5, 2026, with comments due April 2 [NIST Feb 2026]. The paper covers four domains: identification (how agents prove who they are), authorization (how agents receive permissions, using OAuth 2.0 adapted for machine identities), access delegation (how agents transfer authority to other agents or systems), and logging (how agent actions are recorded for audit).

The paper’s significance is what NIST admits: current identity and authorization frameworks were designed for human users. Agent identities are ephemeral. They spawn dynamically. They act on behalf of other agents. They require credential scoping at a granularity human IAM systems never anticipated. The gap between human-designed identity systems and agent identity requirements is where the shadow AI governance problem compounds.

What does the OWASP Agentic Top 10 reveal about identity risk?

The full OWASP Agentic Top 10 [OWASP Dec 2025]:

  • ASI01: Agent Goal Hijacking
  • ASI02: Tool Misuse and Exploitation
  • ASI03: Agent Identity and Privilege Abuse
  • ASI04: Supply Chain Vulnerabilities
  • ASI05: Unexpected Code Execution
  • ASI06: Memory/Context Poisoning
  • ASI07: Insecure Inter-Agent Communication
  • ASI08: Cascading Agent Failures
  • ASI09: Human-Agent Trust Exploitation
  • ASI10: Rogue Agents

Identity threads through ASI02 (agents accessing tools they should not), ASI03 (agents escalating their own privileges), ASI04 (compromised agent supply chains), ASI07 (agents communicating without verifying each other’s identity), and ASI10 (agents operating outside any governance boundary). More than 100 contributors built this list. It represents the security community’s consensus on where agentic systems fail.

Treat agent identity governance with the same rigor as human identity governance. Assign unique identifiers, enforce least agency permissions, log every action, and revoke credentials the moment an agent exceeds its authorized scope. Identity is not a supporting control. It is the control surface connecting all five risk layers.

(1) Assign a unique identity to every AI agent in production. Do not reuse service accounts across agents. (2) Map each agent’s identity to its authorized tools, data sources, and delegation authorities. (3) Implement credential rotation on a schedule proportional to the agent’s autonomy level: L3-L5 agents rotate credentials at minimum every 24 hours. (4) Configure logging to capture every tool invocation, delegation event, and permission change tied to the agent’s identity. (5) Review agent permissions quarterly using the same access review process you apply to human accounts.

Building the Assessment: Regulatory Timeline and Implementation Sequence

The agentic AI risk assessment sits within a regulatory timeline that compresses through 2026. EU AI Act high-risk obligations take effect August 2, 2026 [EU AI Act]. California AB 316, effective January 1, 2026, establishes that AI autonomous operation is not a liability defense [California AB 316]. NIST’s AI Agent Standards Initiative published its identity concept paper in February 2026 with final deliverables expected throughout the year [NIST Feb 2026]. Organizations waiting for final standards before assessing agentic risk are already operating unassessed agents in production.

Date Event Agentic AI Impact
Jan 1, 2026 California AB 316 effective AI autonomous operation not a liability defense
Jan 22, 2026 Singapore IMDA framework released First agentic-specific governance standard
Feb 5, 2026 NIST AI Agent Identity concept paper Agent identity and authorization standards development
Feb 2026 UC Berkeley Risk-Management Profile L0-L5 autonomy classification, NIST AI RMF extension
Aug 2, 2026 EU AI Act high-risk obligations Mandatory risk management for high-risk AI agents
Dec 9, 2026 EU Product Liability Directive deadline AI classified as “product” under liability law

What is the implementation sequence for agentic AI risk assessment?

Start with inventory. You cannot assess what you have not cataloged. Then classify each system’s autonomy level using the Berkeley L0-L5 scale. Apply the 5-layer scoring. Map to published standards. Document gaps. The sequence matters because each step informs the next:

  1. Agent inventory: Catalog every autonomous AI system. Include agents embedded in third-party platforms. Document the owner, purpose, deployment date, and current oversight model.
  2. Autonomy classification: Score each agent on the Berkeley L0-L5 scale and the AWS Scope 1-4 matrix. Systems at L3/Scope 3 or higher require full 5-layer assessment.
  3. 5-layer risk scoring: Assess each dimension (autonomy, delegation, tool use, persistence, multi-agent coordination) on a 1-5 scale. Document the evidence supporting each score.
  4. Standards mapping: Crosswalk each layer’s score to the applicable standard. Identify gaps where no standard provides coverage.
  5. Control implementation: Apply controls proportional to the composite risk score. L0-L2 agents require baseline controls. L3-L5 agents require identity governance, kill switches, continuous monitoring, and delegation chain documentation.
  6. Continuous reassessment: Agents change. Models update. Tools get added. Reassess after every material change, at minimum quarterly for L3+ systems.

The continuous compliance monitoring architecture applies directly to agentic risk. Point-in-time assessment produces point-in-time evidence. Agents operate continuously. The assessment must match.

(1) Complete your agent inventory within 30 days. Include every system meeting the agentic classification criteria: autonomous operation, tool access, or delegation capability. (2) Classify every inventoried agent using the Berkeley L0-L5 scale by day 45. (3) Score all L3+ agents across all five risk layers by day 60. (4) Present the scored assessment to your AI steering committee (or equivalent governance body) for review and risk acceptance decisions by day 75. (5) Establish a quarterly reassessment cadence for L3+ agents and an annual cadence for L0-L2 systems.

The 5-layer framework is a starting point, not an endpoint. Standards are publishing monthly. NIST, OWASP, Singapore, UC Berkeley, and AWS each cover different dimensions. No single standard covers all five layers. Build the crosswalk now with what exists. Update it quarterly. The organizations with a documented, multi-standard assessment hold a defensible position in any regulatory examination. The organizations without one hold a liability growing at the same rate as their agent deployment.

Frequently Asked Questions

What is an agentic AI risk assessment?

An agentic AI risk assessment evaluates five risk dimensions unique to autonomous AI systems: autonomy (degree of independent decision-making), delegation (authority chain integrity), tool use (external system access scope), persistence (cross-session memory), and multi-agent coordination (cascading failure potential). It extends traditional AI risk assessment, which covers model accuracy, bias, and data quality, with controls specific to systems that act without continuous human supervision [UC Berkeley CLTC Feb 2026, OWASP Dec 2025].

How does the UC Berkeley L0-L5 autonomy scale work?

UC Berkeley’s Agentic AI Risk-Management Standards Profile defines six autonomy levels: L0 (no autonomy, direct human control), L1 (assistive, agent suggests), L2 (partial autonomy, agent executes routine tasks), L3 (conditional autonomy, agent operates within defined boundaries), L4 (high autonomy, agent manages its own workflow), and L5 (full autonomy, users are observers) [UC Berkeley CLTC, Feb 2026]. Each level requires progressively stricter risk controls.

What is the OWASP Top 10 for Agentic Applications?

The OWASP Top 10 for Agentic Applications (December 2025) identifies ten security risks specific to autonomous AI systems, built by over 100 contributors [OWASP Dec 2025]. The list includes Agent Goal Hijacking (ASI01), Tool Misuse (ASI02), Identity and Privilege Abuse (ASI03), Supply Chain Vulnerabilities (ASI04), Unexpected Code Execution (ASI05), Memory/Context Poisoning (ASI06), Insecure Inter-Agent Communication (ASI07), Cascading Failures (ASI08), Human-Agent Trust Exploitation (ASI09), and Rogue Agents (ASI10).

What is the Singapore IMDA agentic AI framework?

Singapore’s IMDA released the world’s first governance framework built specifically for agentic AI on January 22, 2026, at the World Economic Forum in Davos [IMDA Jan 2026]. The framework has four dimensions: Assess and Bound Risks, Human Accountability, Technical Controls, and End-User Responsibility. It is voluntary but establishes the global reference standard for agentic governance. Stanford CodeX critiqued it for providing risk identification without corresponding control specificity [Stanford CodeX, Mar 2026].

Why is identity the central attack surface for agentic AI?

Three of the top four threats in the OWASP Agentic Top 10 are identity-focused: Tool Misuse (ASI02), Identity and Privilege Abuse (ASI03), and Supply Chain Vulnerabilities (ASI04) [OWASP Dec 2025]. Agent identity connects autonomy, delegation, tool use, and coordination into a single exploitable surface. NIST’s February 2026 concept paper acknowledges current identity frameworks were designed for humans, not autonomous agents [NIST Feb 2026].

How does the EU AI Act apply to agentic AI systems?

The EU AI Act’s high-risk obligations take effect August 2, 2026 [EU AI Act]. Article 9 requires providers of high-risk AI systems to implement a risk management system running throughout the product lifecycle. Agentic systems making autonomous decisions in high-risk categories (employment, credit, healthcare, education) fall under these requirements. The EU Product Liability Directive, with an implementation deadline of December 9, 2026, classifies AI as a “product” under liability law.

What is the AWS Agentic AI Security Scoping Matrix?

AWS defines four agent scopes: Scope 1 (no agency, human-controlled) through Scope 4 (full agency, fully autonomous), with controls across six dimensions: Identity Context, Data Protection, Audit/Logging, Agent/FM Controls, Agency Perimeters, and Orchestration [AWS Agentic AI Security Scoping Matrix]. The matrix has been adopted by OWASP and CoSAI as a reference architecture for scoping agentic security controls.

Get The Authority Brief

Weekly compliance intelligence for security leaders. Frameworks decoded. Audit strategies explained. Regulatory updates analyzed.

Need hands-on guidance? Book a free technical discovery call to discuss your compliance program.

Book a Discovery Call

Discipline in preparation. Confidence in the room.

Josef Kamara, CPA, CISSP, CISA, Security+
Josef Kamara
Josef Kamara
CPA · CISSP · CISA · Security+

Former KPMG and BDO. Senior manager over third-party risk attestations and IT audits at a top-five global firm, and former technology risk leader directing the IT audit function at a Fortune 500 medical technology company. Advises growth-stage SaaS companies on SOC 2, HIPAA, and AI governance certifications.