All ArticlesRuntime Truth
Category & Definitions
Definitional
What Is Agentic Ai Security

What Is Agentic AI Security? A Security-First Definition

Most AI agents running in enterprise environments today hold more access than any single workflow they execute actually requires. They act autonomously, chain actions across SaaS applications, and authenticate using embedded credentials that your IAM program has never reviewed. Understanding what is agentic AI security means accepting one uncomfortable truth first: the security model you built for human users does not apply to machines that act like insiders but appear in no insider risk program you own.

Obsidian Editorial Team
Security Research
·
Obsidian Security
·
May 13, 2026
May 14, 2026
Key Takeaways
  • Agentic AI security is not an extension of GenAI security. It governs autonomous agents that take actions, not models that generate text.
  • The core risk is effective authority: what an agent can actually do inside your SaaS stack, which is almost always broader than any configuration document shows.
  • Probabilistic agents require deterministic guardrails. Agents deviate from intended goals. Your access controls cannot.
  • The attack surface spans four layers: agent identity, MCP server connections, action chaining across apps, and orphaned or shadow agents with no owner.
  • Security teams need four capabilities in sequence: inventory, runtime visibility, guardrails, and machine identity governance.

The Agentic AI Security Definition Most Teams Are Using Is Wrong

Most security teams currently treat agentic AI security as an extension of two existing categories: GenAI security (controlling what users submit to language models) and SaaS security (governing application configurations and user access). Both framings are wrong for the same reason. They focus on the human user as the unit of risk.

GenAI security asks: what sensitive data is an employee pasting into ChatGPT? That is a real problem. It is not the agentic problem. The agentic problem is what happens after the human is no longer in the loop. An AI agent does not wait for a user to paste data. It retrieves data, processes it, passes it to another agent, writes it to an external system, and completes a multi-step workflow, all within seconds, using credentials the user never explicitly handed over.

SaaS security asks: is this application configured correctly? Again, a real problem. But configuration is not reality. An agent configured with read-only access to Salesforce may, through maker mode credential inheritance, execute with the full administrative permissions of whoever built it. The configuration says one thing. The runtime says another. Security teams chasing theoretical configuration signals without runtime evidence are ghost chasing. They are building a picture of what could happen, with no visibility into what did happen.

The framing failure matters because it shapes tooling choices. Teams that treat agentic AI security as a GenAI problem buy prompt filtering tools. Teams that treat it as a SaaS problem buy posture management tools. Neither category shows you what agents are actually doing at runtime, who they are acting on behalf of, or what their real blast radius is across your environment.

The Security-First Definition of What Is Agentic AI Security

Agentic AI security is the discipline of governing the identity, authority, and runtime behavior of autonomous AI agents operating across enterprise systems, enforcing least privilege and deterministic guardrails on systems that are probabilistic by design.

Break that definition into its load-bearing components:

Governing identity means treating every AI agent as a non-human identity (NHI) with its own credential lifecycle. Agents hold OAuth tokens, service account keys, and API credentials. NHIs already outnumber human identities 25 to 50 times in modern enterprises, and that ratio accelerates with every agent deployment. Machine identity management for AI agents means knowing which agent holds which credential, who provisioned it, and whether that credential is still attached to an active owner.

Governing authority means measuring effective authority, not theoretical configuration. Effective authority is what an agent can actually execute inside a SaaS application after all entitlements resolve. Theoretical configuration is what the policy document or platform UI says the agent should be able to do. The gap between those two numbers is where privilege escalation lives.

Governing runtime behavior means observing what agents do while they are doing it, not reviewing logs after the fact. MCP server interactions, action chaining sequences, and agent-to-agent data handoffs cannot be reconstructed from retroactive log review with any reliability. Runtime truth requires hooks into the platforms where agents execute, capturing tool calls, identity context, and data access in real time.

Deterministic guardrails for probabilistic agents is the central enforcement concept. AI agents are non-deterministic by design. They interpret instructions, make decisions, and can deviate from intended goals. Deterministic guardrails apply fixed, predictable enforcement rules regardless of what the agent decides to do. The agent may choose to request administrative data. The guardrail does not negotiate.

This definition includes what most others miss: MCP server inventory, action chaining detection, maker mode credential abuse, and the machine insider risk created when agents inherit human-level permissions without human-level oversight.

Why Agentic AI Changes the Threat Model

Pre-agentic AI security dealt with a human making a request to a model and receiving a response. The human remained the actor. The model was a tool. Four specific shifts break that model entirely.

1. The human exits the authorization loop. When a user invokes an agent, the agent takes subsequent actions autonomously. It does not re-authenticate the user for each downstream step. It uses its own embedded credentials. A user without Salesforce access can invoke a Copilot agent built in maker mode by a Salesforce administrator. That agent executes with the administrator's credentials. The user retrieves CRM records they were never provisioned to see. Nothing in your IAM flagged it. The agent did exactly what it was designed to do.

2. Agents move data at machine speed. AI agents transfer up to 16 times more data than traditional SaaS integrations. A human exfiltrating data is constrained by the speed of manual action. An agent is not. By the time a SIEM alert fires on anomalous data volume, the sequence is complete.

3. Action chaining compounds the blast radius. A single agent request can trigger a sequence: retrieve customer records from Salesforce, summarize them using an LLM, write the summary to a shared document, and send a notification via Slack. Each step in that chain touches a different system. Each step expands the blast radius. No single-platform monitoring tool sees the full chain.

4. Shadow agents and orphaned agents operate without any owner. Enterprise inventories routinely surface hundreds of Copilot agents that no one had catalogued, and thousands of agents created before any inventory existed. When the agent's creator account is disabled, the agent continues running with inherited credentials. That is an orphaned agent: a machine insider with no human owner and no expiration. Shadow AI at the agent layer is categorically more dangerous than shadow AI at the application layer because agents take actions. Applications receive data.

Composite scenario: A business analyst builds a Copilot Studio agent in maker mode using their own Microsoft 365 admin credentials. They leave the company. IT disables their account. The agent continues running, org-wide accessible, with the former admin's embedded credentials intact. A new employee invokes the agent to pull a report. The agent retrieves files tagged with Microsoft Information Protection sensitivity labels. The new employee sees data they have no clearance to access. No alert fires. No log connects the invoker's identity to the agent's authority. This is the machine insider risk that agentic AI security must address.

The Agentic AI Attack Surface, Mapped

Security teams need a structured view of where agentic AI risk actually lives. The table below maps the primary attack surface layers to OWASP and NIST AI RMF reference points.

Attack Surface LayerRisk MechanismOWASP / NIST Reference
Maker Mode Credential InheritanceAgent executes with creator's admin credentials regardless of invoker's actual permissionsOWASP LCNC-SEC-01: Account Impersonation
Orphaned Agent PersistenceAgent continues running after creator account is disabled; credentials remain validOWASP LCNC-SEC-05: Security Misconfiguration
Shadow Agent DeploymentAgents built and deployed without IT or security oversight; no inventory record existsNIST AI RMF: Govern 1.1 (AI Risk Policies)
MCP Server Connections (Unsanctioned)Agents connect to unregistered MCP servers; tools inside those servers are invisible until runtimeNIST AI RMF: Map 2.2 (AI Risk Identification)
Action ChainingSequential tool calls across multiple SaaS apps compound access and blast radius with each stepOWASP Agentic AI: Excessive Agency
Confused Deputy AttackAgent with elevated permissions is manipulated into performing actions for an unauthorized userOWASP LCNC-SEC-01: Account Impersonation
Over-Permissioned IAM RolesAgent's service account or OAuth scope exceeds what the workflow actually requiresNIST AI RMF: Manage 2.4 (Risk Treatment)
Toxic CombinationsMultiple medium-severity risk factors on one agent compound to critical priorityOWASP Agentic AI: Excessive Agency + Misconfiguration

OWASP Excessive Agency is the most directly applicable framework reference for agentic AI security. It defines the risk that arises when an AI agent is granted more capability, more access, or more autonomy than the task requires. The NIST AI RMF's Govern and Map functions address the organizational accountability gap: who is responsible for the agent, what risks has the organization identified, and what policies govern agent behavior. Neither framework, on its own, tells you what an agent is actually doing at runtime. That is the gap the security discipline must close.

For teams securing Salesforce Agentforce or Microsoft Copilot environments specifically, the toxic combination pattern is the highest-priority risk: an org-wide accessible agent, built in maker mode, with sensitive data access, whose creator account is now disabled. Each factor is medium severity in isolation. Combined, they represent a critical-priority machine insider risk with no human owner and unlimited invokers.

What Agentic AI Security Requires

Security teams that want to govern agentic AI effectively need four capabilities, in sequence. This is a framework, not a product pitch. The sequence matters because each layer is a prerequisite for the next.

1. Inventory: You cannot govern what you cannot see.

The starting point is a complete, continuously updated AI agent inventory across every platform where agents run. That inventory must answer: which agents exist, who built them, what SaaS connections they hold, what MCP servers they connect to, and whether the creator account is still active. Without this single pane of glass, every subsequent security conversation is ghost chasing. Enterprise assessments routinely surface hundreds of Copilot agents the security team did not know existed. Inventory is not optional. It is the prerequisite for all other controls.

2. Runtime Visibility: Configuration is not reality.

Once you know what agents exist, you need to know what they are actually doing. Runtime visibility means observing tool calls, data access events, and identity context as they happen, not reconstructing them from siloed platform logs after the fact. The specific question runtime visibility must answer is: what is this agent's effective authority inside each SaaS application it connects to, and does that match what the configuration says it should have? The gap between those two answers is where privilege escalation and unauthorized data access live. AI agent monitoring at the runtime layer is the only way to produce that answer.

3. Deterministic Guardrails: Probabilistic agents require fixed enforcement.

Detection without prevention is expensive logging. The third capability layer applies deterministic guardrails to agent behavior at runtime: blocking unauthorized tool calls, flagging maker mode privilege escalation, and enforcing least privilege before a sequence completes rather than alerting after it finishes. The core conceptual argument is simple. AI agents are probabilistic. They interpret instructions and make decisions that can deviate from intent. Your access controls cannot be probabilistic. Guardrails must be fixed, predictable, and enforced at the moment of action, not reviewed in a report the following morning. Runtime guardrail enforcement is generally available on Microsoft Copilot today, with expanded platform coverage on the roadmap.

4. Machine Identity Governance: Agents are insiders without an insider risk program.

The fourth capability closes the identity gap. Every AI agent is a non-human identity holding credentials, making access decisions, and moving data. No existing insider risk program covers them. Machine identity governance means applying the same lifecycle rigor to agent credentials that mature IAM programs apply to human accounts: provisioning with least privilege, continuous access review, and immediate revocation when the owning account is disabled. The AI agent governance layer connects agent identity to the broader NHI framework that IAM leaders already own, extending existing programs rather than replacing them.

Start With Inventory: You Can't Protect What You Can't See

Agentic AI security is not a feature you add to your existing security stack. It is a discipline that addresses a category of risk your current tools were not designed to see. Probabilistic agents operating with embedded credentials, chaining actions across SaaS applications, and running without active human owners represent a machine insider risk class that bypasses IAM, outruns SIEM, and operates below the visibility threshold of every platform-native monitoring tool you already own.

The sequence is clear: build the inventory, establish runtime truth, enforce deterministic guardrails, and extend machine identity governance to cover every agent credential the same way you cover every human account. Start with the inventory. You cannot protect what you cannot see, and right now, most security teams cannot see most of what is running.

If you want to understand the real blast radius of your current agent deployments, an AI agent risk assessment is the fastest way to move from theoretical configuration to runtime truth.

Frequently Asked Questions

What is agentic AI security, and how is it different from GenAI security?

GenAI security focuses on controlling what human users submit to language models, primarily through prompt filtering and data loss prevention at the input layer. Agentic AI security governs what autonomous AI agents do after the human is no longer in the loop: what credentials they use, what data they access, what actions they chain across SaaS applications, and whether any of that is policy-aligned. The human is the unit of risk in GenAI security. The agent is the unit of risk in agentic AI security.

What is a machine insider risk in the context of AI agents?

A machine insider risk arises when an AI agent holds credentials, accesses sensitive data, and makes autonomous decisions, just as a human insider would, but falls outside every existing insider risk program. Agents are not covered by behavioral monitoring designed for humans, do not trigger standard access reviews, and can persist with valid credentials long after their creator's account is disabled. The machine insider is the agent acting with effective authority that no security team has explicitly reviewed or approved.

What is maker mode, and why does it create a privilege escalation risk?

Maker mode is a configuration in platforms like Microsoft Copilot Studio where an agent is built using the creator's own credentials as the fixed authentication mechanism. Any user who invokes that agent, regardless of their own permission level, executes actions at the creator's privilege level. A user without Salesforce access who invokes a maker mode agent built by a Salesforce administrator effectively bypasses all access controls and can retrieve data they were never provisioned to see. The agent does nothing wrong. Your IAM was simply never designed to account for this delegation pattern.

What is the difference between theoretical configuration and effective authority?

Theoretical configuration is what a policy document, platform UI, or configuration file says an agent should be able to do. Effective authority is what the agent can actually execute inside a SaaS application after all entitlements, inherited credentials, OAuth scopes, and platform permissions resolve. Most security tools see theoretical configuration. Effective authority requires correlating agent configuration with live SaaS entitlements, identity context, and runtime behavior into a single picture. The gap between the two is where most agentic AI risk lives.

What are deterministic guardrails for AI agents?

Deterministic guardrails are fixed, predictable enforcement rules applied to AI agent behavior at runtime. Because AI agents are probabilistic by design, they can deviate from intended goals and request access or take actions beyond their defined scope. Deterministic guardrails do not negotiate with the agent's decision-making process. They enforce a fixed rule: if this agent requests this action under these conditions, block it. The contrast matters: probabilistic agents require deterministic guardrails precisely because you cannot rely on the agent to self-limit.