AI agents are getting the headlines right now.
That makes sense. They are visible, new enough to feel disruptive, and increasingly able to act on behalf of people, systems, and business processes.
But they did not create a new category of identity risk.
They have mostly dragged an old one into the light.
For years, organisations have been surrounded by non-human identities that were trusted too broadly, monitored too lightly, and governed too weakly. Service accounts, scheduled tasks, application identities, API clients, certificates, bots, devices, and machine-to-machine integrations have all been sitting in the estate for a long time.
AI agents are simply the latest reminder that “who” in an enterprise is no longer limited to a human user.
The Problem Was Already Here
The security industry did not need AI to discover that non-human actors can create disproportionate risk.
We have already seen the pattern in:
- Service accounts with broad standing privilege
- Integration identities that survive long after the original project ends
- Shared secrets that are never rotated
- Certificates with no clear owner
- Devices that authenticate, connect outward, and sit in production environments for years
- Automation tooling that can trigger change at speed without enough policy control
Cheap and poorly governed IoT devices made this problem visible to the public. People understand the risk instinctively: a connected thing appears inside the environment, talks to external services, accumulates trust, and then outlives the attention it got on day one.
That same pattern exists well beyond consumer-grade IoT.
It exists in enterprise applications, cloud workloads, automation pipelines, integration platforms, and now AI-enabled systems.
AI Agents Raise the Stakes
AI agents are not just passive software components.
They can retrieve information, invoke tools, trigger workflows, and act with delegated authority. In many environments, they sit on top of existing APIs, service identities, and workflow engines.
That means an AI agent does not need to be inherently dangerous to become high risk.
It only needs:
- Excessive privilege
- Weak identity boundaries
- Access to sensitive data
- Unclear ownership
- Poor logging and review
When those conditions exist, the issue is not that AI is mysterious.
The issue is that a non-human actor has been allowed to operate without enough governance around it.
The Common Failure Pattern
Across devices, services, workloads, and AI agents, the failure pattern is remarkably consistent.
- The non-human identity is introduced to solve a practical problem quickly.
- It is granted broad access because fine-grained design takes longer.
- Ownership becomes blurred once the project moves on.
- Credentials or tokens persist longer than anyone intended.
- Monitoring focuses on uptime and functionality rather than trust and accountability.
Eventually, the organisation discovers that something non-human can access far more than it should, with too little evidence about what it has done and too few people who can clearly explain why.
That is not an edge case.
It is a governance gap.
flowchart LR
A["Devices / IoT"] --> G["Non-Human Identity Risk"]
B["Service Accounts"] --> G
C["Workload Identities"] --> G
D["API / Integration Accounts"] --> G
E["Automation / Bots"] --> G
F["AI Agents"] --> G
G --> H["Excessive Privilege"]
G --> I["Weak Ownership"]
G --> J["Long-Lived Credentials"]
G --> K["Poor Monitoring"]
G --> L["Hidden Trust"]
Zero Trust Has To Include Non-Humans
Zero Trust is often explained in terms of users, devices, and applications.
That is necessary, but incomplete.
A serious Zero Trust model must also ask:
- Which non-human identities exist?
- What are they allowed to do?
- Who owns them?
- What credentials, certificates, or tokens do they rely on?
- What evidence remains after they act?
If those questions cannot be answered quickly, the environment is carrying hidden trust that has never really been challenged.
This is why non-human identities should be treated as first-class security objects, not background implementation detail.
What Good Looks Like
The answer is not to fear automation, AI, or connected devices.
The answer is to govern them properly.
That starts with a few practical disciplines:
- Maintain an inventory of non-human identities, including devices, workloads, integrations, service accounts, and automation tools
- Assign clear ownership to each identity and review it regularly
- Replace broad standing access with least-privilege permissions
- Reduce reliance on shared secrets and long-lived credentials where possible
- Rotate certificates, keys, and tokens with discipline
- Segment access and constrain what non-human actors can reach
- Log actions in ways that support real accountability and review
- Design architecture so non-human identities are visible to governance, not hidden inside technical debt
None of that is exotic.
It is just identity hygiene applied to the parts of the environment many organisations still struggle to see clearly.
From AI Concern to Identity Discipline
It is reasonable that AI agents are forcing this discussion.
They are powerful, fast-moving, and increasingly close to decision-making and operational workflows. They make the consequences of weak non-human identity governance easier to imagine.
But the lesson should not be that AI has created a whole new problem category.
The lesson is that many organisations already had a non-human identity problem across devices, workloads, integrations, and automation. AI agents have simply made that problem harder to ignore.
The organisations that respond well will not treat this as a temporary AI scare.
They will use it as the prompt to govern every non-human actor with the same seriousness they apply to human access.
This is not an emerging problem.
It is an existing governance gap that AI has made impossible to ignore.