In this blog post How to Evaluate Agent Platforms in 2026 with Identity First in Mind we will look at why agent identity is the first security issue most enterprises need to solve, why silent failures worry me more than obvious outages, and how I think technology leaders should evaluate agent platforms in 2026.
After more than 20 years in enterprise IT, I have learned that the biggest risks rarely arrive with dramatic alarms. They show up as small design shortcuts that seem harmless in a pilot. In agentic AI, the shortcut I keep seeing is this: teams focus on what the agent can do before they decide who the agent is, what it is allowed to do, and how anyone will know when it quietly does the wrong thing.
At a high level, an enterprise agent is not just a chatbot with a nicer interface. It is a language model connected to instructions, memory, tools, APIs, and business systems so it can take actions, not just generate text. That shift from answering to acting is exactly why identity matters so much more than many teams first expect.
From Melbourne, working with organisations across Australia and internationally, I see the same pattern in Azure, Microsoft 365, OpenAI, Claude, and broader cloud environments. The technology is moving quickly, but the architectural lesson is surprisingly old-fashioned: if you cannot name the actor, control its privileges, and trace its actions, you do not really control the system.
The technology behind agentic AI in plain language
It helps to strip the jargon back. Most enterprise agents are built from five moving parts: a model, a set of instructions, a memory or state layer, a tool layer, and an orchestration layer. The orchestration layer decides when the model should answer directly, when it should call a tool, and when it should hand work to another agent or workflow.
That sounds abstract, so here is the practical version. An agent might read a policy document from SharePoint, query a CRM record, check a calendar, create a draft in Microsoft 365, and then ask a human for approval. In other words, it behaves less like search and more like a junior digital worker operating across systems.
Under the hood, most modern platforms now rely on standard identity patterns such as OAuth, OpenID Connect, scoped tokens, role-based access control, and increasingly non-human or workload identities. That is good news. It means agent security should not be treated as magic. It should be treated as identity architecture, policy design, logging, and resilience engineering.
Why agent identity is the first security problem to solve
One lesson from cybersecurity has held up for years: privilege is risk. The moment an agent can open files, call systems, or trigger actions, it becomes a privileged actor. If the enterprise has not assigned it a distinct identity, it usually ends up borrowing one through shared service accounts, embedded secrets, or broad delegated permissions.
That is where the trouble starts. Shared accounts destroy accountability. Embedded secrets leak into code, prompts, notebooks, pipelines, and support logs. Broad delegated permissions make it almost impossible to explain what the agent was truly allowed to do at the moment something went wrong.
In my experience, leaders often assume the model is the main security issue. Usually it is not. The larger issue is the identity and permission boundary around the tools the model can reach.
This is also where Australian cyber discipline becomes useful. The Essential Eight mindset already taught many teams to take service accounts, privileged access, and managed credentials seriously. Agent identities deserve the same treatment: unique identity, least privilege, managed lifecycle, auditable use, and no casual reuse of credentials across environments.
If I had to reduce agent security to one sentence, it would be this: every meaningful enterprise agent should have its own governed identity, not borrowed trust.
Why silent failures are the hidden risk
Traditional software often fails loudly. A server crashes, an API returns an error, or a dashboard goes red. Agents are different because they can fail in ways that still look plausible to the user.
That is why I call silent failures the hidden risk in enterprise agentic AI. The user receives an answer. The workflow completes. Nobody opens an incident. But underneath, the agent may have skipped a tool, used stale context, lost a handoff, hit a permission boundary, or quietly guessed when it should have stopped.
I have seen versions of this in pilots and production-like environments. The most dangerous outcomes are not spectacular hallucinations. They are confident partial truths delivered with a professional tone.
Common silent failure patterns include:
The agent cannot access a system and gives a best-effort answer instead of surfacing a clear access failure.
A tool times out, but the model fills in the gap with an assumption.
A handoff between agents loses a critical instruction, approval state, or user context.
A safety or policy control blocks one step, but the application reports only a generic response.
A fallback model or alternate workflow behaves differently from the tested path.
The agent uses the wrong tool with the right intent, producing believable but incorrect output.
For executives, the business problem is simple. Silent failures damage trust before they trigger security alerts. They show up as wrong summaries, missing evidence, privacy mistakes, inconsistent recommendations, and approvals made on incomplete information.
What I look for in agent platforms in 2026
As a solution architect and enterprise architect, I have become much less interested in polished demos. In 2026, I think platforms should be judged less by how impressive the agent looks in a workshop and more by how well the platform handles identity, traceability, and controlled failure.
Here are the areas I would evaluate first.
1. First-class agent identity
The platform should support distinct non-human identities for agents, not just API keys and generic service accounts. I want clear support for scoped tokens, managed identities or federation where possible, and separation between agent identity and end-user identity.
I also want lifecycle controls. Who owns the agent identity, who approves access, how is it reviewed, and what happens when the agent is retired?
2. Fine-grained authorization
Least privilege matters more with agents than with humans because agents operate at machine speed. A platform should let me restrict permissions per tool, per environment, and ideally per action type. Reading a document should not automatically imply sending email, updating records, or triggering downstream workflows.
High-risk actions should support step-up controls such as approval gates, human confirmation, or just-in-time access. If the platform treats every tool call as equally safe, that is a red flag for me.
3. Real observability, not just chat history
Chat transcripts are not enough. I want traces showing prompts, tool selection, tool arguments, handoffs, retries, latency, blocked actions, and final outputs. If an agent gave the wrong answer, the team should be able to reconstruct what happened without guesswork.
This is one of the biggest ecosystem shifts I have noticed recently. The better platforms are moving beyond simple logs and into tracing, evaluation, and operational telemetry because they know enterprise agents are too nondeterministic to manage any other way.
4. Explicit handling of failure
A good platform should make failure visible. Tool failures, authorization failures, partial results, and policy blocks should be first-class events. The platform should not quietly smooth over important control failures in the name of user experience.
I would ask one direct question: when the agent cannot safely complete a task, does it stop clearly, or does it improvise? The answer tells you a lot about production risk.
5. Governance for connectors and tool ecosystems
In 2026, the connector layer is becoming one of the most important attack surfaces. Whether the platform uses native connectors, plugins, or MCP-style tool integration, I want allowlists, version control, environment separation, and clear trust boundaries around third-party tools.
If anyone can attach a new tool with broad permissions and limited review, the platform is not enterprise-ready. It is simply powerful.
6. Privacy, data handling, and Australian compliance fit
If the agent touches personal information, leaders need to ask harder questions. Where is state stored, how long is it retained, what can be exported to SIEM or compliance tooling, and how are deletion and retention policies handled?
For Australian organisations, I also think the privacy conversation needs to mature. The OAIC has been clear that organisations using AI in decision-support contexts need to understand how outputs are produced, verify accuracy, and ensure meaningful human oversight where decisions affect people. If your agent platform makes that explanation difficult, governance will eventually become painful.
7. Evaluation and regression discipline
Every serious platform now needs a way to test agents continuously, not just once before launch. I want support for repeatable evaluations across tool choice, argument accuracy, policy adherence, hallucination risk, prompt injection resistance, and handoff quality.
This matters because agent behaviour changes as prompts evolve, tools change, models are updated, and policies shift. If the platform has no strong eval story, the burden quietly falls back on your engineering team.
A practical pattern I trust
When I review enterprise designs, I usually look for a pattern closer to this:
Agent identity
- one named identity per agent or agent class
- managed or federated credentials where possible
- no secrets stored in prompts or notebooks
Access model
- read only by default
- separate permissions for read, draft, approve, and execute
- high risk actions require human approval
Operations
- full tracing of prompts, tool calls, retries, and failures
- logs exported to existing monitoring and security workflows
- regular access reviews and evaluation runs
It is not glamorous, but it works. As a published author, I have learned that the same idea applies to both writing and architecture: clarity beats cleverness when the stakes are high.
The strategic question for tech leaders
My view is that enterprises should stop asking whether an agent platform is powerful and start asking whether it is governable. In 2026, impressive reasoning is becoming easier to find. Controlled identity, observable behaviour, and trustworthy failure handling are still much rarer.
The organisations that get this right will not be the ones with the flashiest demos. They will be the ones that treat agents as digital actors inside the security model, not as smart features sitting outside it.
That is the real shift I think leaders need to absorb this year. The more capable agents become, the less sensible it is to govern them like chatbots. The better question now is not whether an agent can act, but whether your architecture can prove who acted, with what authority, and what really happened when the outcome looked fine but was not.
- MCP A2A OpenTelemetry and OAuth Every Architect Must Track in 2026
- Enterprise AI Agents Need Standards Before They Need Scale in 2026
- Anthropic’s DoD stance just changed what “safe” enterprise AI means
- OpenAI Agents SDK vs LangGraph in 2026 What CIOs should standardise on
- OpenAI Hosted on Azure What Microsoft Really Means for Enterprises