James White, CalypsoAI CTO
At the halfway point of 2025, it’s timely to take a snapshot of enterprise AI adoption and the state of AI security.
Enterprise GenAI usage is now standard
Across the board, enterprises have at least a base level understanding of GenAI. They understand what it is, and where it can be put to use. If you’re in a job that doesn’t have access to GenAI, you’re behind the curve right now.
Building applications on top of GenAI is quickly becoming standard within enterprises. First-hand, we see CalypsoAI customers are deploying AI apps to production, both internally and externally.
When it comes to agent adoption, there are two strands: adoption through third-parties and building proprietary agents. Agent adoption through third parties is quite well advanced, but companies don’t always know they’ve done it. Organisations that started with a GenAI model interface in a browser may be unaware that they’re also using agentic AI.
For example, when Gemini launched, it didn’t have any agentic elements built in; now Deep Research is an agent within Gemini. When it’s given a prompt, it figures out what tasks to do, which web searches to do, brings back information and creates reports. Likewise, there are agents within Microsoft 365 Copilot.
Building proprietary agents is mainly at the prototyping and experimentation stage, but agents are moving much faster than GenAI did. In 2024, the questions were: Is this going to happen? Is there really a difference in risk? Now those questions are gone. Enterprises have no doubts.
That’s why understanding agents and agentic resilience is so important.
MCP is not shorthand for security
Many enterprises have begun to think of Model Context Protocol (MCP) as the way of introducing agentic AI. Organisations are almost replacing the word ‘agent’ with MCP, they talk about ‘building an MCP’.
That’s not strictly accurate. MCP effectively represents the menu of options that an agent can use to do things. Think of it as the MCP server saying ‘here’s what’s on the menu’, and the MCP client says ‘I’m going to use that one in this way’. It then uses an AI model to figure out what to do and how.
In that structure, risk still stems from the model. Other risks sit within the server software. Those risks aren’t automatically removed by MCP. The protocol is very good but it is by no means secure by default, as we saw in recent security incidents. For example, Asana recently identified and disclosed a bug that could have exposed data belonging to Asana MCP users to users in other accounts.
By assuming that MCP minimizes risk, enterprises may actually increase their risk exposure. A lot of our customers and prospects are now asking, ‘Tell me how you secure MCP’. They need to take a holistic view that considers and stress-tests the use case, the model, deployment patterns, and interactions with tools via MCP and other standards.
With agentic AI, the ‘how’ is crucial
Most companies using AI over-index on what an AI model or application achieves. They are looking at the work product, the output. With agents working autonomously, the risk levels are much higher, and the how becomes incredibly important.
That’s because there are many ways to achieve a what but a lot of those ways are appropriate or incorrect. Imagine an agent is tasked with keeping a database up to date, and has access and permissions to delete or insert data. It could delete entries relating to CalypsoAI, for example, by accurately finding and removing the exact matches of the company name.
However, the agent could equally decide to issue an instruction to ‘delete C*’ records, deleting records of all companies beginning with C. This crude action would achieve the same goal, but with a cascade of unintended consequences that may not be easily remediated. In the age of agents, when actions are driven by non-deterministic models, unintentional behavior is the breach – especially if safeguards are inadequate. Understanding agent behaviour is absolutely critical to their successful adoption.
The enterprise security stack is struggling
Securing AI represents a whole new way of operating, not just some slight changes to existing security solutions. All of the traditional software that companies use will, over time, start morphing into other manifestations. And how you protect those is not completely clear right now.
For organisations that have a traditional stack of security, there is already evidence that it’s creaking at the seams. That’s because they were designed and built for a pre-AI era. In the AI era, the volume of net new information being generated - ranging from garbage to high-quality, mission-critical data - is exploding.
The systems that were looking at human-generated traffic for the past several decades are now coming to terms with humans-using-AI generated traffic. That already represents a huge multiplier. Now imagine when it is agents-generating-information traffic. It just goes exponential.
The traditional security stack is not built for that reality; it will have to change very dramatically to keep up - and sooner rather than later.
A glimpse at the future: atomic agents
Today, an agent requires an AI model to act as its brain. That brain usually sits centrally somewhere and the agent framework asks questions of the model and the answers come back. That’s the cycle, over and back between the agent and the model.
If an agent goes rogue in that scenario, an organisation can apply a simple networking rule to block the agent from access to the model. In the worst-case scenario, the model can be turned off.
In future, models will get small enough that they don’t require GPUs and won’t have to sit in server rooms or in the cloud. The ‘brain’ will run in the same place as the agent, so the agent can efficiently operate anywhere and scale up and down. When that happens, security risk explodes.
The big concept here is atomic agents, and there is evidence they are coming. Microsoft’s BitNet b1.58 2B4T, a 1-bit model that runs on CPU, shows what’s possible. It’s not small enough, not fast enough, and probably not good enough - yet. But ‘yet’ becomes reality very quickly in this industry.