Your Most Dangerous User Is Not Human: How AI Agents and MCP Servers Broke the Internal API Walled Garden

Highlights

  • The Perimeter is Porous: Modern Agentic AI and the Model Context Protocol (MCP) have effectively turned internal data centers inside out, making the “internal API” security model obsolete.
  • The “Confused Deputy” Risk: Legitimate AI agents act as trusted internal entities but can be exploited to bypass Data Loss Prevention (DLP) policies, as seen in recent Microsoft Office vulnerabilities.
  • Beyond the WAF: Traditional WAFs and API Gateways are blind to lateral “East-West” traffic and cannot detect the subtle behavioral anomalies inherent in AI-to-API interactions.
  • Salt’s Three-Pillar Defense: To secure the Agentic Action Layer, organizations need continuous discovery, adaptive governance, and intent-based behavioral protection.

Last month, Microsoft quietly confirmed something that should keep every CISO up at night.

As first reported by BleepingComputer and later detailed by TechCrunch, a bug in Microsoft Office allowed Copilot, the AI assistant embedded in millions of enterprise environments, to summarize confidential emails and hand them to users who had no business seeing them. Sensitivity labels? Ignored. Data loss prevention (DLP) policies? Bypassed entirely.

This wasn’t the work of a hacker or malware. This was a trusted internal tool doing exactly what it was designed to do: processing data. The AI didn’t break in. It was already inside.

The Illusion of the Internal Safe Zone

For years, security teams have operated under a comforting assumption: internal APIs are safe because they sit behind the gateway. We challenged this myth in our latest Field Guide, but the Microsoft incident proves the reality is far more volatile.

When you deploy an AI agent, you are handing a highly privileged entity the keys to your internal data. You are trusting it to respect every access policy, sensitivity label, and permission boundary you have built. When it doesn’t: when it incorrectly processes a context or misreads a label, there is no alarm. No blocked requests at the edge. Just sensitive data, silently served to the wrong person.

The “Confused Deputy” Is Already on Your Payroll

Security researchers call this the confused deputy problem. It occurs when a trusted entity with legitimate access is tricked (or simply misconfigured) into acting against your interests.

With the rise of the Model Context Protocol (MCP), this problem is about to get dramatically worse. MCP is the “USB-C for AI,” designed to let agents plug into any internal data source with universal ease. For productivity, it is a breakthrough. For security, it is a nightmare: every MCP connection is a new pipeline that bypasses your perimeter entirely.

A developer spins up an MCP server to let an AI agent query a customer database. That agent now has a direct, authenticated connection to sensitive data. It does not traverse your API gateway. It does not pass through your WAF. It just talks to the data, deep inside your network, in a conversation your security stack never sees.

Why Your WAF Is Watching the Wrong Door

Here is the uncomfortable truth: your WAF and API gateway were built for a world that no longer exists.

They analyze North-South traffic: requests coming in from the outside world. They are excellent at catching known attack signatures hitting your front door. But the Microsoft Copilot bug didn’t come through the front door. It happened in the hallways.

East-West traffic: the lateral communication between microservices, AI agents, and data stores, is where the real risk lives now. Traditional perimeter tools are completely blind to it. The Copilot vulnerability wasn’t a malicious payload; it was a context validation failure. No signature to detect. No anomaly at the edge. By the time anything could have been flagged, the data was already exposed.

Securing the Conversations You Can’t See

Stopping these risks requires a fundamentally different approach: one that moves past perimeter defense and into the Agentic Action Layer where AI agents actually operate.