
If you've been following the rapid rise of Agentic AI, you've likely heard of the Model Context Protocol (MCP). Anthropic calls it the "USB-C for AI applications"—a standard way to plug your LLMs into your data, whether it's in Google Drive, Slack, or your internal PostgreSQL database.
It is a game-changer for utility. But it is also a game-changer for risk.
Recent research from BitSight has uncovered a startling reality: roughly 1,000 MCP servers are currently exposed to the public internet, many with zero authentication [1]. For the enterprise, this isn't just a "data leak"—it is a full-blown "agency leak."
As a Fractional CAIO, I help organizations architect AI systems that are powerful and protected. Today, I'm going to break down why this is happening, why standard firewalls aren't enough, and how I implement the leading-edge stacks from Cloudflare, Kong, and Palo Alto Networks to secure your agency.
In the old world of cybersecurity, we worried about data leakage—someone reading your files. In the new world of Agentic AI, we have to worry about "Action Risk."
MCP servers are designed to do things. They have tools like execute_command, write_file, or send_email. When these servers are exposed without strict authorization, you aren't just letting attackers see your data; you are giving them your hands.

| Traditional Risk | Agentic AI Risk |
|---|---|
| Data exfiltration | Command execution |
| Read access | Write/delete access |
| Passive observation | Active manipulation |
| Information theft | System control |
This fundamental shift requires a completely new security paradigm. Your existing data loss prevention (DLP) tools weren't designed to prevent an AI agent from executing arbitrary commands on your behalf.
The most dangerous vulnerability we're seeing right now is the "Confused Deputy" attack. Many developers use libraries like fastmcp or mcp-remote to quickly spin up a server [2]. By default, some of these tools bind to 0.0.0.0 (the public internet) and skip authentication to make development easier.
Here is the nightmare scenario: An attacker scans the web, finds your exposed MCP server on port 8080, and connects to it. They don't need to hack your password. They just ask the MCP server (the "Deputy") to execute a command. Since the server trusts its own tools, it executes the command on your internal network.

tools/listexecute_commandThe server has no concept of "who" is asking—it simply executes because it was told to. This is the essence of the Confused Deputy problem: the trusted intermediary (your MCP server) can't distinguish between legitimate requests and malicious ones.
You cannot rely on "security by obscurity." You need a defense-in-depth architecture that treats every AI agent as a user that must be verified.
In my practice, I evaluate vendors based on "Permissions Fit"—how granularly they can control what an agent can do. We aren't just looking for a firewall; we need an AI Control Plane.
Here is how the top players stack up:

| Vendor | Layer | Key Capability | Best For |
|---|---|---|---|
| Cloudflare One | Network | Zero Trust tunnels | Distributed teams |
| Kong AI Gateway | Governance | OAuth 2.1, rate limiting | Complex internal tools |
| Palo Alto Prisma | Runtime | Prompt inspection | High-stakes environments |
| Lakera Guard | Runtime | Injection detection | Real-time defense |
For distributed teams, I implement Cloudflare One. It allows us to put your MCP servers behind a "Zero Trust" tunnel. The server is never actually exposed to the internet; it's only visible to authenticated users via the Cloudflare edge. This solves the "visibility" problem identified by BitSight immediately.
# cloudflared config example
tunnel: your-tunnel-id
credentials-file: /path/to/credentials.json
ingress:
- hostname: mcp.internal.yourcompany.com
service: http://localhost:8080
originRequest:
access:
required: true
teamName: your-team
- service: http_status:404
With this configuration, your MCP server binds only to localhost, and Cloudflare handles all external access with mandatory authentication.
If you are building complex internal tools, we need Kong. Kong acts as the traffic cop for your agents. It enforces OAuth 2.1 centrally, meaning your developers don't have to write auth code for every single server.
More importantly, it allows us to rate-limit agents based on token usage, preventing a runaway bot from bankrupting your API budget.
Finally, for high-stakes environments, I deploy runtime defenders like Palo Alto Prisma AIRS or Lakera Guard. These tools sit inside the conversation. They inspect the payload of the prompt to stop "Tool Poisoning" and injection attacks before they ever reach your server.

┌─────────────────────────────────────────────────────────────┐
│ INTERNET / EDGE │
├─────────────────────────────────────────────────────────────┤
│ Layer 1: Cloudflare One (Zero Trust Tunnel) │
│ - Identity verification │
│ - Device posture checks │
│ - Geo-blocking │
├─────────────────────────────────────────────────────────────┤
│ Layer 2: Kong AI Gateway (Governance) │
│ - OAuth 2.1 enforcement │
│ - Rate limiting by token usage │
│ - Request/response logging │
├─────────────────────────────────────────────────────────────┤
│ Layer 3: Palo Alto / Lakera (Runtime Defense) │
│ - Prompt inspection │
│ - Injection detection │
│ - Tool poisoning prevention │
├─────────────────────────────────────────────────────────────┤
│ YOUR MCP SERVER │
│ - Bound to localhost only │
│ - Minimal tool permissions │
│ - Audit logging enabled │
└─────────────────────────────────────────────────────────────┘
The "USB-C for AI" is here to stay, but you shouldn't be plugging it into a dirty socket.
Implementing a secure Agentic AI stack doesn't have to slow you down. In fact, by using gateways like Kong or Cloudflare, we can actually accelerate development by offloading the heavy lifting of authentication and logging to the infrastructure.
If you are ready to move from "experimental" to "enterprise-grade," the path is clear: implement network isolation, enforce governance controls, and deploy runtime defense. Your AI agents are only as secure as the infrastructure that hosts them.

Ryan previously served as a PCI Professional Forensic Investigator (PFI) of record for 3 of the top 10 largest data breaches in history. With over two decades of experience in cybersecurity, digital forensics, and executive leadership, he has served Fortune 500 companies and government agencies worldwide.

How Apple Intelligence hallucinations exposed fragile market microstructure, and why iOS 26's Liquid Glass UI and FinanceKit API are fundamentally reshaping fintech data provenance, algorithmic trading, and the death of screen scraping.

A deep technical analysis of Notion's architectural security gaps, permission model failures, AI exfiltration vulnerabilities, and why enterprise IT leaders should look past the polished UI before adopting it as a system of record.

With DORA, NIS2, and SEC disclosure rules in full enforcement, compliance is no longer a check-the-box exercise—it's an engineering constraint. Here's how to navigate supply chain security and regulatory convergence in 2026.