Check Point acquired Lakera for $300 million. CrowdStrike added AI usage monitoring. Cloudflare launched AI Gateway with guardrails. AWS shipped Bedrock Guardrails. Every major security vendor now has an “AI Security” checkbox on their product page.

And yet none of them can tell you whether your AI agent just read .env, stored the contents in context, and exfiltrated them three tool calls later in a POST request disguised as a webhook test.

The reason isn’t that these companies lack engineering talent or resources. It’s that their architecture makes it structurally impossible to do runtime agent security. Here’s why.

1. Architecture Path Dependency

Every major security vendor’s AI security product is a variation of the same pattern: take their existing deployment model and bolt on content filtering.

VendorCore ProductAI Security Approach
Check PointNetwork firewallAdd AI content filter at network boundary
CrowdStrikeEDR/XDRAdd AI usage monitoring to endpoint telemetry
CloudflareCDN/WAFAdd guardrails to reverse proxy layer
F5Load balancer/WAFAdd prompt inspection at app gateway
AWS/AzureCloud infrastructureAdd guardrails at API call layer

Check Point didn’t pay $300M for Lakera because their DeBERTa classifier was uniquely advanced — anyone can train one. They paid for the fastest path to adding an “AI Security” line item to their firewall product for existing enterprise customers.

This is rational behavior. But it means every acquisition gets squeezed into a pass-through filtering architecture that was designed for HTTP request inspection, not agent behavior analysis.

2. The Information Boundary Problem

This is the most important structural issue, and it’s rarely discussed.

A security gateway sits between a client and an API. It sees individual requests and responses. That’s it.

What a gateway seesWhat a gateway can’t see
Request/response contentAgent session state (what it did last turn, what it plans next)
Individual tool call parametersCross-call semantic chains (read .env → POST external = theft chain)
API key patterns in textMCP tool registration changes, tool tampering
Prompt injection textAgent intent drift, persistent memory mutations, context window state

Gateways are stateless by design. Each request is processed independently, then forgotten. They fundamentally cannot do cross-request behavioral analysis.

This isn’t a limitation that can be fixed with better engineering. It’s determined by deployment position. The information simply doesn’t exist at the gateway layer.

Analogy: A highway toll booth can check every car’s license plate. But it has no idea where each car came from, where it’s going, or which other toll booths it passed through. A dashcam inside the car knows all of this.

What This Means in Practice

Consider this attack sequence:

Turn 1: Agent reads a webpage containing hidden instructions
Turn 2: Agent runs `cat ~/.aws/credentials` (looks like a config check)
Turn 3: Agent writes credentials to a variable in working memory
Turn 4: Agent sends a POST request with the credentials embedded in a "webhook test"

A gateway sees four independent, individually-normal requests. It might flag the credentials in Turn 4, but only if they appear in plaintext — and a sophisticated attack would encode them.

A runtime security layer inside the agent session sees the complete chain: external content → credential access → data staging → exfiltration. It can flag the pattern at Turn 2, before data ever leaves.

3. Business Model Priorities

Even if gateways could do runtime security, the business case doesn’t align.

DimensionGateway modelRuntime model
BuyerNetwork/security teamDev/platform team
BudgetSecurity budget (existing)Dev tools budget (must be justified)
IntegrationChange DNS/proxy configEmbed in application deployment
Sales cycleShort (existing relationship)Long (requires developer buy-in)
ScaleOne customer = entire companyProject-by-project adoption

For a company like Check Point, integrating an acquisition into the existing firewall product and selling it to existing customers is the highest-ROI path. Building a fundamentally different product that targets a different buyer in a different budget? That requires a strategic pivot few large companies are willing to make while the current approach still generates revenue.

4. The Timing Gap

The market dynamics create a structural window:

  • 2024–2025: 80% of AI deployments are chatbots and RAG pipelines. Gateway-level content filtering is sufficient. No urgent demand for runtime security.
  • 2025–2026: AI agents are growing rapidly. Runtime security needs are emerging, but the market is small. Big vendors are watching.
  • 2026–2027: Agents become mainstream. Runtime security becomes a hard requirement for production deployments.

Large vendors are rationally waiting for the market to mature before investing (or acquiring). They have money and customers — they don’t need to bet on early-stage markets.

This creates a 12–24 month window where the runtime security space is structurally underserved by incumbents.

The Runtime Alternative

What does “runtime security” actually look like?

Gateway:  Client → [Cloud Gateway] → LLM
          (sees request/response pairs only)

Runtime:  Client → [Local Proxy] → LLM
          (sees full session, tool chains, state, local environment)

A local proxy deployed alongside the agent has access to:

  • Complete session history — every turn, every tool call, every response
  • Cross-request correlation — detecting multi-step attack chains
  • Tool chain monitoring — what tools are registered, what’s been modified, what’s being called
  • Local environment context — file system access patterns, environment variables, process state
  • Real-time intervention — blocking a dangerous tool call before it executes, not after

This isn’t about having better ML models or more training data. It’s about information access. The runtime layer simply has more signal to work with than a gateway ever can.

What This Means for the Industry

The implications are clear:

  1. Gateway AI security products will converge on commodity content filtering. Every vendor will have roughly equivalent prompt injection detection and DLP scanning within 12 months. Differentiation will collapse.

  2. Runtime security will become a distinct category. Just as EDR didn’t replace firewalls but created a new category alongside them, agent runtime security will complement gateway security rather than replace it.

  3. The EU AI Act accelerates the shift. Article 15 requires detailed logging and monitoring of AI system behavior. Gateway-level request logs don’t meet this standard. Runtime-level session recording does. Data sovereignty requirements also favor local deployment over cloud gateways.

  4. Developer adoption is the moat. Gateway security is sold top-down to CISOs. Runtime security is adopted bottom-up by developers who are actually building agent applications. Different go-to-market, different competitive dynamics.

The security giants aren’t wrong to invest in gateway-level AI security. It solves real problems for chatbot and RAG deployments. But as agents become the dominant AI application pattern, the architectural limitations of the gateway model become impossible to ignore.

The toll booth catches cars without plates. The dashcam catches the ones that are stealing.


Bastion is an open-source local security proxy for AI agents. It provides DLP, prompt injection detection, tool call blocking, and full audit logging — all running locally, with zero cloud dependency.