No items found.

Overhaul of Agent Gateway supporting A2A, MCP, and Kubernetes Gateway API

Today, we’re excited to share the next major milestone: Agent Gateway is now a full-featured, AI-native gateway that combines deep MCP and A2A protocol awareness, robust traffic policy controls, inference gateway support, Kubernetes Gateway API support, and unified access to major LLMs, all purpose-built with Rust for real-world agentic systems.

Back when we first introduced Agent Gateway, it was designed to fill a critical gap in the AI stack: enabling structured, secure, and scalable communication between agents, tools, and LLMs using protocols like MCP and A2A. We wrote about this in our last blog post where we explained why traditional API gateways fall short in agentic environments. Since then, the project has grown tremendously.

Today, we’re excited to share the next major milestone: Agent Gateway is now a full-featured, AI-native gateway that combines deep MCP and A2A protocol awareness, robust traffic policy controls, inference gateway support, Kubernetes Gateway API support, and unified access to major LLMs, all purpose-built with Rust for real-world agentic systems.

And notably, Agent Gateway now fully supports:

Let’s take a closer look at the updated capabilities.

Deep Protocol Awareness: MCP and A2A

Agent Gateway continues to deepen its native support for the emerging Model Context Protocol (MCP) and Agent-to-Agent Protocol (A2A) with updated support including:

  • Protocol-aware routing, telemetry, and tracing
  • Support for MCP server/tool aggregation (virtualized MCP servers)
  • Updated MCP Authorization support per the June 2025 spec
  • Native authorization policy engine using Cedar for fine-grained authorizations
  • Support for exposing local stdio-based MCP servers as remote targets

Agent Gateway can be used to implement authentication and authorization for your MCP servers with minimal set up. We will keep Agent Gateway updated as these specs continue to evolve.

Kubernetes Gateway API Support

Agent Gateway now implements the Kubernetes Gateway API with support for HTTPRoute, GRPCRoute, TCPRoute, and TLSRoute. This isn’t just basic support, we’ve implemented all core and extended features, and most experimental ones too. This means you can:

  • Match traffic based on headers, paths, and query params
  • Apply CORS, redirects, header rewrites, request mirroring
  • Configure timeouts, retries, and direct responses

For platform teams already using Kubernetes-native tooling, this makes Agent Gateway a seamless drop-in with advanced L7 control. It’s a critical step toward integrating agentic workloads into the broader platform ecosystem.

Fine-Grained Traffic Policy Controls

In real enterprise deployments, security and reliability aren't optional. That’s why Agent Gateway now includes advanced policy features:

  • Local and remote rate limiting
  • JWT authentication and external auth, (ie, extAuthz) hooks
  • Upstream auth (cloud identity, TLS, etc)
  • Full OpenTelemetry support for metrics, logs, and distributed tracing

These controls allow you to run production-grade agent infrastructure that meets enterprise security and observability requirements.

LLM Gateway

Agent Gateway can now be used as an “AI Gateway” and route traffic directly to LLMs (OpenAI, Anthropic, Gemini, Bedrock, etc). Top usecases here are unified LLM API, prompt guarding, and resilience (failover, rate limiting, etc).

Agent Gateway now provides a unified OpenAI-compatible API across:

  • OpenAI / Azure OpenAI
  • Anthropic
  • Google Gemini
  • Amazon Bedrock
  • Google Vertex

This enables users to seamlessly move between providers without changes to their application,  even dynamically based on the health or performance of a specific provider.

Agent gateway can also be used to implement prompt guarding to scan/filter for sensitive information or full on direct prompt injection attacks:

  • Regex-based filters to block prompts with known unsafe patterns
  • Optional webhook integration to run pre-flight validation using custom logic

Combined with the fine-grained traffic policy controls, Agent Gateway can act as a powerful LLM Gateway.

Inference Routing

If you are running your own models on self-hosted GPU infrastructure, Agent Gateway now implements the Inference Gateway extensions for more accurate, efficient prompt routing. Using the InferencePool API, you can route based on

  • Prompt criticality
  • GPU and KV cache utilization
  • Work queue / waiting queue depth
  • Lora adapters

This gives AI platform operators more efficient and cost-conscious routing. This can also form the foundation for deeper optimizations like llm-d.

Get Involved

We’re thrilled to see the community interest grow around this project. If you’re building agent infrastructure, dealing with LLM routing, or trying to enforce policy in AI-native environments, we’d love your feedback and contributions:

Agent Gateway has evolved far beyond just a proxy. It’s becoming the control point for secure, scalable agentic systems. We’re just getting started.

Cloud connectivity done right