Rust Performance
L2 + L3 built with axum + tokio. 599 req/s on cache-hit path. 18ms P50. Zero-copy SSE streaming. No Python wrappers in the hot path.
The self-hosted gateway that unifies API traffic, LLM calls, and autonomous agents in one control plane.
Three layers. One audit trail. Zero data leaving your network.
Platform engineers managing AI workloads are running separate tools for the edge, the LLM layer, and agents. Three auth systems. Three dashboards. Three 2am incidents.
Teams calling LLM APIs with no governance. A runaway prompt loop burns $50K over a weekend. Finance cannot forecast AI spend.
Kong for the edge. LiteLLM in the middle. A separate agent runtime. Three repos, three auth configs, three observability stacks.
Autonomous agents calling tools with no allowlists, no session budgets, no human approval for high-risk operations. MCP + A2A are here, but who governs them?
There's a better way.
Click a layer to explore. All three share tenant_id, Redis, ClickHouse, and OTel traces.
Battle-tested L7 proxy handling all inbound traffic. 50k TPS at P99 < 50ms on production hardware. Per-tenant rate limiting, JWT auth, and every request logged to ClickHouse.
rl:{tenant_id}:{route}:{window}Not marketing claims. Measured results from real deployments.
L2 + L3 built with axum + tokio. 599 req/s on cache-hit path. 18ms P50. Zero-copy SSE streaming. No Python wrappers in the hot path.
tenant_id in every API call, every log entry, every cache key. Rate limit buckets, token budgets, tool allowlists — all tenant-isolated from day one. Not bolted on.
One span tree from APISIX → AI Gateway → agent tool call. See an HTTP request become an LLM call become a tool invocation. OTel + Jaeger. No competitor has this.
MCP + A2A protocol support. HITL gates for high-risk tool calls. Tool allowlists (deny-by-default). Session budgets. Tool poisoning detection. A2A loop detection.
Docker Compose to start. Kubernetes for production. Air-gap capable. Your data never leaves your infrastructure. HIPAA, GDPR, ISO 42001 ready.
Apache 2.0. No "community edition" limitations. No license traps. No open-core bait-and-switch. Every feature ships in the open-source build.
Not every tool does every job. Here is what each platform actually ships.
| Feature | gatez | Kong | Portkey | agentgateway |
|---|---|---|---|---|
| API Gateway (L1) | ✓ | ✓ | ✕ | ✕ |
| AI Gateway (L2) | ✓ | Partial | ✓ | ✕ |
| Agent Gateway (L3) | ✓ | ✕ | ✕ | ✓ |
| All 3 layers unified | ✓ | ✕ | ✕ | ✕ |
| Cross-layer traces | ✓ | ✕ | ✕ | ✕ |
| Multi-tenant native | ✓ | Partial | ✕ | Partial |
| Self-hosted / air-gap | ✓ | ✓ | ✕ | ✓ |
| MCP + A2A protocols | ✓ | ✕ | ✕ | ✓ |
| HITL gates | ✓ | ✕ | ✕ | ✕ |
| Per-tenant token budgets | ✓ | Enterprise | ✕ | ✕ |
| Two-portal control plane | ✓ | Partial | ✕ | ✕ |
| API lifecycle governance | ✓ | Partial | ✕ | ✕ |
| Open source | ✓ | Partial | ✕ | ✕ |
See full comparisons:
Clone, start, call. That request goes through rate limiting (L1), PII redaction (L2), and token budget enforcement (L2). All logged. All traced.
Stop shadow AI from burning your budget. Token budgets, semantic caching, per-tenant cost analytics.
Learn moreGive agents tools. Keep humans in control. HITL gates, tool allowlists, session budgets, A2A policies.
Learn moreAI governance that satisfies your auditor. Full air-gap, PII redaction, complete audit trail, no data leaves your network.
Learn moreMeter, limit, and bill every API call per tenant. Self-service developer portal, API key workflows, usage export for billing.
Learn moreApache 2.0 licensed. No open-core. No feature gates. Every feature ships in the open-source build. The code is public, so we have no incentive to lie about features.
On-prem deployment assistance. Custom Keycloak/SSO integration. SLA guarantees. Architecture review. Dedicated support channel.