Google Cloud's Apigee includes AI gateway capabilities for managing and securing generative AI API traffic, with model routing, token-based rate limiting, content moderation, and comprehensive analytics.
LiteLLM is an open-source LLM proxy that translates OpenAI-format API calls to 100+ LLM providers. It provides a standardized interface for calling models from Anthropic, Google, Azure, AWS Bedrock, and dozens more. LiteLLM is popular as a self-hosted gateway with features like spend tracking, rate limiting, and team management.
Unified API platforms and proxies that aggregate multiple LLM providers behind a single endpoint, providing model routing, fallback, caching, rate limiting, cost optimization, and access control.