LLM Gateway
Summary
A unified gateway server for managing interactions with multiple LLM providers (OpenAI, Anthropic, DeepSeek, Gemini) with built-in cost optimization, caching, and monitoring. Features include automatic model selection based on task requirements, semantic caching to reduce redundant API calls, detailed usage analytics, and a CLI for direct interaction. The implementation focuses on reliability with comprehensive error handling, request retries, and rate limiting, making it particularly valuable for production deployments that need to balance performance and costs across multiple LLM providers.
Available Actions
No explicit actions found
This MCP server may use standard commands or have its functionality documented in the README. Check the Setup or README tabs for more information.
커뮤니티 리뷰
아직 리뷰가 없습니다. 첫 번째 리뷰를 작성해 보세요!
대화에 참여하려면 로그인하세요