An MCP server that analyzes the cost of LLM API calls by tracking tokens used and calculating costs based on model pricing. It provides detailed cost breakdowns and usage statistics for better budget management and optimization of AI applications.
No explicit actions found
This MCP server may use standard commands or have its functionality documented in the README. Check the Setup or README tabs for more information.
Aún no hay reseñas. ¡Sé el primero en reseñar!
Inicia sesión para unirte a la conversación