Enterprise LLM Gateway Implementation Skill

Community

Provider-agnostic LLM gateway with routing, caching, and guardrails.

Authorngoquytuan
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill enables building a production-grade multi-provider LLM gateway with a unified OpenAI-compatible API.

Core Features & Use Cases

  • Intelligent routing across providers (OpenAI, Claude, Gemini, local models) with fallback, exponential backoff retry, and load balancing.
  • Semantic caching plus simple exact-match caching to dramatically reduce LLM usage and latency.
  • Guardrails including PII detection, content safety checks, and a plugin-based architecture for extensibility.
  • Cost tracking and per-tenant budget enforcement for real-time cost visibility.
  • Observability with structured logging, Prometheus metrics, and Grafana-ready dashboards for monitoring.
  • Use Case: Enterprise chatbots requiring high reliability, compliance, and cost discipline at scale.

Quick Start

Start by reviewing the implementation guidance in SKILL.md, install dependencies from requirements.txt, and run the FastAPI gateway locally with docker-compose as described in the deployment guide.

Dependency Matrix

Required Modules

None required

Components

Standard package

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: Enterprise LLM Gateway Implementation Skill
Download link: https://github.com/ngoquytuan/thietKeHeThongchatbot/archive/main.zip#enterprise-llm-gateway-implementation-skill

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.