Enterprise LLM Gateway Implementation Skill
CommunityProvider-agnostic LLM gateway with routing, caching, and guardrails.
Software Engineering#routing#observability#enterprise#guardrails#multi-provider#gateway#semantic-cache
Authorngoquytuan
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill enables building a production-grade multi-provider LLM gateway with a unified OpenAI-compatible API.
Core Features & Use Cases
- Intelligent routing across providers (OpenAI, Claude, Gemini, local models) with fallback, exponential backoff retry, and load balancing.
- Semantic caching plus simple exact-match caching to dramatically reduce LLM usage and latency.
- Guardrails including PII detection, content safety checks, and a plugin-based architecture for extensibility.
- Cost tracking and per-tenant budget enforcement for real-time cost visibility.
- Observability with structured logging, Prometheus metrics, and Grafana-ready dashboards for monitoring.
- Use Case: Enterprise chatbots requiring high reliability, compliance, and cost discipline at scale.
Quick Start
Start by reviewing the implementation guidance in SKILL.md, install dependencies from requirements.txt, and run the FastAPI gateway locally with docker-compose as described in the deployment guide.
Dependency Matrix
Required Modules
None requiredComponents
Standard package💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: Enterprise LLM Gateway Implementation Skill Download link: https://github.com/ngoquytuan/thietKeHeThongchatbot/archive/main.zip#enterprise-llm-gateway-implementation-skill Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.