LLMOps

Community

Master LLM deployment and RAG pipelines.

Authordoanchienthangdev
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill addresses the complexities of deploying, managing, and optimizing Large Language Models (LLMs) in production environments, including prompt engineering, retrieval-augmented generation (RAG), and model evaluation.

Core Features & Use Cases

  • Prompt Management: Register, version, and A/B test prompts for optimal performance.
  • RAG Pipelines: Ingest documents, process them into chunks, and retrieve relevant information for LLM context.
  • LLM Evaluation: Frameworks for assessing model relevance, faithfulness, and other quality metrics.
  • Cost Management: Tools for tracking token usage and estimating operational costs.
  • Use Case: Deploy a customer support chatbot that leverages RAG to answer user queries based on your company's knowledge base, while continuously evaluating and improving prompt performance and cost-efficiency.

Quick Start

Use the LLMOps skill to ingest documents from the 'knowledge_base/' directory into the RAG pipeline.

Dependency Matrix

Required Modules

langchainopenaipinecone-clienttiktokenrank_bm25

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: LLMOps
Download link: https://github.com/doanchienthangdev/omgkit/archive/main.zip#llmops

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.