RAG Implementer

Community

Ground LLMs in your data, eliminate hallucinations.

Authordaffy0208
Version1.0.0
Installs0

System Documentation

What problem does it solves? This Skill addresses the challenge of making Large Language Models (LLMs) reliable and factual by grounding their responses in up-to-date, domain-specific, or proprietary data, thereby reducing hallucinations and providing source attribution.

Core Features & Use Cases

  • Knowledge Base Design: Guides through data source mapping, intelligent chunking strategies, and metadata enrichment for effective retrieval.
  • Retrieval Pipeline: Implements advanced techniques like hybrid search, re-ranking, and query enhancement to fetch the most relevant information.
  • Evaluation & Monitoring: Establishes metrics for retrieval and generation quality, ensuring production-ready performance and continuous improvement.

Quick Start

Fastest path to RAG:

  1. Define knowledge scope
    • Identify data sources
    • Choose chunking strategy (500-1000 tokens)
    • Add metadata for filtering
  2. Choose embedding model
    • General: text-embedding-3-large (OpenAI)
    • Code: code-search-babbage-code-001
    • Test on sample queries
  3. Set up vector store
    • Managed: Pinecone
    • Self-hosted: Weaviate or Qdrant
    • Lightweight: Chroma or pgvector

Dependency Matrix

Required Modules

None required

Components

Standard package

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: RAG Implementer
Download link: https://github.com/daffy0208/ai-dev-standards/archive/main.zip#rag-implementer

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository