RAG Implementer
CommunityGround LLMs in your data, eliminate hallucinations.
Authordaffy0208
Version1.0.0
Installs0
System Documentation
What problem does it solves? This Skill addresses the challenge of making Large Language Models (LLMs) reliable and factual by grounding their responses in up-to-date, domain-specific, or proprietary data, thereby reducing hallucinations and providing source attribution.
Core Features & Use Cases
- Knowledge Base Design: Guides through data source mapping, intelligent chunking strategies, and metadata enrichment for effective retrieval.
- Retrieval Pipeline: Implements advanced techniques like hybrid search, re-ranking, and query enhancement to fetch the most relevant information.
- Evaluation & Monitoring: Establishes metrics for retrieval and generation quality, ensuring production-ready performance and continuous improvement.
Quick Start
Fastest path to RAG:
- Define knowledge scope
- Identify data sources
- Choose chunking strategy (500-1000 tokens)
- Add metadata for filtering
- Choose embedding model
- General:
text-embedding-3-large(OpenAI) - Code:
code-search-babbage-code-001 - Test on sample queries
- General:
- Set up vector store
- Managed: Pinecone
- Self-hosted: Weaviate or Qdrant
- Lightweight: Chroma or pgvector
Dependency Matrix
Required Modules
None requiredComponents
Standard package💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: RAG Implementer Download link: https://github.com/daffy0208/ai-dev-standards/archive/main.zip#rag-implementer Please download this .zip file, extract it, and install it in the .claude/skills/ directory.