LLM Tuning Patterns

Community

Master LLM fine-tuning techniques.

AuthorHermeticOrmus
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill provides expert patterns and code examples for fine-tuning Large Language Models (LLMs), addressing the complexities of efficient training, dataset preparation, and model evaluation.

Core Features & Use Cases

  • Efficient Fine-Tuning: Implement QLoRA and LoRA for memory-efficient training on consumer hardware.
  • Dataset Preparation: Format instruction datasets and understand label masking for optimal training.
  • Preference Alignment: Utilize Direct Preference Optimization (DPO) to align models with human preferences.
  • Model Evaluation: Integrate with lm-evaluation-harness for standardized benchmarking.
  • Use Case: Fine-tune a Llama-2 7B model for a specific task like customer support summarization using QLoRA, ensuring efficient use of GPU VRAM and achieving high performance.

Quick Start

Use the LLM Tuning Patterns skill to perform QLoRA fine-tuning on a Llama-2 7B model with the provided instruction dataset.

Dependency Matrix

Required Modules

transformerspefttrldatasetstorch

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: LLM Tuning Patterns
Download link: https://github.com/HermeticOrmus/LibreMLOps-Claude-Code/archive/main.zip#llm-tuning-patterns

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.