llm-fine-tuning

Community

Fine-tune LLMs efficiently

AuthorBagelHole
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill provides the infrastructure and code to fine-tune Large Language Models (LLMs) using various techniques like QLoRA, LoRA, and full fine-tuning, enabling customization for specific domains and tasks.

Core Features & Use Cases

  • QLoRA/LoRA Fine-Tuning: Efficiently fine-tune large models on consumer-grade GPUs.
  • Distributed Training: Supports DeepSpeed and FSDP for multi-node, multi-GPU training.
  • Dataset Preparation: Guides on preparing datasets for instruction tuning, DPO, or RLHF.
  • Model Export: Merging LoRA adapters and exporting models for serving.
  • Use Case: Fine-tune an open-source LLM on your company's internal documentation to create a specialized chatbot that can answer employee questions accurately.

Quick Start

Use the llm-fine-tuning skill to set up QLoRA fine-tuning for a Llama 3.1 model using the provided Axolotl configuration.

Dependency Matrix

Required Modules

transformersdatasetstrlpeftbitsandbytesacceleratetorchwandbaxolotl

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: llm-fine-tuning
Download link: https://github.com/BagelHole/DevOps-Security-Agent-Skills/archive/main.zip#llm-fine-tuning

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.