hugging-face-model-trainer

Community

Train LLMs on Hugging Face Jobs, no local GPU needed.

Author3kpro
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill simplifies the process of training and fine-tuning large language models (LLMs) by leveraging Hugging Face's managed cloud infrastructure, eliminating the need for local GPU setup and complex environment management.

Core Features & Use Cases

  • Cloud-Based Training: Train models on Hugging Face Jobs infrastructure without requiring local GPUs.
  • Multiple Training Methods: Supports SFT, DPO, GRPO, and reward modeling using the TRL library.
  • GGUF Conversion: Converts trained models to GGUF format for local deployment with tools like Ollama and LM Studio.
  • Use Case: You want to fine-tune a Qwen2.5 model on a custom dataset using DPO. This Skill allows you to submit the training job directly to Hugging Face Jobs, monitor its progress, and automatically save the resulting model to the Hugging Face Hub.

Quick Start

Use the hugging face model trainer skill to fine-tune the Qwen2.5-0.5B model using SFT on the trl-lib/Capybara dataset for 3 epochs.

Dependency Matrix

Required Modules

trl>=0.12.0peft>=0.7.0transformers>=4.36.0accelerate>=0.24.0trackio

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: hugging-face-model-trainer
Download link: https://github.com/3kpro/aiarsenal/archive/main.zip#hugging-face-model-trainer

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.