hugging-face-model-trainer
CommunityTrain LLMs on Hugging Face Jobs, no local GPU needed.
Author3kpro
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill simplifies the process of training and fine-tuning large language models (LLMs) by leveraging Hugging Face's managed cloud infrastructure, eliminating the need for local GPU setup and complex environment management.
Core Features & Use Cases
- Cloud-Based Training: Train models on Hugging Face Jobs infrastructure without requiring local GPUs.
- Multiple Training Methods: Supports SFT, DPO, GRPO, and reward modeling using the TRL library.
- GGUF Conversion: Converts trained models to GGUF format for local deployment with tools like Ollama and LM Studio.
- Use Case: You want to fine-tune a Qwen2.5 model on a custom dataset using DPO. This Skill allows you to submit the training job directly to Hugging Face Jobs, monitor its progress, and automatically save the resulting model to the Hugging Face Hub.
Quick Start
Use the hugging face model trainer skill to fine-tune the Qwen2.5-0.5B model using SFT on the trl-lib/Capybara dataset for 3 epochs.
Dependency Matrix
Required Modules
trl>=0.12.0peft>=0.7.0transformers>=4.36.0accelerate>=0.24.0trackio
Components
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: hugging-face-model-trainer Download link: https://github.com/3kpro/aiarsenal/archive/main.zip#hugging-face-model-trainer Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.