deepspeed

Community

Scale training with DeepSpeed efficiently.

Authorovachiever
Version1.0.0
Installs0

System Documentation

What problem does it solve?

Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism, FP16/BF16/FP8, 1-bit Adam, sparse attention.

Core Features & Use Cases

  • Guidance on setting up distributed training with DeepSpeed and ZeRO
  • DeepSpeed features: ZeRO, pipeline parallelism, mixed precision (FP16/ BF16 / FP8)
  • Advanced optimization techniques: 1-bit Adam, sparse attention
  • Debugging, best practices, and performance tuning for large-scale training

Quick Start

Install and configure DeepSpeed with a simple training script, then iterate on ZeRO stages and precision modes to achieve scalable training with reduced memory overhead.

Dependency Matrix

Required Modules

deepspeedtorchtransformersaccelerate

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: deepspeed
Download link: https://github.com/ovachiever/droid-tings/archive/main.zip#deepspeed

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository