deepspeed
CommunityScale training with DeepSpeed efficiently.
Authorovachiever
Version1.0.0
Installs0
System Documentation
What problem does it solve?
Expert guidance for distributed training with DeepSpeed - ZeRO optimization stages, pipeline parallelism, FP16/BF16/FP8, 1-bit Adam, sparse attention.
Core Features & Use Cases
- Guidance on setting up distributed training with DeepSpeed and ZeRO
- DeepSpeed features: ZeRO, pipeline parallelism, mixed precision (FP16/ BF16 / FP8)
- Advanced optimization techniques: 1-bit Adam, sparse attention
- Debugging, best practices, and performance tuning for large-scale training
Quick Start
Install and configure DeepSpeed with a simple training script, then iterate on ZeRO stages and precision modes to achieve scalable training with reduced memory overhead.
Dependency Matrix
Required Modules
deepspeedtorchtransformersaccelerate
Components
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: deepspeed Download link: https://github.com/ovachiever/droid-tings/archive/main.zip#deepspeed Please download this .zip file, extract it, and install it in the .claude/skills/ directory.