pytorch-quantization
CommunityOptimize PyTorch models with INT8 quantization.
Authorcuba6112
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill addresses the need to reduce model size and accelerate inference for PyTorch models, particularly for deployment on resource-constrained environments.
Core Features & Use Cases
- INT8 Quantization: Convert floating-point models to use 8-bit integers for reduced memory footprint and faster computation.
- Post-Training Quantization (PTQ): Apply quantization to an already trained model without retraining.
- Quantization Aware Training (QAT): Train a model with quantization simulation to maintain higher accuracy.
- Use Case: Deploy a computer vision model on an edge device where memory and processing power are limited, significantly improving its performance and reducing power consumption.
Quick Start
Use the pytorch-quantization skill to demonstrate manual tensor quantization and load a pre-quantized ResNet50 model.
Dependency Matrix
Required Modules
torchtorchvision
Components
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: pytorch-quantization Download link: https://github.com/cuba6112/skillfactory/archive/main.zip#pytorch-quantization Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.