data-engineering-data-pipeline
CommunityDesign and build robust data pipelines.
AuthorIndustrial
Version1.0.0
Installs0
System Documentation
What problem does it solve?
This Skill addresses the complexity of designing, implementing, and managing scalable, reliable, and cost-effective data pipelines for both batch and streaming data.
Core Features & Use Cases
- Architecture Design: Create ETL/ELT, Lambda, Kappa, and Lakehouse architectures.
- Implementation: Build ingestion, transformation (dbt, Spark), and orchestration (Airflow, Prefect) layers.
- Data Quality & Storage: Implement data quality frameworks and manage data in Delta Lake/Iceberg.
- Monitoring & Optimization: Ensure pipeline health and optimize costs.
- Use Case: Design a real-time streaming data pipeline to ingest user clickstream data, process it for sessionization, and store it in a Lakehouse for analytics.
Quick Start
Use the data-engineering-data-pipeline skill to design a batch ETL architecture for processing daily sales transactions from a PostgreSQL database into a Delta Lake table.
Dependency Matrix
Required Modules
None requiredComponents
scriptsreferences
💻 Claude Code Installation
Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.
Please help me install this Skill: Name: data-engineering-data-pipeline Download link: https://github.com/Industrial/rust-symphony/archive/main.zip#data-engineering-data-pipeline Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
Agent Skills Search Helper
Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.