data-engineering-data-pipeline

Community

Design and build robust data pipelines.

AuthorIndustrial
Version1.0.0
Installs0

System Documentation

What problem does it solve?

This Skill addresses the complexity of designing, implementing, and managing scalable, reliable, and cost-effective data pipelines for both batch and streaming data.

Core Features & Use Cases

  • Architecture Design: Create ETL/ELT, Lambda, Kappa, and Lakehouse architectures.
  • Implementation: Build ingestion, transformation (dbt, Spark), and orchestration (Airflow, Prefect) layers.
  • Data Quality & Storage: Implement data quality frameworks and manage data in Delta Lake/Iceberg.
  • Monitoring & Optimization: Ensure pipeline health and optimize costs.
  • Use Case: Design a real-time streaming data pipeline to ingest user clickstream data, process it for sessionization, and store it in a Lakehouse for analytics.

Quick Start

Use the data-engineering-data-pipeline skill to design a batch ETL architecture for processing daily sales transactions from a PostgreSQL database into a Delta Lake table.

Dependency Matrix

Required Modules

None required

Components

scriptsreferences

💻 Claude Code Installation

Recommended: Let Claude install automatically. Simply copy and paste the text below to Claude Code.

Please help me install this Skill:
Name: data-engineering-data-pipeline
Download link: https://github.com/Industrial/rust-symphony/archive/main.zip#data-engineering-data-pipeline

Please download this .zip file, extract it, and install it in the .claude/skills/ directory.
View Source Repository

Agent Skills Search Helper

Install a tiny helper to your Agent, search and equip skill from 223,000+ vetted skills library on demand.