ml-pipeline-workflow
构建从数据准备到模型训练、验证及生产部署的端到端MLOps流水线。适用于创建机器学习管道、实施MLOps实践或自动化模型训练与部署工作流的场景。
name:ml-pipeline-workflowdescription:Build end-to-end MLOps pipelines from data preparation through model training, validation, and production deployment. Use when creating ML pipelines, implementing MLOps practices, or automating model training and deployment workflows.
ML Pipeline Workflow
Complete end-to-end MLOps pipeline orchestration from data preparation through model deployment.
Do not use this skill when
Instructions
resources/implementation-playbook.md.Overview
This skill provides comprehensive guidance for building production ML pipelines that handle the full lifecycle: data ingestion → preparation → training → validation → deployment → monitoring.
Use this skill when
What This Skill Provides
Core Capabilities
- End-to-end workflow design
- DAG orchestration patterns (Airflow, Dagster, Kubeflow)
- Component dependencies and data flow
- Error handling and retry strategies
- Data validation and quality checks
- Feature engineering pipelines
- Data versioning and lineage
- Train/validation/test splitting strategies
- Training job orchestration
- Hyperparameter management
- Experiment tracking integration
- Distributed training patterns
- Validation frameworks and metrics
- A/B testing infrastructure
- Performance regression detection
- Model comparison workflows
- Model serving patterns
- Canary deployments
- Blue-green deployment strategies
- Rollback mechanisms
Reference Documentation
See the references/ directory for detailed guides:
Assets and Templates
The assets/ directory contains:
Usage Patterns
Basic Pipeline Setup
# 1. Define pipeline stages
stages = [
"data_ingestion",
"data_validation",
"feature_engineering",
"model_training",
"model_validation",
"model_deployment"
]2. Configure dependencies
See assets/pipeline-dag.yaml.template for full example
Production Workflow
- Ingest raw data from sources
- Run data quality checks
- Apply feature transformations
- Version processed datasets
- Load versioned training data
- Execute training jobs
- Track experiments and metrics
- Save trained models
- Run validation test suite
- Compare against baseline
- Generate performance reports
- Approve for deployment
- Package model artifacts
- Deploy to serving infrastructure
- Configure monitoring
- Validate production traffic
Best Practices
Pipeline Design
Data Management
Model Operations
Deployment Strategies
Integration Points
Orchestration Tools
Experiment Tracking
Deployment Platforms
Progressive Disclosure
Start with the basics and gradually add complexity:
Common Patterns
Batch Training Pipeline
# See assets/pipeline-dag.yaml.template
stages:
- name: data_preparation
dependencies: []
- name: model_training
dependencies: [data_preparation]
- name: model_evaluation
dependencies: [model_training]
- name: model_deployment
dependencies: [model_evaluation]Real-time Feature Pipeline
# Stream processing for real-time features
Combined with batch training
See references/data-preparation.md
Continuous Training
# Automated retraining on schedule
Triggered by data drift detection
See references/model-training.md
Troubleshooting
Common Issues
Debugging Steps
Next Steps
After setting up your pipeline: