machine-learning-ops-ml-pipeline
Design and implement a complete ML pipeline for: $ARGUMENTS
Author
Category
AI Skill DevelopmentInstall
Download and extract to your skills directory
Copy command and send to OpenClaw for auto-install:
Machine Learning Pipeline - Multi-Agent MLOps Orchestration Skill
Skill Overview
This is an automated skill that, through collaborative multi-agent coordination, designs and implements production-grade machine learning pipelines. It covers the entire MLOps lifecycle end to end, from data engineering, model development, and deployment to ongoing monitoring.
Applicable Scenarios
1. Build an Enterprise ML Platform from Scratch
When you need to set up a complete production environment for an ML project, this skill coordinates multiple specialized agents—including data engineers, data scientists, ML engineers, MLOps engineers, and observability engineers—to systematically complete the full workflow: data pipeline design, feature engineering, model training, production deployment, and monitoring and alerting.
2. Modernize an Existing ML System
If your current ML system suffers from issues such as excessive manual operations, poor reproducibility, or missing monitoring, this skill can help you assess the current state and design improvement plans aligned with modern MLOps best practices. This includes introducing key capabilities such as experiment tracking, CI/CD automation, model drift detection, and more.
3. Deep Consulting for Specific ML Technical Solutions
When you need expert guidance on a particular technical area—such as choosing between Kubeflow and Airflow, designing a feature store architecture, or implementing a model drift detection approach—this skill can provide detailed analysis and implementation recommendations based on hands-on experience.
Core Features
1. Multi-Agent Collaborative Orchestration
The skill uses a staged coordination approach, where each stage is handled by a specialized domain agent: data engineers manage data ingestion and quality assurance; data scientists design features and experiment plans; ML engineers implement training pipelines; MLOps engineers handle production deployment; and observability engineers ensure the monitoring system is complete. Clear handoffs and quality gates between stages ensure that every part is handled by the right experts.
2. Integration of Modern MLOps Tooling
Supports selecting and integrating leading MLOps tools, including experiment tracking (MLflow, Weights & Biases, Neptune, ClearML), feature stores (Feast, Tecton, Databricks), model serving (KServe, Seldon, TorchServe, Triton), orchestration platforms (Kubeflow, Airflow, Prefect, Dagster), monitoring stacks (Prometheus, Datadog, NewRelic), etc. Customized recommendations can be provided based on your specific needs and environment.
3. Production-Ready Delivery Standards
More than just solution design, it focuses on real production requirements: 99.9% service availability, P99 inference latency under 200 ms, automated rollback within 5 minutes, a complete observability system, cost optimization strategies, disaster recovery procedures, and more. The final deliverables include an end-to-end automated pipeline, infrastructure as code, and complete documentation and operations manuals.
Common Questions
What are the main differences between an ML Pipeline and a traditional data pipeline?
Traditional data pipelines (such as ETL) primarily focus on moving and transforming data, usually in deterministic ways. ML Pipelines, in addition to data processing, also include ML-specific stages such as model training, evaluation, version management, A/B testing, and drift detection. ML processes are iterative and involve trial and error. ML Pipelines must manage additional assets such as model versions, experiment metadata, and feature definitions, and therefore have higher requirements for reproducibility and experiment tracking.
When is it necessary to introduce a multi-agent collaboration approach?
For simple, small ML projects, a single engineer or a small team can handle everything. However, when the project reaches a certain scale (e.g., cross-team collaboration, strict SLA requirements, complex compliance needs) or involves deep technical decisions across multiple professional domains, multi-agent collaboration helps ensure each stage has professional support, reduces cross-domain communication costs, and improves delivery quality. Especially for enterprise-level projects, this approach can prevent issues in the overall system caused by weakness in any single stage.
Which cloud platforms and deployment modes does this skill support?
The skill is designed for a cloud-native architecture and supports AWS, Azure, GCP, or multi-cloud hybrid deployment strategies. Deployment modes include real-time inference (REST/gRPC API), batch predictions (scheduled jobs), streaming processing (Kafka/Kinesis), or hybrid modes. Outputs include infrastructure-as-code artifacts such as Terraform modules, Kubernetes Helm charts, Docker build configurations, etc., which can be used directly in your chosen cloud environment.