Operationalize the machine-learning lifecycle so models are reproducible, releasable, observable, and governable. Use when work involves training pipelines, artifact and feature versioning, model registries, CI/CD/CT, environment promotion, serving release controls, drift and performance monitoring, retraining policy, lineage, rollback, or MLOps maturity planning. Do not use for standalone model design when lifecycle automation and operational controls are not the core challenge.
Make ML dependable after the notebook ends.
Use this skill to:
Do not use this skill for:
ml-engineerCollect:
Return one or more of:
Use tables for artifact flow, ownership, release gates, and alerts.
A practical maturity lens:
Do not prescribe a high-ceremony platform a small team cannot run.
Track and recover at minimum:
If production artifacts cannot be recreated, the system is not under control.
Define gates for:
Every gate needs an owner, threshold, and action on failure.
Specify:
Model release policy should match business risk, not team optimism.
Track at least:
A healthy endpoint can still be a failing model.
Define:
Automatic retraining without release controls is just automated risk.
A strong result should:
prompt.md — response posture and structureexamples/README.md — representative request and output patternsguides/qa-checklist.md — final quality gatemeta/skill.json — aliases, boundaries, and metadata