Skip to content

Assessment Checklist

Before you submit your M9 final assessment, check every box below to ensure you have covered the full ML workflow.

The MVP Submission

  • Have I formulated a clear ML business problem?
  • Did I acquire, clean, and engineer features appropriately?
  • Have I trained at least one baseline and one advanced model?
  • Is there rigorous validation (CV, training vs test analysis)?
  • Have I explicitly translated the model metric (Accuracy/RMSE) into a business metric (£ saved or time reduced)?
  • Did I communicate the results effectively with a visualisation?

Data Preparation

  • Missing values handled (dropped, imputed, or flagged)
  • Categorical features encoded (One-Hot, Ordinal, or Target Encoding)
  • Numerical features scaled where required (StandardScaler for distance-based models)
  • No data leakage — preprocessing fitted on training data only

Modelling

  • At least two algorithms compared (e.g., Logistic Regression vs Random Forest)
  • Hyperparameters tuned systematically (GridSearch, RandomSearch, or Optuna)
  • Cross-validation used — not a single train/test split
  • Overfitting checked (training vs test score gap)

Evaluation

  • Appropriate metric chosen (not just accuracy — consider F1, ROC AUC, RMSE)
  • Confusion matrix or classification report included (for classification)
  • Residual analysis included (for regression)
  • Confidence intervals or standard deviation reported

Communication

  • Feature importance visualised (SHAP, permutation, or tree-based)
  • Results presented in plain English for a non-technical audience
  • Business impact quantified (ROI, cost savings, efficiency gains)
  • Limitations and next steps acknowledged

KSB Mapping

KSB Description How This Addresses It
K1 Context of Data Science Understanding where ML sits within the broader discipline
S3 Programming languages and tools Setting up the development environment and dependencies
B6 Commitment to keeping up to date Engaging with current ML resources and research