Skip to content

Scikit-Learn Regressors Reference

Quick lookup for commonly utilized continuous machine learning algorithms within sklearn.

Linear Models

LinearRegression

Use Case: Baseline continuous relationships. Key Parameters: None (Standard Ordinary Least Squares).

Ridge and Lasso

Use Case: Regularised linear regression to drastically prevent overfitting. Key Parameters: * alpha: Regularisation strength (Larger means stronger penalty).

Tree-Based Models

DecisionTreeRegressor

Use Case: Non-linear rule-based regression. Key Parameters: * max_depth: Limits geometric tree depth.

RandomForestRegressor

Use Case: Averaged, robust ensemble predictions. Key Parameters: * n_estimators: Count of trees constructed. * max_features: Maximum variables considered at each dynamic split.

GradientBoostingRegressor

Use Case: Corrective step-wise ensemble. Key Parameters: * learning_rate: Shrinkage parameter. * loss: Loss function to minimize (e.g., 'squared_error').

Advanced Models

SVR (Support Vector Regressor)

Use Case: Geometric tube boundary analysis. Key Parameters: * kernel: Space transformation ('linear', 'rbf'). * C: Margin strictness. * epsilon: The mathematical width of the no-penalty tube.

KSB Mapping

KSB Description How This Addresses It
K4.1 Statistical models and methods Understanding the statistical basis of regression and classification
K4.2 ML and AI techniques Implementing and comparing supervised learning algorithms
K4.4 Resource constraints and trade-offs Model complexity vs interpretability; computational cost
S1 Scientific methods and hypothesis testing Formulating hypotheses and testing with rigorous validation
S4 Building models and validating Cross-validation, train/test evaluation, performance metrics
B5 Impartial, hypothesis-driven approach Honest evaluation of model performance and limitations