Topics

SL is about learning mapping from input X to output Y. It uses labeled training data (X, Y).

  • Goal: predict Y for new, unseen X accurately
  • Core tasks:: Classification (predict discrete category, e.g., spam/not spam), Regression (predict continuous value, e.g., price)

It contrasts unsupervised learning (no labels, finds patterns) & reinforcement learning (learns via rewards/penalties). The reliance on labeled data is SL’s defining feature. Data quality, labeling effort are critical prerequisites. Classification/regression distinction is fundamental: dictates algorithm choice and evaluation metrics.

Common algorithms: logistic regression, support vector machines (SVM), K-Nearest Neighbors (KNN algorithm), decision trees (DT), bagging (random forests - RF), boosting (adaboost - AB, gradient boosting - GB) etc.

High-level comparison of fundamental supervised algorithms:

AlgorithmKey IdeaCore Math Concept(s)
Logistic RegressionModel probability of binary outcome via sigmoid functionSigmoid, Log Loss, gradient descent
SVMMaximize margin between classes (Classification)Hyperplane, Margin, Hinge Loss, Kernels, Quadratic Programming
KNNClassify/predict based on K nearest neighborsdistance metrics (Euclidean, Manhattan), Majority Vote/Avg.
Decision TreeRecursive partitioning of data based on feature testsEntropy, Gini Impurity, Information Gain, Variance Reduction
Random ForestEnsemble of DTs via bagging + feature randomnessBagging, Decision Trees, Feature Importance, OOB Error
AdaBoostSequential ensemble, weights misclassified points higherBoosting, Weak Learners, Weighted Voting, Exponential Loss
Gradient BoostingSequential ensemble, fits new models to residual errors (gradients)Boosting, Weak Learners, Gradient Descent, Loss Functions