Topics

Vector norm: function that assigns non-negative length/size to a vector. For vector in , -norm written as , satisfying:

  1. Non-negativity: , only if
  2. Absolute homogeneity: for any scalar
  3. Triangle inequality:
  4. Definiteness:

L1 Norm (Manhattan Norm)

  • Notation:
  • Definition: Sum of absolute components
  • Interpretation: “Taxicab”/“Manhattan” distance from origin
  • Use in ML: As a regularizer, it encourages sparsity (L1 regularization)

L₂ Norm (Euclidean Norm)

  • Notation: (often just )
  • Definition: Square root of sum of squares
  • Interpretation: Standard Euclidean distance from origin
  • Use in ML: Common regularizer (L2 regularization), ridge regression, weight decay

Max (∞) Norm

  • Notation:
  • Definition: Maximum absolute component
  • Interpretation: Greatest distance along any single coordinate axis
  • Use in ML: Bounding weights (max-norm regularization in neural nets)

General p-Norms and Properties

  • -norm for :
  • As ,
  • All satisfy triangle inequality and definiteness
from math import inf
from numpy import array
from numpy.linalg import norm
 
 
a = array([1, 2, 3])
 
l1 = norm(a, 1)
l2 = norm(a, 2)
l_inf = norm(a, inf)

Why Vector Norms Matter in Machine Learning

  • Regularization: Adding or penalties controls model complexity
  • Distance Measures: Nearest-neighbor classifiers, clustering rely on L1/L2 distances
  • Optimization Geometry: Norms define contours, influence convergence of gradient-based methods