Topics
Vector norm: function that assigns non-negative length/size to a vector. For vector in , -norm written as , satisfying:
- Non-negativity: , only if
- Absolute homogeneity: for any scalar
- Triangle inequality:
- Definiteness:
L1 Norm (Manhattan Norm)
- Notation:
- Definition: Sum of absolute components
- Interpretation: “Taxicab”/“Manhattan” distance from origin
- Use in ML: As a regularizer, it encourages sparsity (L1 regularization)
L₂ Norm (Euclidean Norm)
- Notation: (often just )
- Definition: Square root of sum of squares
- Interpretation: Standard Euclidean distance from origin
- Use in ML: Common regularizer (L2 regularization), ridge regression, weight decay
Max (∞) Norm
- Notation:
- Definition: Maximum absolute component
- Interpretation: Greatest distance along any single coordinate axis
- Use in ML: Bounding weights (max-norm regularization in neural nets)
General p-Norms and Properties
- -norm for :
- As ,
- All satisfy triangle inequality and definiteness
from math import inf
from numpy import array
from numpy.linalg import norm
a = array([1, 2, 3])
l1 = norm(a, 1)
l2 = norm(a, 2)
l_inf = norm(a, inf)
Why Vector Norms Matter in Machine Learning
- Regularization: Adding or penalties controls model complexity
- Distance Measures: Nearest-neighbor classifiers, clustering rely on L1/L2 distances
- Optimization Geometry: Norms define contours, influence convergence of gradient-based methods