
The most fundamental challenge in ML: building a model that generalizes to new data rather than memorizing training data.
Model memorizes training data — performs great on training set, poorly on test set. Too complex for the data.
Model too simple to capture patterns — performs poorly on both training and test sets.
High bias = underfitting. High variance = overfitting. The goal is the sweet spot of low bias and low variance.
Reference:
TaskLoco™ — The Sticky Note GOAT