Exactly what it sounds like. Worth a note just because it keeps coming up.
Actually, you know what, the bias-variance decomposition is possibly worth discussing. Given a predictor and a target , the squared error can be decomposed as
where , and is the irreducible noise in the data (i.e., given , outcomes are drawn as where has variance .
As you increase model complexity, bias typically decreases and variance increases. This is known as the bias-variance tradeoff. So to minimize squared-error, you want to find a appropriate compromise between the two.