Applied ML is highly iterative - decide hyperparameters, code, evaluate, reiterate. So quick prototyping is key for ensuring success of the ML project
Bias is an indicator of training set performance. If it is high, then it means we are underfitting on our train set and there is scope to learn more from the train data
Variance is an indicator of test set performance. If it is high, it means we are overfitting on train set and the model cannot generalize well on the data it has not seen
Image 1: Bias and Variance Visualized for 2 dimensional data
Image 2: High Bias and High Variance visualized for 2 dimension data. The classifier is mostly linear with some highly complex non-linear elements
Visualizing bias and variance beyond 2 dimensions is difficult. In that case, we resort to Train Set Error and Dev Set Error
Image 3: Example of low/high bias and variance values based on Train and Dev set errors. Optimal Bayes Error is ~0%
Optimal or Bayes error: This gives an indication of the error value of an optimal classifier. Think of it like a benchmark value for the error. In the above example, we assume that the Optimal error is nearly 0% since humans can identify a cat image with almost no error. Optimal error could be 14% if the images were blurred to the point where machines or even humans cannot determine the right class of the object
Bias and Variance Trade-off? Not Really in DL
Image 4: Flow chart describing steps to tackle a ML problem. Try and reduce high bias first, then try reducing high variance