Learning curve

From Machinelearning
Revision as of 01:39, 19 June 2014 by Vipul (talk | contribs) (Created page with "==Definition== The term '''learning curve''' is typically used in the context of a graph where the axes are as follows: * The horizontal axis represents a hyperparameter...")
(diff) ← Older revision | Latest revision (diff) | Newer revision → (diff)

Definition

The term learning curve is typically used in the context of a graph where the axes are as follows:

  • The horizontal axis represents a hyperparameter. This could be a model hyperparameter (a hyperparameter controlling the choice of model), regularization hyperparameter (a hyperparameter controlling how we regularize), or learning algorithm hyperparameter (a hyperparameter controlling how the learning algorithm proceeds).
  • The vertical axis represents the cost function value. There are several different cost function values that could be plotted:
    • The value of the regularized cost function on the training set (this is the function that we are ostensibly trying to optimize with the learning algorithm).
    • The value of the unregularized cost function on the training set.
    • The value of the unregularized cost function on the cross-validation set (or test set).

In many cases, we plot all the curves together in the same picture, so that we can compare the training and test errors.

When we plot the learning curve with respect to a particular hyperparameter, we are holding everything else fixed.

General concepts

We say that a cost function value is "high" if it is similar to or more than the cost function value one could get without any knowledge of the training data. For instance, in the logistic regression problem with a logarithmic cost function, always predicting a probability of 0.5 yields an error of on all data sets, so a cost function that is close to, or greater than, , is high.

We say that a cost function value is "low" if it is close to zero.

  • We say that a particular hyperparameter value exhibits high bias if the training error and cross-validation error are both high, we may also say that the model is underfitted.
  • We say that a particular hyperparameter value exhibits high variance if the training error is low but the cross-validation error is high. We also call this a situation of overfitting.

The interpretation of the high bias and high variance situations depends on what sort of hyperparameter our model is in terms of. Some cases are discussed below.