In machine learning, early stopping is a form of regularization used when a machine learning model (such as a neural network) is trained by on-line gradient descent. In early stopping, the training set is split into a new training set and a validation set. Gradient descent is applied to the new training set. After each sweep through the new training set, the network is evaluated on the validation set. When the performance with the validation test stops improving, the algorithm halts. The network with the best performance on the validation set is then used for actual testing, with a separate set of data (the validation set is used in learning to decide when to stop).
This technique is a simple but efficient hack to deal with the problem of overfitting. Overfitting is a phenomenon in which a learning system, such as a neural network gets very good at dealing with one data set at the expense of becoming very bad at dealing with other data sets. Early stopping is effectively limiting the used weights in the network and thus imposes a regularization, effectively lowering the VC dimension.
Early stopping is a very common practice in neural network training and often produces networks that generalize well. However, while often improving the generalization it does not do so in a mathematically well-defined way.
Read more about Early Stopping: Method, Advantages, Issues
Famous quotes containing the words early and/or stopping:
“I realized how for all of us who came of age in the late sixties and early seventies the war was a defining experience. You went or you didnt, but the fact of it and the decisions it forced us to make marked us for the rest of our lives, just as the depression and World War II had marked my parents.”
—Linda Grant (b. 1949)
“Hamlet. To what base uses we may return, Horatio! Why may
not imagination trace the noble dust of Alexander, till a
find it stopping a bung-hole?
Horatio. Twere to consider too curiously to consider so.”
—William Shakespeare (15641616)