U&O Fitting
U&O Fitting
Goodness of Fit
The "Goodness of fit" term is taken from the statistics, and the goal of the
machine learning models to achieve the goodness of fit. In statistics modeling,
it defines how closely the result or predicted values match the true values of
the dataset.
The model with a good fit is between the underfitted and overfitted model, and
ideally, it makes predictions with 0 errors, but in practice, it is difficult to
achieve it.
As when we train our model for a time, the errors in the training data go down,
and the same happens with test data. But if we train the model for a long
duration, then the performance of the model may decrease due to the
overfitting, as the model also learn the noise present in the dataset. The errors
in the test dataset start increasing, so the point, just before the raising of
errors, is the good point, and we can stop here for achieving a good model.
Good Fit in a Statistical Model: Ideally, the case when the model makes the
predictions with 0 error, is said to have a good fit on the data. This
situation is achievable at a spot between overfitting and underfitting. In
order to understand it, we will have to look at the performance of our
model with the passage of time, while it is learning from the training
dataset.
With the passage of time, our model will keep on learning, and thus the
error for the model on the training and testing data will keep on
decreasing. If it will learn for too long, the model will become more prone
to overfitting due to the presence of noise and less useful details. Hence
the performance of our model will decrease. In order to get a good fit, we
will stop at a point just before where the error starts increasing. At this
point, the model is said to have good skills in training datasets as well as
our unseen testing dataset.