A k-fold cross validation technique is used to minimize the overfitting or bias in the predictive modeling techniques used in this study. k-Fold cross validation is used with decision tree and neural network with MLP and RBF to generate more flexible models to reduce overfitting or underfitting.

The complete dataset R is randomly split into k-mutually exclusive subsets or folds of approximately equal size (R_1 , R_2 , . . . , R_k). In classification model each of the k-subsets are trained and tested k times.

One of the k-subsets is used as the test set and the other k-1 subsets are used as training dataset.

Then the overall accuracy is calculated as the average of the k individual accuracy measures. All the machine learning techniques in this study use 10- fold cross validation.

Cross-Validation

Cross-Validation