K folds cross-validation

K folds cross-validation is a much better estimator of our model's performance, even more so than our train-test split. Here's how it works:

We will take a finite number of equal slices of our data (usually 3, 5, or 10). Assume that this number is called k.

  1. For each "fold" of the cross-validation, we will treat k-1 of the sections as the training set, and the remaining section as our test set.
  2. For the remaining folds, a different arrangement of k-1 sections is considered for our training set and a different section is our training set.
  3. We compute a set metric for each fold of the cross-validation.
  4. We average our scores at the end.

Cross-validation is effectively using multiple train-test splits being done on the same dataset. ...

Get Principles of Data Science - Second Edition now with the O’Reilly learning platform.

O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.