k-nearest neighbors (KNN) is an example of a non-parametric model, meaning it has no parameters to be learned from data; however, it has one very important hyperparameter—the number of neighbors. Back in Chapter 4, Predicting Numerical Values with Machine Learning, we used a KNN model with 12 neighbors in our diamond prices dataset, and that model gave us the best result when compared with multiple linear regression and lasso. We used 12 because I thought it was a good value, however, there is no guarantee it is the best value; why not nine or 15 neighbors? Maybe 10 works best. That is what the activity of hyperparameter optimization is all about—finding the set of best hyperparameters, or, more often, a ...
Optimizing a single hyperparameter
Get Hands-On Predictive Analytics with Python now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.