In Bayesian regression, we place a prior on the coefficients of the linear regression model that biases the model parameters to 0. The hyperparameters of the model control this prior and to what extent it biases the coefficients of the model to 0. In Bayesian ridge regression, the estimated parameters are the maximum a posteriori probability (MAP) estimates. While I haven't discussed it, there is, in fact, a prior distribution that we are imposing upon the coefficients of the model, and Bayesian ridge regression will estimate a posterior distribution. The coefficients that maximize this posterior distribution are referred to as the MAP estimates of the coefficients. We will choose a polunomial using the following steps: ...
Choosing a polynomial
Get Training Systems Using Python Statistical Modeling now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.