Chapter 13. Linear Regression
13.0 Introduction
Linear regression is one of the simplest supervised learning algorithms in our toolkit. If you have ever taken an introductory statistics course in college, likely the final topic you covered was linear regression. Linear regression and its extensions continue to be a common and useful method of making predictions when the target vector is a quantitative value (e.g., home price, age). In this chapter we will cover a variety of linear regression methods (and some extensions) for creating well-performing prediction models.
13.1 Fitting a Line
Problem
You want to train a model that represents a linear relationship between the feature and target vector.
Solution
Use a linear regression (in scikit-learn, LinearRegression
):
# Load libraries
from
sklearn.linear_model
import
LinearRegression
from
sklearn.datasets
import
make_regression
# Generate features matrix, target vector
features
,
target
=
make_regression
(
n_samples
=
100
,
n_features
=
3
,
n_informative
=
2
,
n_targets
=
1
,
noise
=
0.2
,
coef
=
False
,
random_state
=
1
)
# Create linear regression
regression
=
LinearRegression
()
# Fit the linear regression
model
=
regression
.
fit
(
features
,
target
)
Discussion
Linear regression assumes that the relationship between the features and the target vector is approximately linear. That is, the effect (also called coefficient, weight, or parameter) of the features on the target vector is constant. In our solution, for the sake of explanation, ...
Get Machine Learning with Python Cookbook, 2nd Edition now with the O’Reilly learning platform.
O’Reilly members experience books, live events, courses curated by job role, and more from O’Reilly and nearly 200 top publishers.