Chapter 13. Linear Regression
13.0 Introduction
Linear regression is one of the simplest supervised learning algorithms in our toolkit. If you have ever taken an introductory statistics course in college, likely the final topic you covered was linear regression. Linear regression and its extensions continue to be a common and useful method of making predictions when the target vector is a quantitative value (e.g., home price, age). In this chapter we will cover a variety of linear regression methods (and some extensions) for creating well-performing prediction models.
13.1 Fitting a Line
Problem
You want to train a model that represents a linear relationship between the feature and target vector.
Solution
Use a linear regression (in scikit-learn, LinearRegression):
# Load librariesfromsklearn.linear_modelimportLinearRegressionfromsklearn.datasetsimportmake_regression# Generate features matrix, target vectorfeatures,target=make_regression(n_samples=100,n_features=3,n_informative=2,n_targets=1,noise=0.2,coef=False,random_state=1)# Create linear regressionregression=LinearRegression()# Fit the linear regressionmodel=regression.fit(features,target)
Discussion
Linear regression assumes that the relationship between the features and the target vector is approximately linear. That is, the effect (also called coefficient, weight, or parameter) of the features on the target vector is constant. In our solution, for the sake of explanation, ...
Become an O’Reilly member and get unlimited access to this title plus top books and audiobooks from O’Reilly and nearly 200 top publishers, thousands of courses curated by job role, 150+ live events each month,
and much more.
Read now
Unlock full access