February 2018
Intermediate to advanced
378 pages
10h 14m
English
To fix the problem with irrelevant features, one can replace the L2 norm with the L1 norm in a penalty term, and instead of penalizing squares of regression coefficients, penalize their absolute values:
![]()
where L1 norm
. This is so-called Least Absolute Shrinkage and Selection Operator (LASSO) regression. Some of the weight coefficients under such a penalty can become exactly zero, which you can think of as a feature selection. If there are several highly correlated features in your dataset, LASSO picks one of them and sets all ...
Read now
Unlock full access