It's known as Ridge Regression in sci-kit learn library. Similar to Linear Regression, except it takes an additional parameter \(alpha\)
\(alpha\) tries to pull coefficients to zero unless large coefficient value reduces training error by large value.
\(alpha\) of 0 becomes linear regression. BP use Ridge regression over linear regression with carefully tuned value of \(alpha\) (has no good default value)
Use RidgeCV (over GridSearchCV) (cross validation) to try different \(alpha\) values
scikit-learn by default uses regularization with LogisticRegression(C=1). C is inverse of \(alpha\) (high values of C leads to weaker regularization)
BP use regularization when n_samples << n_features