from sklearn.preprocessing import PolynomialFeatures # Import PolynomialFeatures class
poly_reg = PolynomialFeatures(degree=2) # Create a PolynomialFeatures object with degree 2
X_poly = poly_reg.fit_transform(X) # Transform the original features to polynomial features
lin_reg_2 = LinearRegression() # Create a LinearRegression object
lin_reg_2.fit(X_poly, y) # Fit the linear regression model to the polynomial features
In machine learning, we often work with regression models to predict continuous outcomes. One common regression model is linear regression, which assumes a linear relationship between the input features (independent variables) and the output (dependent variable).
However, not all relationships in data are linear. To capture more complex patterns, we can use polynomial regression, which extends linear regression by adding polynomial terms to the model. This allows us to fit non-linear relationships.
Here’s a breakdown of the concepts and steps involved:
PolynomialFeatures
class from Scikit-Learn, we can transform our original feature matrix X into a new feature matrix X_poly that includes polynomial terms. The degree of the polynomial determines the highest power of the original features included.$$
y = b_0 + b_1x_1 + b_2x_2 + \ldots + b_nx_n
$$
$$
y = b_0 + b_1x + b_2x^2
$$
By transforming the features into polynomial terms, we can use the linear regression algorithm to fit a polynomial model to the data. This approach combines the simplicity of linear regression with the flexibility of polynomial regression to capture more complex patterns in the data.