In this post, you will learn about a feature selection technique called as Sequential Backward Selection using Python code example.
Feature selection is one of the key steps in training the most optimal model in order to achieve higher computational efficiency while training the model, and also reduce the the generalization error of the model by removing irrelevant features or noise. Some of the important feature selection techniques includes L-norm regularization and greedy search algorithms such as sequential forward or backward feature selection, especially for algorithms which don’t support regularization. It is of utmost importance for data scientists to learn these techniques in order to build optimal models.
Sequential backward selection algorithm aims to reduce the dimensionality of the initial feature subspace from N to K-features with a minimum reduction in the model performance to improve upon computational efficiency and reduce generalization error. The primary idea is to sequentially remove features from the given features list consisting of N features to reach to the list of K-features. At each stage of removal, the feature that causes the least performance loss gets removed. The search for features is based on combinatorial search algorithm where the subset of features get selected from combination and score is calculated for the subset and compared to other subsets. Note the usage of combinations class of itertools package in the code given below.
Here is the python code for sequential backward selection algorithm. The class takes the constructor as an instance of an estimator and subset of features to which the original feature space have to be reduced to. One can pass the training and test data set after feature scaling is done to determine the subset of features. The same is depicted in the python code given below.
The class provides a fit and transform method. Fit method determines the subset of features indices along with the accuracy scores. Pay attention to some of the following in the code:
- scores_ consists of best score with different number of dimensions
- subsets_ consists of feature indices representing features that led to best scores
from sklearn.metrics import accuracy_score from itertools import combinations from sklearn.base import clone from numpy as np class SequentialBackwardSearch(): ''' Instantiate with Estimator and given number of features ''' def __init__(self, estimator, k_features): self.estimator = clone(estimator) self.k_features = k_features ''' X_train - Training data Pandas dataframe X_test - Test data Pandas dataframe y_train - Training label Pandas dataframe y_test - Test data Pandas dataframe ''' def fit(self, X_train, X_test, y_train, y_test): dim = X_train.shape self.indices_ = tuple(range(dim)) self.subsets_ = [self.indices_] score = self._calc_score(X_train.values, X_test.values, y_train.values, y_test.values, self.indices_) self.scores_ = [score]''' Iterate through all the dimensions until k_features is reached At the end of loop, dimension count is reduced by 1 ''' while dim > k_features: scores =  subsets =  ''' Iterate through different combinations of features, train the model, record the score ''' for p in combinations(self.indices_, r=dim - 1): score = self._calc_score(X_train.values, X_test.values, y_train.values, y_test.values, p) scores.append(score) subsets.append(p) # # Get the index of best score # best_score_index = np.argmax(scores) # # Record the best score # self.scores_.append(scores[best_score_index]) # # Get the indices of features which gave best score # self.indices_ = subsets[best_score_index]# # Record the indices of features for best score # self.subsets_.append(self.indices_) dim -= 1 # Dimension is reduced by 1 ''' Transform training, test data set to the data set havng features which gave best score ''' def transform(self, X): return X.values[:, self.indices_] ''' Train models with specific set of features indices - indices of features ''' def _calc_score(self, X_train, X_test, y_train, y_test, indices): self.estimator.fit(X_train[:, indices], y_train.ravel()) y_pred = self.estimator.predict(X_test[:, indices]) score = accuracy_score(y_test, y_pred) return score
Here is the python code representing how the instance of LogisticRegression (default solver = lbfgs) is used as an estimator. This is just for illustration purpose as one could apply L1 norm regularization technique with solver as liblinear for training the model with most appropriate features.
# # Instantiate the estimator - LogisticRegression # lr = LogisticRegression(C=1.0, random_state=1) # # Number of features - 5 # k_features = 5 # # Instantiate SequentialBackwardSearch # sbs = SequentialBackwardSearch(lr, k_features) # # Fit the data to determine the k_features which give the # most optimal model performance # sbs.fit(X_train, X_test, y_train, y_test) # # Transform the training data set to dataset having k_features # giving most optimal model performance # X_train_kfeatures = sbs.transform(X_train) # # Transform the test data set to dataset having k_features # X_test_kfeatures = sbs.transform(X_test)
One could find the most optimal K-features using the following code:
X_train.columns[[sbs.indices_]] # sbs is an instance of SequentialBackwardSearch class
One could do the plot displaying most optimal features as each stage of reduction from N to K features vis-a-vis the model scores. Here is the python code:
k_features = [len(k) for k in sbs.subsets_] from matplotlib.pyplot as plt plt.plot(k_features, sbs.scores_, marker='o') plt.ylim([0.7, 1.02]) plt.ylabel('Accuracy') plt.xlabel('Number of features') plt.grid() plt.tight_layout() plt.show()
Here is how the plot would look like:
Some of the code snippets have been taken from Python Machine Learning book by Dr. Sebastian Raschka
- Fixed vs Random vs Mixed Effects Models – Examples - November 29, 2020
- Hierarchical Clustering Explained with Python Example - November 26, 2020
- Negative Binomial Distribution Python Examples - November 24, 2020