Information based feature selection algorithm based on the mutual information criteria of
max-relevance and min-redundancy presented by Peng et al. in http://ieeexplore.ieee.org/document/1453511/.
Because the algorithm makes use of information metric functions, such as Shanon Entropy, it requires discretized data.
from mrmr import MRMR
selection = MRMR(n_features=5)
indices = selection.fit(X, y)
X_selection = X[:, indices]
In a dataset with correlated features, using the mutual information between single feature vectors and the target vector as criterion for feature selection may result in selecting a fature set with some redundancy.
Within the selected features set, features that depend on each other would not improve class-discriminative power and could therefore be removed.
The goal of this algorithm is to find a set of feature vectors
The computation is performed incrementally, starting from the feature that shows larger mutual information with the y, and then selecting the
that maximize:
where