Sparse minimum Redundancy Maximum Relevance for feature selection
By: Peter Naylor , Benjamin Poignard , Héctor Climente-González and more
Potential Business Impact:
Finds important clues in messy data.
We propose a feature screening method that integrates both feature-feature and feature-target relationships. Inactive features are identified via a penalized minimum Redundancy Maximum Relevance (mRMR) procedure, which is the continuous version of the classic mRMR penalized by a non-convex regularizer, and where the parameters estimated as zero coefficients represent the set of inactive features. We establish the conditions under which zero coefficients are correctly identified to guarantee accurate recovery of inactive features. We introduce a multi-stage procedure based on the knockoff filter enabling the penalized mRMR to discard inactive features while controlling the false discovery rate (FDR). Our method performs comparably to HSIC-LASSO but is more conservative in the number of selected features. It only requires setting an FDR threshold, rather than specifying the number of features to retain. The effectiveness of the method is illustrated through simulations and real-world datasets. The code to reproduce this work is available on the following GitHub: https://github.com/PeterJackNaylor/SmRMR.
Similar Papers
Robust Gene Prioritization via Fast-mRMR Feature Selection in high-dimensional omics data
Machine Learning (CS)
Finds important genes for health research faster.
When Features Beat Noise: A Feature Selection Technique Through Noise-Based Hypothesis Testing
Machine Learning (Stat)
Finds the most important information in messy data.
When Features Beat Noise: A Feature Selection Technique Through Noise-Based Hypothesis Testing
Machine Learning (Stat)
Finds the most important information in data.