Feature Importance Guided Random Forest Learning with Simulated Annealing Based Hyperparameter Tuning
By: Kowshik Balasubramanian, Andre Williams, Ismail Butun
Potential Business Impact:
Makes computer predictions more accurate and reliable.
This paper introduces a novel framework for enhancing Random Forest classifiers by integrating probabilistic feature sampling and hyperparameter tuning via Simulated Annealing. The proposed framework exhibits substantial advancements in predictive accuracy and generalization, adeptly tackling the multifaceted challenges of robust classification across diverse domains, including credit risk evaluation, anomaly detection in IoT ecosystems, early-stage medical diagnostics, and high-dimensional biological data analysis. To overcome the limitations of conventional Random Forests, we present an approach that places stronger emphasis on capturing the most relevant signals from data while enabling adaptive hyperparameter configuration. The model is guided towards features that contribute more meaningfully to classification and optimizing this with dynamic parameter tuning. The results demonstrate consistent accuracy improvements and meaningful insights into feature relevance, showcasing the efficacy of combining importance aware sampling and metaheuristic optimization.
Similar Papers
An explainable Recursive Feature Elimination to detect Advanced Persistent Threats using Random Forest classifier
Cryptography and Security
Finds hidden computer attacks with clear reasons.
Enhancing Password Security Through a High-Accuracy Scoring Framework Using Random Forests
Cryptography and Security
Makes passwords much harder for hackers to guess.
When Features Beat Noise: A Feature Selection Technique Through Noise-Based Hypothesis Testing
Machine Learning (Stat)
Finds the most important information in messy data.