Score: 0

A Multi-Component Reward Function with Policy Gradient for Automated Feature Selection with Dynamic Regularization and Bias Mitigation

Published: October 9, 2025 | arXiv ID: 2510.09705v1

By: Sudip Khadka, L. S. Paudel

Potential Business Impact:

Makes AI fair by choosing the right information.

Business Areas:
Machine Learning Artificial Intelligence, Data and Analytics, Software

Static feature exclusion strategies often fail to prevent bias when hidden dependencies influence the model predictions. To address this issue, we explore a reinforcement learning (RL) framework that integrates bias mitigation and automated feature selection within a single learning process. Unlike traditional heuristic-driven filter or wrapper approaches, our RL agent adaptively selects features using a reward signal that explicitly integrates predictive performance with fairness considerations. This dynamic formulation allows the model to balance generalization, accuracy, and equity throughout the training process, rather than rely exclusively on pre-processing adjustments or post hoc correction mechanisms. In this paper, we describe the construction of a multi-component reward function, the specification of the agents action space over feature subsets, and the integration of this system with ensemble learning. We aim to provide a flexible and generalizable way to select features in environments where predictors are correlated and biases can inadvertently re-emerge.

Country of Origin
πŸ‡ΊπŸ‡Έ United States

Page Count
14 pages

Category
Computer Science:
Machine Learning (CS)