Feasibility-Guided Fair Adaptive Offline Reinforcement Learning for Medicaid Care Management
By: Sanjay Basu , Sadiq Y. Patel , Parth Sheth and more
Potential Business Impact:
Makes AI fairer and safer for everyone.
We introduce Feasibility-Guided Fair Adaptive Reinforcement Learning (FG-FARL), an offline RL procedure that calibrates per-group safety thresholds to reduce harm while equalizing a chosen fairness target (coverage or harm) across protected subgroups. Using de-identified longitudinal trajectories from a Medicaid population health management program, we evaluate FG-FARL against behavior cloning (BC) and HACO (Hybrid Adaptive Conformal Offline RL; a global conformal safety baseline). We report off-policy value estimates with bootstrap 95% confidence intervals and subgroup disparity analyses with p-values. FG-FARL achieves comparable value to baselines while improving fairness metrics, demonstrating a practical path to safer and more equitable decision support.
Similar Papers
Hybrid Adaptive Conformal Offline Reinforcement Learning for Fair Population Health Management
Machine Learning (CS)
Helps doctors help sick people safely and fairly.
Failure-Aware RL: Reliable Offline-to-Online Reinforcement Learning with Self-Recovery for Real-World Manipulation
Robotics
Robots learn to avoid mistakes and work better.
FairGRPO: Fair Reinforcement Learning for Equitable Clinical Reasoning
Machine Learning (CS)
Makes AI doctors treat everyone fairly.