Efficient Reinforcement Learning from Human Feedback via Bayesian Preference Inference
By: Matteo Cercola, Valeria Capretti, Simone Formentin
Potential Business Impact:
Teaches computers faster by asking them what they like.
Learning from human preferences is a cornerstone of aligning machine learning models with subjective human judgments. Yet, collecting such preference data is often costly and time-consuming, motivating the need for more efficient learning paradigms. Two established approaches offer complementary advantages: RLHF scales effectively to high-dimensional tasks such as LLM fine-tuning, while PBO achieves greater sample efficiency through active querying. We propose a hybrid framework that unifies RLHF's scalability with PBO's query efficiency by integrating an acquisition-driven module into the RLHF pipeline, thereby enabling active and sample-efficient preference gathering. We validate the proposed approach on two representative domains: (i) high-dimensional preference optimization and (ii) LLM fine-tuning. Experimental results demonstrate consistent improvements in both sample efficiency and overall performance across these tasks.
Similar Papers
Policy-labeled Preference Learning: Is Preference Enough for RLHF?
Machine Learning (CS)
Teaches computers to learn better from people.
Maximizing the efficiency of human feedback in AI alignment: a comparative analysis
Human-Computer Interaction
Teaches AI to learn faster from people's choices.
Maximizing the efficiency of human feedback in AI alignment: a comparative analysis
Human-Computer Interaction
Teaches computers to learn what people like faster.