BoRP: Bootstrapped Regression Probing for Scalable and Human-Aligned LLM Evaluation
By: Peng Sun, Xiangyu Zhang, Duan Wu
Potential Business Impact:
Makes AI assistants understand if people are happy.
Accurate evaluation of user satisfaction is critical for iterative development of conversational AI. However, for open-ended assistants, traditional A/B testing lacks reliable metrics: explicit feedback is sparse, while implicit metrics are ambiguous. To bridge this gap, we introduce BoRP (Bootstrapped Regression Probing), a scalable framework for high-fidelity satisfaction evaluation. Unlike generative approaches, BoRP leverages the geometric properties of LLM latent space. It employs a polarization-index-based bootstrapping mechanism to automate rubric generation and utilizes Partial Least Squares (PLS) to map hidden states to continuous scores. Experiments on industrial datasets show that BoRP (Qwen3-8B/14B) significantly outperforms generative baselines (even Qwen3-Max) in alignment with human judgments. Furthermore, BoRP reduces inference costs by orders of magnitude, enabling full-scale monitoring and highly sensitive A/B testing via CUPED.
Similar Papers
Bootstrapping LLMs via Preference-Based Policy Optimization
Artificial Intelligence
Teaches AI to follow human wishes better.
IB-GRPO: Aligning LLM-based Learning Path Recommendation with Educational Objectives via Indicator-Based Group Relative Policy Optimization
Artificial Intelligence
Helps students learn better with smart lesson plans.
Prompted Policy Search: Reinforcement Learning through Linguistic and Numerical Reasoning in LLMs
Machine Learning (CS)
Teaches robots to learn faster with words.