Score: 0

A Generalized Bias-Variance Decomposition for Bregman Divergences

Published: November 11, 2025 | arXiv ID: 2511.08789v1

By: David Pfau

Potential Business Impact:

Makes computer learning more accurate for certain tasks.

Business Areas:
A/B Testing Data and Analytics

The bias-variance decomposition is a central result in statistics and machine learning, but is typically presented only for the squared error. We present a generalization of the bias-variance decomposition where the prediction error is a Bregman divergence, which is relevant to maximum likelihood estimation with exponential families. While the result is already known, there was not previously a clear, standalone derivation, so we provide one for pedagogical purposes. A version of this note previously appeared on the author's personal website without context. Here we provide additional discussion and references to the relevant prior literature.

Page Count
6 pages

Category
Computer Science:
Machine Learning (CS)