Functional uniqueness and stability of Gaussian priors in optimal L1 estimation
By: Leighton Barnes, Alex Dytso
Potential Business Impact:
Makes computers guess better with less data.
This paper studies the functional uniqueness and stability of Gaussian priors in optimal $L^1$ estimation. While it is well known that the Gaussian prior uniquely induces linear conditional means under Gaussian noise, the analogous question for the conditional median (i.e., the optimal estimator under absolute-error loss) has only recently been settled. Building on the prior work establishing this uniqueness, we develop a quantitative stability theory that characterizes how approximate linearity of the optimal estimator constrains the prior distribution. For $L^2$ loss, we derive explicit rates showing that near-linearity of the conditional mean implies proximity of the prior to the Gaussian in the Lévy metric. For $L^1$ loss, we introduce a Hermite expansion framework and analyze the adjoint of the linearity-defining operator to show that the Gaussian remains the unique stable solution. Together, these results provide a more complete functional-analytic understanding of linearity and stability in Bayesian estimation under Gaussian noise.
Similar Papers
Linearity-Inducing Priors for Poisson Parameter Estimation Under $L^{1}$ Loss
Statistics Theory
Lets computers guess numbers better from messy data.
CLT in high-dimensional Bayesian linear regression with low SNR
Statistics Theory
Helps understand data when signals are weak.
Weak convergence of Bayes estimators under general loss functions
Statistics Theory
Makes computer guesses about data more accurate.