Uncertainty-driven Embedding Convolution
By: Sungjun Lim , Kangjun Noh , Youngjun Choi and more
Potential Business Impact:
Makes computer language understanding more reliable.
Text embeddings are essential components in modern NLP pipelines. While numerous embedding models have been proposed, their performance varies across domains, and no single model consistently excels across all tasks. This variability motivates the use of ensemble techniques to combine complementary strengths. However, most existing ensemble methods operate on deterministic embeddings and fail to account for model-specific uncertainty, limiting their robustness and reliability in downstream applications. To address these limitations, we propose Uncertainty-driven Embedding Convolution (UEC). UEC first transforms deterministic embeddings into probabilistic ones in a post-hoc manner. It then computes adaptive ensemble weights based on embedding uncertainty, grounded in a Bayes-optimal solution under a surrogate loss. Additionally, UEC introduces an uncertainty-aware similarity function that directly incorporates uncertainty into similarity scoring. Extensive experiments on retrieval, classification, and semantic similarity benchmarks demonstrate that UEC consistently improves both performance and robustness by leveraging principled uncertainty modeling.
Similar Papers
Calibrated and uncertain? Evaluating uncertainty estimates in binary classification models
Machine Learning (CS)
Helps computers know when they are unsure.
Epistemic Uncertainty-aware Recommendation Systems via Bayesian Deep Ensemble Learning
Information Retrieval
Suggests better movies and songs you'll like.
On Uncertainty Calibration for Equivariant Functions
Machine Learning (CS)
Makes AI smarter and more careful with new data.