UncertaintyZoo: A Unified Toolkit for Quantifying Predictive Uncertainty in Deep Learning Systems
By: Xianzong Wu , Xiaohong Li , Lili Quan and more
Potential Business Impact:
Helps AI know when it's wrong.
Large language models(LLMs) are increasingly expanding their real-world applications across domains, e.g., question answering, autonomous driving, and automatic software development. Despite this achievement, LLMs, as data-driven systems, often make incorrect predictions, which can lead to potential losses in safety-critical scenarios. To address this issue and measure the confidence of model outputs, multiple uncertainty quantification(UQ) criteria have been proposed. However, even though important, there are limited tools to integrate these methods, hindering the practical usage of UQ methods and future research in this domain. To bridge this gap, in this paper, we introduce UncertaintyZoo, a unified toolkit that integrates 29 uncertainty quantification methods, covering five major categories under a standardized interface. Using UncertaintyZoo, we evaluate the usefulness of existing uncertainty quantification methods under the code vulnerability detection task on CodeBERT and ChatGLM3 models. The results demonstrate that UncertaintyZoo effectively reveals prediction uncertainty. The tool with a demonstration video is available on the project site https://github.com/Paddingbuta/UncertaintyZoo.
Similar Papers
From Calibration to Collaboration: LLM Uncertainty Quantification Should Be More Human-Centered
Computation and Language
Helps people know when to trust AI answers.
The Illusion of Certainty: Uncertainty quantification for LLMs fails under ambiguity
Machine Learning (CS)
Makes AI understand when it's unsure.
Torch-Uncertainty: A Deep Learning Framework for Uncertainty Quantification
Machine Learning (CS)
Makes AI know when it's not sure.