The Effects of Data Augmentation on Confidence Estimation for LLMs
By: Rui Wang , Renyu Zhu , Minmin Lin and more
Potential Business Impact:
Makes AI more honest about what it knows.
Confidence estimation is crucial for reflecting the reliability of large language models (LLMs), particularly in the widely used closed-source models. Utilizing data augmentation for confidence estimation is viable, but discussions focus on specific augmentation techniques, limiting its potential. We study the impact of different data augmentation methods on confidence estimation. Our findings indicate that data augmentation strategies can achieve better performance and mitigate the impact of overconfidence. We investigate the influential factors related to this and discover that, while preserving semantic information, greater data diversity enhances the effectiveness of augmentation. Furthermore, the impact of different augmentation strategies varies across different range of application. Considering parameter transferability and usability, the random combination of augmentations is a promising choice.
Similar Papers
Evaluating the Effectiveness and Scalability of LLM-Based Data Augmentation for Retrieval
Information Retrieval
Makes search engines smarter with less effort.
Text Data Augmentation for Large Language Models: A Comprehensive Survey of Methods, Challenges, and Opportunities
Computation and Language
Makes AI write better by teaching it with more examples.
Semantic-preserved Augmentation with Confidence-weighted Fine-tuning for Aspect Category Sentiment Analysis
Computation and Language
Teaches computers to understand opinions better.