LAUD: Integrating Large Language Models with Active Learning for Unlabeled Data
By: Tzu-Hsuan Chou, Chun-Nan Chou
Potential Business Impact:
Teaches computers to learn from less data.
Large language models (LLMs) have shown a remarkable ability to generalize beyond their pre-training data, and fine-tuning LLMs can elevate performance to human-level and beyond. However, in real-world scenarios, lacking labeled data often prevents practitioners from obtaining well-performing models, thereby forcing practitioners to highly rely on prompt-based approaches that are often tedious, inefficient, and driven by trial and error. To alleviate this issue of lacking labeled data, we present a learning framework integrating LLMs with active learning for unlabeled dataset (LAUD). LAUD mitigates the cold-start problem by constructing an initial label set with zero-shot learning. Experimental results show that LLMs derived from LAUD outperform LLMs with zero-shot or few-shot learning on commodity name classification tasks, demonstrating the effectiveness of LAUD.
Similar Papers
Complementary Learning Approach for Text Classification using Large Language Models
Computation and Language
Helps people and computers work together better.
Exploring Fine-Tuning of Large Audio Language Models for Spoken Language Understanding under Limited Speech data
Sound
Teaches computers to understand speech better with less data.
Just Because You Can, Doesn't Mean You Should: LLMs for Data Fitting
Machine Learning (CS)
Computers change answers if you rename data.