Score: 0

LAUD: Integrating Large Language Models with Active Learning for Unlabeled Data

Published: November 18, 2025 | arXiv ID: 2511.14738v1

By: Tzu-Hsuan Chou, Chun-Nan Chou

Potential Business Impact:

Teaches computers to learn from less data.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Large language models (LLMs) have shown a remarkable ability to generalize beyond their pre-training data, and fine-tuning LLMs can elevate performance to human-level and beyond. However, in real-world scenarios, lacking labeled data often prevents practitioners from obtaining well-performing models, thereby forcing practitioners to highly rely on prompt-based approaches that are often tedious, inefficient, and driven by trial and error. To alleviate this issue of lacking labeled data, we present a learning framework integrating LLMs with active learning for unlabeled dataset (LAUD). LAUD mitigates the cold-start problem by constructing an initial label set with zero-shot learning. Experimental results show that LLMs derived from LAUD outperform LLMs with zero-shot or few-shot learning on commodity name classification tasks, demonstrating the effectiveness of LAUD.

Page Count
7 pages

Category
Computer Science:
Machine Learning (CS)