PANER: A Paraphrase-Augmented Framework for Low-Resource Named Entity Recognition
By: Nanda Kumar Rengarajan, Jun Yan, Chun Wang
Potential Business Impact:
Helps computers find specific words in text.
Named Entity Recognition (NER) is a critical task that requires substantial annotated data, making it challenging in low-resource scenarios where label acquisition is expensive. While zero-shot and instruction-tuned approaches have made progress, they often fail to generalize to domain-specific entities and do not effectively utilize limited available data. We present a lightweight few-shot NER framework that addresses these challenges through two key innovations: (1) a new instruction tuning template with a simplified output format that combines principles from prior IT approaches to leverage the large context window of recent state-of-the-art LLMs; (2) introducing a strategic data augmentation technique that preserves entity information while paraphrasing the surrounding context, thereby expanding our training data without compromising semantic relationships. Experiments on benchmark datasets show that our method achieves performance comparable to state-of-the-art models on few-shot and zero-shot tasks, with our few-shot approach attaining an average F1 score of 80.1 on the CrossNER datasets. Models trained with our paraphrasing approach show consistent improvements in F1 scores of up to 17 points over baseline versions, offering a promising solution for groups with limited NER training data and compute power.
Similar Papers
Meta-Pretraining for Zero-Shot Cross-Lingual Named Entity Recognition in Low-Resource Philippine Languages
Computation and Language
Teaches computers to understand new languages faster.
Enhancing NER Performance in Low-Resource Pakistani Languages using Cross-Lingual Data Augmentation
Computation and Language
Helps computers understand rare languages better.
Positional Attention for Efficient BERT-Based Named Entity Recognition
Computation and Language
Makes computers find important words faster.