KuBERT: Central Kurdish BERT Model and Its Application for Sentiment Analysis
By: Kozhin muhealddin Awlla, Hadi Veisi, Abdulhady Abas Abdullah
Potential Business Impact:
Helps computers understand feelings in Kurdish.
This paper enhances the study of sentiment analysis for the Central Kurdish language by integrating the Bidirectional Encoder Representations from Transformers (BERT) into Natural Language Processing techniques. Kurdish is a low-resourced language, having a high level of linguistic diversity with minimal computational resources, making sentiment analysis somewhat challenging. Earlier, this was done using a traditional word embedding model, such as Word2Vec, but with the emergence of new language models, specifically BERT, there is hope for improvements. The better word embedding capabilities of BERT lend to this study, aiding in the capturing of the nuanced semantic pool and the contextual intricacies of the language under study, the Kurdish language, thus setting a new benchmark for sentiment analysis in low-resource languages.
Similar Papers
Emotion Recognition for Low-Resource Turkish: Fine-Tuning BERTurk on TREMO and Testing on Xenophobic Political Discourse
Computation and Language
Finds angry feelings in Turkish social media.
TWSSenti: A Novel Hybrid Framework for Topic-Wise Sentiment Analysis on Social Media Using Transformer Models
Computation and Language
Reads feelings from online words better.
KyrgyzBERT: A Compact, Efficient Language Model for Kyrgyz NLP
Computation and Language
Helps computers understand the Kyrgyz language better.