Subword Tokenization Strategies for Kurdish Word Embeddings
By: Ali Salehi, Cassandra L. Jacobs
Potential Business Impact:
Helps computers understand Kurdish words better.
We investigate tokenization strategies for Kurdish word embeddings by comparing word-level, morpheme-based, and BPE approaches on morphological similarity preservation tasks. We develop a BiLSTM-CRF morphological segmenter using bootstrapped training from minimal manual annotation and evaluate Word2Vec embeddings across comprehensive metrics including similarity preservation, clustering quality, and semantic organization. Our analysis reveals critical evaluation biases in tokenization comparison. While BPE initially appears superior in morphological similarity, it evaluates only 28.6\% of test cases compared to 68.7\% for morpheme model, creating artificial performance inflation. When assessed comprehensively, morpheme-based tokenization demonstrates superior embedding space organization, better semantic neighborhood structure, and more balanced coverage across morphological complexity levels. These findings highlight the importance of coverage-aware evaluation in low-resource language processing and offers different tokenization methods for low-resourced language processing.
Similar Papers
Tokens with Meaning: A Hybrid Tokenization Approach for NLP
Computation and Language
Helps computers understand languages better.
Tokenization Strategies for Low-Resource Agglutinative Languages in Word2Vec: Case Study on Turkish and Finnish
Computation and Language
Makes computers understand languages better.
Evaluating Subword Tokenization Techniques for Bengali: A Benchmark Study with BengaliBPE
Computation and Language
Helps computers understand Bengali words better.