Teaching Old Tokenizers New Words: Efficient Tokenizer Adaptation for Pre-trained Models
By: Taido Purason , Pavel Chizhov , Ivan P. Yamshchikov and more
Potential Business Impact:
Makes computer language models understand new words better.
Tokenizer adaptation plays an important role in transferring pre-trained language models to new domains or languages. In this work, we address two complementary aspects of this process: vocabulary extension and pruning. The common approach to extension trains a new tokenizer on domain-specific text and appends the tokens that do not overlap with the existing vocabulary, which often results in many tokens that are unreachable or never used. We propose continued BPE training, which adapts a pre-trained tokenizer by continuing the BPE merge learning process on new data. Experiments across multiple languages and model families show that this approach improves tokenization efficiency and leads to better utilization of added vocabulary. We also introduce leaf-based vocabulary pruning, which removes redundant tokens while preserving model quality. Together, these methods provide practical tools for controlled vocabulary modification, which we release as an open-source package.
Similar Papers
The Art of Breaking Words: Rethinking Multilingual Tokenizer Design
Computation and Language
Makes computers understand many languages faster.
One Tokenizer To Rule Them All: Emergent Language Plasticity via Multilingual Tokenizers
Computation and Language
Helps computers learn many new languages faster.
Parity-Aware Byte-Pair Encoding: Improving Cross-lingual Fairness in Tokenization
Computation and Language
Makes computer language tools fair for all languages.