XLM: A Python package for non-autoregressive language models
By: Dhruvesh Patel , Durga Prasad Maram , Sai Sreenivas Chintha and more
Potential Business Impact:
Makes computers write text faster and easier.
In recent years, there has been a resurgence of interest in non-autoregressive text generation in the context of general language modeling. Unlike the well-established autoregressive language modeling paradigm, which has a plethora of standard training and inference libraries, implementations of non-autoregressive language modeling have largely been bespoke making it difficult to perform systematic comparisons of different methods. Moreover, each non-autoregressive language model typically requires it own data collation, loss, and prediction logic, making it challenging to reuse common components. In this work, we present the XLM python package, which is designed to make implementing small non-autoregressive language models faster with a secondary goal of providing a suite of small pre-trained models (through a companion xlm-models package) that can be used by the research community. The code is available at https://github.com/dhruvdcoder/xlm-core.
Similar Papers
Diffusion Language Models are Super Data Learners
Machine Learning (CS)
Makes AI better at writing code with less data.
A Survey on Diffusion Language Models
Computation and Language
Makes computers write faster and understand better.
Beyond Next-Token Prediction: A Performance Characterization of Diffusion versus Autoregressive Language Models
Machine Learning (CS)
Makes computers write faster and understand longer stories.