Progress Ratio Embeddings: An Impatience Signal for Robust Length Control in Neural Text Generation
By: Ivanhoé Botcazou , Tassadit Amghar , Sylvain Lamprier and more
Potential Business Impact:
Makes AI write stories exactly as long as you want.
Modern neural language models achieve high accuracy in text generation, yet precise control over generation length remains underdeveloped. In this paper, we first investigate a recent length control method based on Reverse Positional Embeddings (RPE) and show its limits when control is requested beyond the training distribution. In particular, using a discrete countdown signal tied to the absolute remaining token count leads to instability. To provide robust length control, we introduce Progress Ratio Embeddings (PRE), as continuous embeddings tied to a trigonometric impatience signal. PRE integrates seamlessly into standard Transformer architectures, providing stable length fidelity without degrading text accuracy under standard evaluation metrics. We further show that PRE generalizes well to unseen target lengths. Experiments on two widely used news-summarization benchmarks validate these findings.
Similar Papers
Length-Aware Rotary Position Embedding for Text-Speech Alignment
Audio and Speech Processing
Makes computer voices sound more natural.
Effective Length Extrapolation via Dimension-Wise Positional Embeddings Manipulation
Computation and Language
Lets computers remember much longer stories.
Selective Rotary Position Embedding
Computation and Language
Makes AI better at remembering and understanding long stories.