Prompt-Based One-Shot Exact Length-Controlled Generation with LLMs
By: Juncheng Xie, Hung-yi Lee
Potential Business Impact:
Makes AI write exactly the right number of words.
Controlling the length of text produced by large language models (LLMs) remains challenging: models frequently overshoot or undershoot explicit length instructions because they cannot reliably keep an internal token count. We present a prompt-based, one-shot strategy that compels an off-the-shelf LLM to generate exactly a desired number of tokens - words (English) or characters (Chinese) - without any fine-tuning or iterative sampling. The prompt appends countdown markers and explicit counting rules so that the model "writes while counting." We evaluate on four settings: open-ended generation (1-1000 tokens), XSUM summarization, MT-Bench-LI instruction following, and the LIFEBENCH equal-length track. On MT-Bench-LI, strict length compliance with GPT-4.1 leaps from below 30% under naive prompts to above 95% with our countdown prompt, surpassing the popular draft-then-revise baseline, while judged answer quality is preserved. These results show that precise length control can be achieved through prompt engineering alone, offering a lightweight alternative to training- or decoding-based methods.
Similar Papers
Plan-and-Write: Structure-Guided Length Control for LLMs without Model Retraining
Computation and Language
Makes AI write exactly the right amount of words.
Prompting LLMs: Length Control for Isometric Machine Translation
Computation and Language
Makes computer translations shorter and better.
Controllable Abstraction in Summary Generation for Large Language Models via Prompt Engineering
Computation and Language
Makes AI write better, shorter summaries.