Controlling Summarization Length Through EOS Token Weighting
By: Zeno Belligoli , Emmanouil Stergiadis , Eran Fainman and more
Potential Business Impact:
Makes AI summaries shorter or longer easily.
Controlling the length of generated text can be crucial in various text-generation tasks, including summarization. Existing methods often require complex model alterations, limiting compatibility with pre-trained models. We address these limitations by developing a simple approach for controlling the length of automatic text summaries by increasing the importance of correctly predicting the EOS token in the cross-entropy loss computation. The proposed methodology is agnostic to architecture and decoding algorithms and orthogonal to other inference-time techniques to control the generation length. We tested it with encoder-decoder and modern GPT-style LLMs, and show that this method can control generation length, often without affecting the quality of the summary.
Similar Papers
A Decoding Algorithm for Length-Control Summarization Based on Directed Acyclic Transformers
Computation and Language
Makes summaries exactly the right length.
Consistency Evaluation of News Article Summaries Generated by Large (and Small) Language Models
Computation and Language
Makes computer summaries more truthful to the original text.
An Empirical Comparison of Text Summarization: A Multi-Dimensional Evaluation of Large Language Models
Computation and Language
Finds best AI for summarizing text.