Sentence-Anchored Gist Compression for Long-Context LLMs
By: Dmitrii Tarasov, Elizaveta Goncharova, Kuznetsov Andrey
Potential Business Impact:
Makes computers understand longer stories with less effort.
This work investigates context compression for Large Language Models (LLMs) using learned compression tokens to reduce the memory and computational demands of processing long sequences. We demonstrate that pre-trained LLMs can be fine-tuned to compress their context by factors of 2x to 8x without significant performance degradation, as evaluated on both short-context and long-context benchmarks. Furthermore, in experiments on a 3-billion-parameter LLaMA model, our method achieves results on par with alternative compression techniques while attaining higher compression ratios.
Similar Papers
CompLLM: Compression for Long Context Q&A
Computation and Language
Makes AI understand long texts much faster.
AdmTree: Compressing Lengthy Context with Adaptive Semantic Trees
Computation and Language
Makes computers understand long stories better.
Learning to Compress: Unlocking the Potential of Large Language Models for Text Representation
Computation and Language
Makes computers understand writing better for searching.