LLM-MemCluster: Empowering Large Language Models with Dynamic Memory for Text Clustering
By: Yuanjie Zhu , Liangwei Yang , Ke Xu and more
Potential Business Impact:
Lets computers group words by meaning better.
Large Language Models (LLMs) are reshaping unsupervised learning by offering an unprecedented ability to perform text clustering based on their deep semantic understanding. However, their direct application is fundamentally limited by a lack of stateful memory for iterative refinement and the difficulty of managing cluster granularity. As a result, existing methods often rely on complex pipelines with external modules, sacrificing a truly end-to-end approach. We introduce LLM-MemCluster, a novel framework that reconceptualizes clustering as a fully LLM-native task. It leverages a Dynamic Memory to instill state awareness and a Dual-Prompt Strategy to enable the model to reason about and determine the number of clusters. Evaluated on several benchmark datasets, our tuning-free framework significantly and consistently outperforms strong baselines. LLM-MemCluster presents an effective, interpretable, and truly end-to-end paradigm for LLM-based text clustering.
Similar Papers
ClusterFusion: Hybrid Clustering with Embedding Guidance and LLM Adaptation
Computation and Language
Helps computers group words by meaning better.
LLM-as-classifier: Semi-Supervised, Iterative Framework for Hierarchical Text Classification using Large Language Models
Computation and Language
Makes smart computer programs sort text better.
An Explorative Study on Distributed Computing Techniques in Training and Inference of Large Language Models
Distributed, Parallel, and Cluster Computing
Lets big AI run on normal computers.