Topic Modelling Black Box Optimization
By: Roman Akramov , Artem Khamatullin , Svetlana Glazyrina and more
Potential Business Impact:
Finds the best number of topics for text.
Choosing the number of topics $T$ in Latent Dirichlet Allocation (LDA) is a key design decision that strongly affects both the statistical fit and interpretability of topic models. In this work, we formulate the selection of $T$ as a discrete black-box optimization problem, where each function evaluation corresponds to training an LDA model and measuring its validation perplexity. Under a fixed evaluation budget, we compare four families of optimizers: two hand-designed evolutionary methods - Genetic Algorithm (GA) and Evolution Strategy (ES) - and two learned, amortized approaches, Preferential Amortized Black-Box Optimization (PABBO) and Sharpness-Aware Black-Box Optimization (SABBO). Our experiments show that, while GA, ES, PABBO, and SABBO eventually reach a similar band of final perplexity, the amortized optimizers are substantially more sample- and time-efficient. SABBO typically identifies a near-optimal topic number after essentially a single evaluation, and PABBO finds competitive configurations within a few evaluations, whereas GA and ES require almost the full budget to approach the same region.
Similar Papers
PABBO: Preferential Amortized Black-Box Optimization
Machine Learning (Stat)
Teaches computers to learn what people like faster.
MALBO: Optimizing LLM-Based Multi-Agent Teams via Multi-Objective Bayesian Optimization
Multiagent Systems
Finds best AI teams, saving money and time.
BAMBO: Construct Ability and Efficiency LLM Pareto Set via Bayesian Adaptive Multi-objective Block-wise Optimization
Machine Learning (CS)
Builds better AI by finding best speed-accuracy mixes.