Leveraging semantic similarity for experimentation with AI-generated treatments
By: Lei Shi , David Arbour , Raghavendra Addanki and more
Potential Business Impact:
Helps computers understand complex ideas better.
Large Language Models (LLMs) enable a new form of digital experimentation where treatments combine human and model-generated content in increasingly sophisticated ways. The main methodological challenge in this setting is representing these high-dimensional treatments without losing their semantic meaning or rendering analysis intractable. Here, we address this problem by focusing on learning low-dimensional representations that capture the underlying structure of such treatments. These representations enable downstream applications such as guiding generative models to produce meaningful treatment variants and facilitating adaptive assignment in online experiments. We propose double kernel representation learning, which models the causal effect through the inner product of kernel-based representations of treatments and user covariates. We develop an alternating-minimization algorithm that learns these representations efficiently from data and provides convergence guarantees under a low-rank factor model. As an application of this framework, we introduce an adaptive design strategy for online experimentation and demonstrate the method's effectiveness through numerical experiments.
Similar Papers
Behavior and Representation in Large Language Models for Combinatorial Optimization: From Feature Extraction to Algorithm Selection
Artificial Intelligence
Helps computers pick the best way to solve problems.
From Knowledge to Treatment: Large Language Model Assisted Biomedical Concept Representation for Drug Repurposing
Computation and Language
Finds new uses for old medicines faster.
Large Language Models as Model Organisms for Human Associative Learning
Machine Learning (CS)
Helps computers learn like brains, remembering new things.