Score: 3

Leveraging semantic similarity for experimentation with AI-generated treatments

Published: October 24, 2025 | arXiv ID: 2510.21119v1

By: Lei Shi , David Arbour , Raghavendra Addanki and more

BigTech Affiliations: University of California, Berkeley

Potential Business Impact:

Helps computers understand complex ideas better.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Large Language Models (LLMs) enable a new form of digital experimentation where treatments combine human and model-generated content in increasingly sophisticated ways. The main methodological challenge in this setting is representing these high-dimensional treatments without losing their semantic meaning or rendering analysis intractable. Here, we address this problem by focusing on learning low-dimensional representations that capture the underlying structure of such treatments. These representations enable downstream applications such as guiding generative models to produce meaningful treatment variants and facilitating adaptive assignment in online experiments. We propose double kernel representation learning, which models the causal effect through the inner product of kernel-based representations of treatments and user covariates. We develop an alternating-minimization algorithm that learns these representations efficiently from data and provides convergence guarantees under a low-rank factor model. As an application of this framework, we introduce an adaptive design strategy for online experimentation and demonstrate the method's effectiveness through numerical experiments.

Country of Origin
🇺🇸 United States


Page Count
31 pages

Category
Statistics:
Methodology