Score: 1

AlignSAE: Concept-Aligned Sparse Autoencoders

Published: December 1, 2025 | arXiv ID: 2512.02004v1

By: Minglai Yang , Xinyu Guo , Mihai Surdeanu and more

Potential Business Impact:

Lets AI understand and change specific ideas easily.

Business Areas:
Semantic Search Internet Services

Large Language Models (LLMs) encode factual knowledge within hidden parametric spaces that are difficult to inspect or control. While Sparse Autoencoders (SAEs) can decompose hidden activations into more fine-grained, interpretable features, they often struggle to reliably align these features with human-defined concepts, resulting in entangled and distributed feature representations. To address this, we introduce AlignSAE, a method that aligns SAE features with a defined ontology through a "pre-train, then post-train" curriculum. After an initial unsupervised training phase, we apply supervised post-training to bind specific concepts to dedicated latent slots while preserving the remaining capacity for general reconstruction. This separation creates an interpretable interface where specific relations can be inspected and controlled without interference from unrelated features. Empirical results demonstrate that AlignSAE enables precise causal interventions, such as reliable "concept swaps", by targeting single, semantically aligned slots.

Country of Origin
πŸ‡ΊπŸ‡Έ πŸ‡¨πŸ‡³ China, United States

Page Count
20 pages

Category
Computer Science:
Machine Learning (CS)