Detecting High-Stakes Interactions with Activation Probes
By: Alex McKenzie , Urja Pawar , Phil Blandfort and more
Potential Business Impact:
Catches AI when it might cause harm.
Monitoring is an important aspect of safely deploying Large Language Models (LLMs). This paper examines activation probes for detecting "high-stakes" interactions -- where the text indicates that the interaction might lead to significant harm -- as a critical, yet underexplored, target for such monitoring. We evaluate several probe architectures trained on synthetic data, and find them to exhibit robust generalization to diverse, out-of-distribution, real-world data. Probes' performance is comparable to that of prompted or finetuned medium-sized LLM monitors, while offering computational savings of six orders-of-magnitude. Our experiments also highlight the potential of building resource-aware hierarchical monitoring systems, where probes serve as an efficient initial filter and flag cases for more expensive downstream analysis. We release our novel synthetic dataset and codebase to encourage further study.
Similar Papers
Investigating task-specific prompts and sparse autoencoders for activation monitoring
Machine Learning (CS)
Helps check if AI is saying safe things.
Detecting Strategic Deception Using Linear Probes
Machine Learning (CS)
Finds if AI is lying by checking its thoughts.
Building Production-Ready Probes For Gemini
Machine Learning (CS)
Keeps AI from being used for bad things.