Few-shot Protein Fitness Prediction via In-context Learning and Test-time Training
By: Felix Teufel , Aaron W. Kollasch , Yining Huang and more
Potential Business Impact:
Helps scientists design better proteins faster.
Accurately predicting protein fitness with minimal experimental data is a persistent challenge in protein engineering. We introduce PRIMO (PRotein In-context Mutation Oracle), a transformer-based framework that leverages in-context learning and test-time training to adapt rapidly to new proteins and assays without large task-specific datasets. By encoding sequence information, auxiliary zero-shot predictions, and sparse experimental labels from many assays as a unified token set in a pre-training masked-language modeling paradigm, PRIMO learns to prioritize promising variants through a preference-based loss function. Across diverse protein families and properties-including both substitution and indel mutations-PRIMO outperforms zero-shot and fully supervised baselines. This work underscores the power of combining large-scale pre-training with efficient test-time adaptation to tackle challenging protein design tasks where data collection is expensive and label availability is limited.
Similar Papers
Exploring zero-shot structure-based protein fitness prediction
Quantitative Methods
Predicts how protein changes affect health.
ProtTeX: Structure-In-Context Reasoning and Editing of Proteins with Large Language Models
Biomolecules
Helps computers understand and design proteins by reading their shapes.
ProteinZero: Self-Improving Protein Generation via Online Reinforcement Learning
Machine Learning (CS)
Designs better proteins, failing less often.