InstructPLM-mu: 1-Hour Fine-Tuning of ESM2 Beats ESM3 in Protein Mutation Predictions
By: Junde Xu , Yapin Shi , Lijun Lang and more
Potential Business Impact:
Makes protein predictions faster and cheaper.
Multimodal protein language models deliver strong performance on mutation-effect prediction, but training such models from scratch demands substantial computational resources. In this paper, we propose a fine-tuning framework called InstructPLM-mu and try to answer a question: \textit{Can multimodal fine-tuning of a pretrained, sequence-only protein language model match the performance of models trained end-to-end? } Surprisingly, our experiments show that fine-tuning ESM2 with structural inputs can reach performance comparable to ESM3. To understand how this is achieved, we systematically compare three different feature-fusion designs and fine-tuning recipes. Our results reveal that both the fusion method and the tuning strategy strongly affect final accuracy, indicating that the fine-tuning process is not trivial. We hope this work offers practical guidance for injecting structure into pretrained protein language models and motivates further research on better fusion mechanisms and fine-tuning protocols.
Similar Papers
InstructPLM-mu: 1-Hour Fine-Tuning of ESM2 Beats ESM3 in Protein Mutation Predictions
Quantitative Methods
Teaches computers to predict protein changes faster.
Understanding protein function with a multimodal retrieval-augmented foundation model
Quantitative Methods
Helps predict how tiny body parts will change.
Protein Language Model Zero-Shot Fitness Predictions are Improved by Inference-only Dropout
Machine Learning (CS)
Improves computer guesses about protein health.