InstructPLM-mu: 1-Hour Fine-Tuning of ESM2 Beats ESM3 in Protein Mutation Predictions
By: Junde Xu , Yapin Shi , Lijun Lang and more
Potential Business Impact:
Teaches computers to predict protein changes faster.
Multimodal protein language models deliver strong performance on mutation-effect prediction, but training such models from scratch demands substantial computational resources. In this paper, we propose a fine-tuning framework called InstructPLM-mu and try to answer a question: \textit{Can multimodal fine-tuning of a pretrained, sequence-only protein language model match the performance of models trained end-to-end? } Surprisingly, our experiments show that fine-tuning ESM2 with structural inputs can reach performance comparable to ESM3. To understand how this is achieved, we systematically compare three different feature-fusion designs and fine-tuning recipes. Our results reveal that both the fusion method and the tuning strategy strongly affect final accuracy, indicating that the fine-tuning process is not trivial. We hope this work offers practical guidance for injecting structure into pretrained protein language models and motivates further research on better fusion mechanisms and fine-tuning protocols.
Similar Papers
InstructPLM-mu: 1-Hour Fine-Tuning of ESM2 Beats ESM3 in Protein Mutation Predictions
Quantitative Methods
Makes protein predictions faster and cheaper.
Understanding protein function with a multimodal retrieval-augmented foundation model
Quantitative Methods
Helps predict how tiny body parts will change.
Elucidating the Design Space of Multimodal Protein Language Models
Machine Learning (CS)
Helps computers understand protein shapes better.