AnnoDPO: Protein Functional Annotation Learning with Direct Preference Optimization
By: Zixuan Jiang, Renjing Xu
Potential Business Impact:
Helps computers understand what proteins do better.
Deciphering protein function remains a fundamental challenge in protein representation learning. The task presents significant difficulties for protein language models (PLMs) due to the sheer volume of functional annotation categories and the highly imbalanced distribution of annotated instances across biological ontologies. Inspired by the remarkable success of reinforcement learning from human feedback (RLHF) in large language model (LLM) alignment, we propose AnnoDPO, a novel multi-modal framework for protein function prediction that leverages Direct Preference Optimization (DPO) to enhance annotation learning. Our methodology addresses the dual challenges of annotation scarcity and category imbalance through preference-aligned training objectives, establishing a new paradigm for biological knowledge integration in protein representation learning.
Similar Papers
Active Learning for Direct Preference Optimization
Machine Learning (CS)
Teaches AI to learn faster from human choices.
A Survey of Direct Preference Optimization
Machine Learning (CS)
Teaches computers to be helpful and safe.
Improving Protein Sequence Design through Designability Preference Optimization
Machine Learning (CS)
Designs proteins that reliably fold into shapes.