Large Language Models for Controllable Multi-property Multi-objective Molecule Optimization
By: Vishal Dey, Xiao Hu, Xia Ning
Potential Business Impact:
Improves medicine design by changing molecules better.
In real-world drug design, molecule optimization requires selectively improving multiple molecular properties up to pharmaceutically relevant levels, while maintaining others that already meet such criteria. However, existing computational approaches and instruction-tuned LLMs fail to capture such nuanced property-specific objectives, limiting their practical applicability. To address this, we introduce C-MuMOInstruct, the first instruction-tuning dataset focused on multi-property optimization with explicit, property-specific objectives. Leveraging C-MuMOInstruct, we develop GeLLMO-Cs, a series of instruction-tuned LLMs that can perform targeted property-specific optimization. Our experiments across 5 in-distribution and 5 out-of-distribution tasks show that GeLLMO-Cs consistently outperform strong baselines, achieving up to 126% higher success rate. Notably, GeLLMO-Cs exhibit impressive 0-shot generalization to novel optimization tasks and unseen instructions. This offers a step toward a foundational LLM to support realistic, diverse optimizations with property-specific objectives. C-MuMOInstruct and code are accessible through https://github.com/ninglab/GeLLMO-C.
Similar Papers
GeLLMO: Generalizing Large Language Models for Multi-property Molecule Optimization
Machine Learning (CS)
Helps design new medicines faster.
Collaborative Expert LLMs Guided Multi-Objective Molecular Optimization
Biomolecules
Finds better medicines faster.
ChatMol: A Versatile Molecule Designer Based on the Numerically Enhanced Large Language Model
Computational Engineering, Finance, and Science
Creates new medicines faster and better.