Atom-anchored LLMs speak Chemistry: A Retrosynthesis Demonstration
By: Alan Kai Hassen , Andrius Bernatavicius , Antonius P. A. Janssen and more
Potential Business Impact:
Teaches computers to invent new medicines.
Applications of machine learning in chemistry are often limited by the scarcity and expense of labeled data, restricting traditional supervised methods. In this work, we introduce a framework for molecular reasoning using general-purpose Large Language Models (LLMs) that operates without requiring labeled training data. Our method anchors chain-of-thought reasoning to the molecular structure by using unique atomic identifiers. First, the LLM performs a one-shot task to identify relevant fragments and their associated chemical labels or transformation classes. In an optional second step, this position-aware information is used in a few-shot task with provided class examples to predict the chemical transformation. We apply our framework to single-step retrosynthesis, a task where LLMs have previously underperformed. Across academic benchmarks and expert-validated drug discovery molecules, our work enables LLMs to achieve high success rates in identifying chemically plausible reaction sites ($\geq90\%$), named reaction classes ($\geq40\%$), and final reactants ($\geq74\%$). Beyond solving complex chemical tasks, our work also provides a method to generate theoretically grounded synthetic datasets by mapping chemical knowledge onto the molecular structure and thereby addressing data scarcity.
Similar Papers
Chemical reasoning in LLMs unlocks strategy-aware synthesis planning and reaction mechanism elucidation
Artificial Intelligence
Computers plan chemical reactions like expert scientists.
LLM-Augmented Chemical Synthesis and Design Decision Programs
Artificial Intelligence
Computers plan how to build new medicines faster.
Large Language Models Transform Organic Synthesis From Reaction Prediction to Automation
Artificial Intelligence
AI helps scientists invent new things faster.