Score: 0

Leveraging Large Language Models for enzymatic reaction prediction and characterization

Published: May 8, 2025 | arXiv ID: 2505.05616v1

By: Lorenzo Di Fruscia, Jana Marie Weber

Potential Business Impact:

Helps computers guess how tiny body machines work.

Business Areas:
Natural Language Processing Artificial Intelligence, Data and Analytics, Software

Predicting enzymatic reactions is crucial for applications in biocatalysis, metabolic engineering, and drug discovery, yet it remains a complex and resource-intensive task. Large Language Models (LLMs) have recently demonstrated remarkable success in various scientific domains, e.g., through their ability to generalize knowledge, reason over complex structures, and leverage in-context learning strategies. In this study, we systematically evaluate the capability of LLMs, particularly the Llama-3.1 family (8B and 70B), across three core biochemical tasks: Enzyme Commission number prediction, forward synthesis, and retrosynthesis. We compare single-task and multitask learning strategies, employing parameter-efficient fine-tuning via LoRA adapters. Additionally, we assess performance across different data regimes to explore their adaptability in low-data settings. Our results demonstrate that fine-tuned LLMs capture biochemical knowledge, with multitask learning enhancing forward- and retrosynthesis predictions by leveraging shared enzymatic information. We also identify key limitations, for example challenges in hierarchical EC classification schemes, highlighting areas for further improvement in LLM-driven biochemical modeling.

Country of Origin
🇳🇱 Netherlands

Page Count
21 pages

Category
Computer Science:
Artificial Intelligence