From Static Structures to Ensembles: Studying and Harnessing Protein Structure Tokenization
By: Zijing Liu , Bin Feng , He Cao and more
Potential Business Impact:
Shows how proteins bend and move naturally.
Protein structure tokenization converts 3D structures into discrete or vectorized representations, enabling the integration of structural and sequence data. Despite many recent works on structure tokenization, the properties of the underlying discrete representations are not well understood. In this work, we first demonstrate that the successful utilization of structural tokens in a language model for structure prediction depends on using rich, pre-trained sequence embeddings to bridge the semantic gap between the sequence and structural "language". The analysis of the structural vocabulary itself then reveals significant semantic redundancy, where multiple distinct tokens correspond to nearly identical local geometries, acting as "structural synonyms". This redundancy, rather than being a flaw, can be exploited with a simple "synonym swap" strategy to generate diverse conformational ensembles by perturbing a predicted structure with its structural synonyms. This computationally lightweight method accurately recapitulates protein flexibility, performing competitively with state-of-the-art models. Our study provides fundamental insights into the nature of discrete protein structure representations and introduces a powerful, near-instantaneous method for modeling protein dynamics. Source code is available in https://github.com/IDEA-XL/TokenMD.
Similar Papers
Protein Structure Tokenization: Benchmarking and New Recipe
Quantitative Methods
Makes computers understand protein shapes better.
ProtTeX: Structure-In-Context Reasoning and Editing of Proteins with Large Language Models
Biomolecules
Helps computers understand and design proteins by reading their shapes.
Innovative tokenisation of structured data for LLM training
Machine Learning (CS)
Turns messy data into neat lists for smart computers.