Text-Trained LLMs Can Zero-Shot Extrapolate PDE Dynamics
By: Jiajun Bao , Nicolas Boullé , Toni J. B. Liu and more
Potential Business Impact:
Computers predict future events from math patterns.
Large language models (LLMs) have demonstrated emergent in-context learning (ICL) capabilities across a range of tasks, including zero-shot time-series forecasting. We show that text-trained foundation models can accurately extrapolate spatiotemporal dynamics from discretized partial differential equation (PDE) solutions without fine-tuning or natural language prompting. Predictive accuracy improves with longer temporal contexts but degrades at finer spatial discretizations. In multi-step rollouts, where the model recursively predicts future spatial states over multiple time steps, errors grow algebraically with the time horizon, reminiscent of global error accumulation in classical finite-difference solvers. We interpret these trends as in-context neural scaling laws, where prediction quality varies predictably with both context length and output length. To better understand how LLMs are able to internally process PDE solutions so as to accurately roll them out, we analyze token-level output distributions and uncover a consistent ICL progression: beginning with syntactic pattern imitation, transitioning through an exploratory high-entropy phase, and culminating in confident, numerically grounded predictions.
Similar Papers
From Text to Trajectories: GPT-2 as an ODE Solver via In-Context
Artificial Intelligence
Teaches computers to solve math problems with examples.
From Equations to Insights: Unraveling Symbolic Structures in PDEs with LLMs
Machine Learning (CS)
AI finds hidden math rules in science problems.
Uncovering Emergent Physics Representations Learned In-Context by Large Language Models
Computation and Language
Computers learn physics concepts from examples.