KD-PINN: Knowledge-Distilled PINNs for ultra-low-latency real-time neural PDE solvers
By: Karim Bounja, Lahcen Laayouni, Abdeljalil Sakat
This work introduces Knowledge-Distilled Physics-Informed Neural Networks (KD-PINN), a framework that transfers the predictive accuracy of a high-capacity teacher model to a compact student through a continuous adaptation of the Kullback-Leibler divergence. To confirm its generality for various dynamics and dimensionalities, the framework is evaluated on a representative set of partial differential equations (PDEs). In all tested cases, the student model preserved the teacher's physical accuracy, with a mean RMSE increase below 0.64%, and achieved inference speedups ranging from 4.8x (Navier-Stokes) to 6.9x (Burgers). The distillation process also revealed a regularizing effect. With an average inference latency of 5.3 ms on CPU, the distilled models enter the ultra-low-latency real-time regime defined by sub-10 ms performance. Finally, this study examines how knowledge distillation reduces inference latency in PINNs to contribute to the development of accurate ultra-low-latency neural PDE solvers.
Similar Papers
KP-PINNs: Kernel Packet Accelerated Physics Informed Neural Networks
Computational Engineering, Finance, and Science
Solves hard math problems more reliably.
Lang-PINN: From Language to Physics-Informed Neural Networks via a Multi-Agent Framework
Artificial Intelligence
Builds smart computer programs from plain English.
Adaptive Physics-informed Neural Networks: A Survey
Machine Learning (CS)
Teaches computers to solve science problems faster.