Continuum Dropout for Neural Differential Equations
By: Jonghun Lee , YongKyung Oh , Sungil Kim and more
Potential Business Impact:
Makes AI better at learning from messy, incomplete data.
Neural Differential Equations (NDEs) excel at modeling continuous-time dynamics, effectively handling challenges such as irregular observations, missing values, and noise. Despite their advantages, NDEs face a fundamental challenge in adopting dropout, a cornerstone of deep learning regularization, making them susceptible to overfitting. To address this research gap, we introduce Continuum Dropout, a universally applicable regularization technique for NDEs built upon the theory of alternating renewal processes. Continuum Dropout formulates the on-off mechanism of dropout as a stochastic process that alternates between active (evolution) and inactive (paused) states in continuous time. This provides a principled approach to prevent overfitting and enhance the generalization capabilities of NDEs. Moreover, Continuum Dropout offers a structured framework to quantify predictive uncertainty via Monte Carlo sampling at test time. Through extensive experiments, we demonstrate that Continuum Dropout outperforms existing regularization methods for NDEs, achieving superior performance on various time series and image classification tasks. It also yields better-calibrated and more trustworthy probability estimates, highlighting its effectiveness for uncertainty-aware modeling.
Similar Papers
Continuum Dropout for Neural Differential Equations
Machine Learning (Stat)
Makes AI better at learning from messy, incomplete data.
Convergence, design and training of continuous-time dropout as a random batch method
Machine Learning (CS)
Makes computer learning faster and more accurate.
Neural CDEs as Correctors for Learned Time Series Models
Machine Learning (CS)
Fixes wrong guesses in future predictions.