Score: 0

Meta-Learning Neural Process for Implied Volatility Surfaces with SABR-induced Priors

Published: September 15, 2025 | arXiv ID: 2509.11928v2

By: Jirong Zhuang, Xuan Wu

Potential Business Impact:

Makes stock price predictions faster and more accurate.

Business Areas:
Semantic Search Internet Services

We treat implied volatility surface (IVS) reconstruction as a two-principle learning problem. First, we adopt a meta-learning view that trains across trading days to learn a procedure that maps sparse option quotes to a full IVS via conditional prediction, avoiding per-day calibration at test time. Second, we impose a structural prior via transfer learning: pre-train on SABR-generated dataset to encode geometric prior, then fine-tune on historical market dataset to align with empirical patterns. We implement both principles in a single attention-based Neural Process (Volatility Neural Process, VolNP) that produces a complete IVS from a sparse context set in one forward pass. On SPX options, the VolNP outperforms SABR, SSVI, and Gaussian Process. Relative to an ablation trained only on market data, the SABR-induced prior reduces RMSE by about 40% and suppresses large errors, with pronounced gains at long maturities where quotes are sparse. The resulting model is fast (single pass), stable (no daily parameter solving), and practical for deployment at scale.

Page Count
24 pages

Category
Quantitative Finance:
Computational Finance