Score: 0

Interpreting Time Series Forecasts with LIME and SHAP: A Case Study on the Air Passengers Dataset

Published: August 17, 2025 | arXiv ID: 2508.12253v1

By: Manish Shukla

Potential Business Impact:

Shows why future predictions are right.

Plain English Summary

Imagine you need to predict how many customers will visit your store next week, or how much electricity your city will need. This new method helps make those predictions much more accurate and understandable. It's like getting a clear explanation of *why* the prediction is what it is, so you can trust it and make better decisions. This means businesses and cities can plan more effectively, saving money and resources.

Time-series forecasting underpins critical decisions across aviation, energy, retail and health. Classical autoregressive integrated moving average (ARIMA) models offer interpretability via coefficients but struggle with nonlinearities, whereas tree-based machine-learning models such as XGBoost deliver high accuracy but are often opaque. This paper presents a unified framework for interpreting time-series forecasts using local interpretable model-agnostic explanations (LIME) and SHapley additive exPlanations (SHAP). We convert a univariate series into a leakage-free supervised learning problem, train a gradient-boosted tree alongside an ARIMA baseline and apply post-hoc explainability. Using the Air Passengers dataset as a case study, we show that a small set of lagged features -- particularly the twelve-month lag -- and seasonal encodings explain most forecast variance. We contribute: (i) a methodology for applying LIME and SHAP to time series without violating chronology; (ii) theoretical exposition of the underlying algorithms; (iii) empirical evaluation with extensive analysis; and (iv) guidelines for practitioners.

Page Count
9 pages

Category
Computer Science:
Machine Learning (CS)