High-Dimensional Learning in Finance
By: Hasan Fallahgoul
Potential Business Impact:
Makes stock market predictions more reliable.
Recent advances in machine learning have shown promising results for financial prediction using large, over-parameterized models. This paper provides theoretical foundations and empirical validation for understanding when and how these methods achieve predictive success. I examine two key aspects of high-dimensional learning in finance. First, I prove that within-sample standardization in Random Fourier Features implementations fundamentally alters the underlying Gaussian kernel approximation, replacing shift-invariant kernels with training-set dependent alternatives. Second, I establish information-theoretic lower bounds that identify when reliable learning is impossible no matter how sophisticated the estimator. A detailed quantitative calibration of the polynomial lower bound shows that with typical parameter choices, e.g., 12,000 features, 12 monthly observations, and R-square 2-3%, the required sample size to escape the bound exceeds 25-30 years of data--well beyond any rolling-window actually used. Thus, observed out-of-sample success must originate from lower-complexity artefacts rather than from the intended high-dimensional mechanism.
Similar Papers
Predicting Realized Variance Out of Sample: Can Anything Beat The Benchmark?
Statistical Finance
Predicts stock prices better for smarter investing.
Macroeconomic Forecasting and Machine Learning
Econometrics
Predicts future money problems more accurately.
Precise asymptotic analysis of Sobolev training for random feature models
Machine Learning (Stat)
Teaches computers better by using more data.