Future Is Unevenly Distributed: Forecasting Ability of LLMs Depends on What We're Asking
By: Chinmay Karkar, Paras Chopra
Potential Business Impact:
Models guess future events better with more facts.
Large Language Models (LLMs) demonstrate partial forecasting competence across social, political, and economic events. Yet, their predictive ability varies sharply with domain structure and prompt framing. We investigate how forecasting performance varies with different model families on real-world questions about events that happened beyond the model cutoff date. We analyze how context, question type, and external knowledge affect accuracy and calibration, and how adding factual news context modifies belief formation and failure modes. Our results show that forecasting ability is highly variable as it depends on what, and how, we ask.
Similar Papers
Pitfalls in Evaluating Language Model Forecasters
Machine Learning (CS)
Makes AI predictions more trustworthy and accurate.
What Does ChatGPT Make of Historical Stock Returns? Extrapolation and Miscalibration in LLM Stock Return Forecasts
General Finance
AI predicts stock prices, but is too optimistic.
How AI Forecasts AI Jobs: Benchmarking LLM Predictions of Labor Market Changes
Computation and Language
Helps predict which jobs AI will change.