Quantifying Semantic Shift in Financial NLP: Robust Metrics for Market Prediction Stability
By: Zhongtian Sun , Chenghao Xiao , Anoushka Harit and more
Potential Business Impact:
Makes money predictions more reliable during crises.
Financial news is essential for accurate market prediction, but evolving narratives across macroeconomic regimes introduce semantic and causal drift that weaken model reliability. We present an evaluation framework to quantify robustness in financial NLP under regime shifts. The framework defines four metrics: (1) Financial Causal Attribution Score (FCAS) for alignment with causal cues, (2) Patent Cliff Sensitivity (PCS) for sensitivity to semantic perturbations, (3) Temporal Semantic Volatility (TSV) for drift in latent text representations, and (4) NLI-based Logical Consistency Score (NLICS) for entailment coherence. Applied to LSTM and Transformer models across four economic periods (pre-COVID, COVID, post-COVID, and rate hike), the metrics reveal performance degradation during crises. Semantic volatility and Jensen-Shannon divergence correlate with prediction error. Transformers are more affected by drift, while feature-enhanced variants improve generalisation. A GPT-4 case study confirms that alignment-aware models better preserve causal and logical consistency. The framework supports auditability, stress testing, and adaptive retraining in financial AI systems.
Similar Papers
Financial Sentiment Analysis on News and Reports Using Large Language Models and FinBERT
Information Retrieval
Helps computers understand money news better.
FinCARE: Financial Causal Analysis with Reasoning and Evidence
Computational Finance
Finds hidden money secrets for better investing.
The New Quant: A Survey of Large Language Models in Financial Prediction and Trading
Portfolio Management
Helps computers make smart money choices from news.