Financial Fraud Detection Using Explainable AI and Stacking Ensemble Methods
By: Fahad Almalki, Mehedi Masud
Potential Business Impact:
Makes fraud detection easier to understand.
Traditional machine learning models often prioritize predictive accuracy, often at the expense of model transparency and interpretability. The lack of transparency makes it difficult for organizations to comply with regulatory requirements and gain stakeholders trust. In this research, we propose a fraud detection framework that combines a stacking ensemble of well-known gradient boosting models: XGBoost, LightGBM, and CatBoost. In addition, explainable artificial intelligence (XAI) techniques are used to enhance the transparency and interpretability of the model's decisions. We used SHAP (SHapley Additive Explanations) for feature selection to identify the most important features. Further efforts were made to explain the model's predictions using Local Interpretable Model-Agnostic Explanation (LIME), Partial Dependence Plots (PDP), and Permutation Feature Importance (PFI). The IEEE-CIS Fraud Detection dataset, which includes more than 590,000 real transaction records, was used to evaluate the proposed model. The model achieved a high performance with an accuracy of 99% and an AUC-ROC score of 0.99, outperforming several recent related approaches. These results indicate that combining high prediction accuracy with transparent interpretability is possible and could lead to a more ethical and trustworthy solution in financial fraud detection.
Similar Papers
Explainable AI in Big Data Fraud Detection
Machine Learning (CS)
Shows how computers catch fraud without secrets.
Enhancing IoMT Security with Explainable Machine Learning: A Case Study on the CICIOMT2024 Dataset
Cryptography and Security
Shows why computers flag medical device attacks.
Enhancing ML Models Interpretability for Credit Scoring
Machine Learning (CS)
Helps banks predict loan risk with clear rules.