Understanding Customer Churn in Retail Banking through Explainable Predictive Analytics: Evidence of a Product Paradox
DOI:
https://doi.org/10.62411/jcta.15870Keywords:
Customer Churn, Explainable AI (XAI), Financial Analytics, Machine Learning, Predictive Analytics, Retail Banking, SHAP, Stacking EnsembleAbstract
The retention of customers in the retail banking sector is a critical economic imperative; however, predictive modeling is frequently hindered by severe class imbalance and the “Black Box” nature of complex algorithms. This study proposes a Heterogeneous Stacking Ensemble framework integrating XGBoost, CatBoost, and Random Forest base learners with a Logistic Regression meta-learner to forecast customer attrition. To overcome the pervasive “Majority Class Bias,” we introduce a “Dual-Imbalance Defense” that synergizes the Synthetic Minority Over-sampling Technique (SMOTE) with algorithmic cost-sensitive penalization. Furthermore, moving beyond standard accuracy metrics, the framework mathematically derives a dynamic classification threshold to guarantee a strict 0.90 recall rate, actively optimizing the capture of at-risk capital. Model opacity is addressed through the integration of a SHapley Additive exPlanations (SHAP) TreeExplainer. This cooperative game theory approach provides localized, patient-level “Reason Codes” for regulatory compliance and reveals global systemic vulnerabilities, including non-linear drivers such as the “Product Paradox.” Achieving a 0.90 recall rate and an AUC of 0.8654, this framework provides a statistically robust and operationally transparent tool for targeted customer retention.References
S. M. Keaveney, “Customer Switching Behavior in Service Industries: An Exploratory Study,” J. Mark., vol. 59, no. 2, pp. 71–82, Apr. 1995, doi: 10.1177/002224299505900206.
A. S. Adepeju, C. A. Edeze, S. Ojuade, M. T. Adenibuyan, F. I. Eneh, and A. S. Adepoju, “Predictive Analytics in Retail Banking Marketing,” Int. J. Manag. Organ. Res., vol. 2, no. 6, pp. 223–229, 2023, doi: 10.54660/IJMOR.2025.4.5.17-23.
A. Prashanthan, “An Integrated Framework for Optimizing Customer Retention Budget using Clustering, Classification, and Mathematical Optimization,” J. Comput. Theor. Appl., vol. 3, no. 1, pp. 45–63, Jul. 2025, doi: 10.62411/jcta.13194.
N. V. Chawla, K. W. Bowyer, L. O. Hall, and W. P. Kegelmeyer, “SMOTE: Synthetic Minority Over-sampling Technique,” J. Artif. Intell. Res., vol. 16, no. February 2017, pp. 321–357, Jun. 2002, doi: 10.1613/jair.953.
S. C. K. Tékouabou, Ștefan C. Gherghina, H. Toulni, P. N. Mata, and J. M. Martins, “Towards Explainable Machine Learning for Bank Churn Prediction Using Data Balancing and Ensemble-Based Methods,” Mathematics, vol. 10, no. 14, p. 2379, Jul. 2022, doi: 10.3390/math10142379.
C. Elkan, “The Foundations of Cost-Sensitive Learning,” in Proceedings of the 17th International Joint Conference on Artificial Intelligence (IJCAI), 2001, pp. 973–978. [Online]. Available: https://dl.acm.org/doi/10.5555/1642194.1642224
J. M. Johnson and T. M. Khoshgoftaar, “Survey on deep learning with class imbalance,” J. Big Data, vol. 6, no. 1, p. 27, Dec. 2019, doi: 10.1186/s40537-019-0192-5.
C. Rudin, “Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead,” Nat. Mach. Intell., vol. 1, no. 5, pp. 206–215, May 2019, doi: 10.1038/s42256-019-0048-x.
S. M. Lundberg et al., “From local explanations to global understanding with explainable AI for trees,” Nat. Mach. Intell., vol. 2, no. 1, pp. 56–67, Jan. 2020, doi: 10.1038/s42256-019-0138-9.
V.-H. Vu, “An Efficient Customer Churn Prediction Technique Using Combined Machine Learning in Commercial Banks,” Oper. Res. Forum, vol. 5, no. 3, p. 66, Jul. 2024, doi: 10.1007/s43069-024-00345-5.
R. Li, “Bank Customer Churn Prediction Based on Stacking Model,” Adv. Econ. Manag. Polit. Sci., vol. 185, no. 1, pp. 42–51, Jun. 2025, doi: 10.54254/2754-1169/2025.LH23930.
R. Ashraf, “Bank Customer Churn Prediction Using Machine Learning Framework,” J. Appl. Financ. Bank., vol. 14, no. 4, pp. 65–109, Jun. 2024, doi: 10.47260/jafb/1445.
S. Dutta, P. Bose, S. K. Bandyopadhyay, and M. Janarthanan, “A Hybrid Machine Learning Model for Bank Customer Churn Prediction,” Int. J. Eng. Trends Technol., vol. 70, no. 6, pp. 13–23, Jun. 2022, doi: 10.14445/22315381/IJETT-V70I6P202.
S. Kumar and C. D., “A Survey on Customer Churn Prediction using Machine Learning Techniques,” Int. J. Comput. Appl., vol. 154, no. 10, pp. 13–16, Nov. 2016, doi: 10.5120/ijca2016912237.
Miriyala Lavanya, “Customer Churn Prediction in Banking Sector Using Machine Learning,” J. Inf. Syst. Eng. Manag., vol. 10, no. 57s, pp. 224–238, Jul. 2025, doi: 10.52783/jisem.v10i57s.12181.
S. L. Kumar, “Bank Customer Churn Prediction Using Machine Learning,” Int. J. Res. Appl. Sci. Eng. Technol., vol. 9, no. VIII, pp. 727–732, Aug. 2021, doi: 10.22214/ijraset.2021.37467.
R. E. Ako et al., “Effects of Data Resampling on Predicting Customer Churn via a Comparative Tree-based Random Forest and XGBoost,” J. Comput. Theor. Appl., vol. 2, no. 1, pp. 86–101, Jun. 2024, doi: 10.62411/jcta.10562.
M. A. Hambali and I. Andrew, “Bank Customer Churn Prediction Using SMOTE: A Comparative Analysis,” Qeios, Mar. 2024, doi: 10.32388/H82XTW.
M. J. Nur, D. R. I. Moses Setiadi, A. A. Ojugo, and M. T. Nguyen, “Improving Customer Churn Prediction Using Domain-Driven Feature Engineering, Resampling, and CatBoost with Explainability Extensions,” in 2025 International Seminar on Application for Technology of Information and Communication (iSemantic), Sep. 2025, pp. 493–499. doi: 10.1109/ISemantic67418.2025.11291801.
L. Breiman, “Random Forests,” Mach. Learn., vol. 45, no. 1, pp. 5–32, Oct. 2001, doi: 10.1023/A:1010933404324.
T. Chen and C. Guestrin, “XGBoost: A Scalable Tree Boosting System,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Aug. 2016, pp. 785–794. doi: 10.1145/2939672.2939785.
L. Prokhorenkova, G. Gusev, A. Vorobev, A. V. Dorogush, and A. Gulin, “CatBoost: unbiased boosting with categorical features,” in Advances in Neural Information Processing Systems 31 (NeurIPS 2018), Jan. 2018. [Online]. Available: http://arxiv.org/abs/1706.09516
D. H. Wolpert, “Stacked generalization,” Neural Networks, vol. 5, no. 2, pp. 241–259, Jan. 1992, doi: 10.1016/S0893-6080(05)80023-1.
T. Fawcett, “An introduction to ROC analysis,” Pattern Recognit. Lett., vol. 27, no. 8, pp. 861–874, Jun. 2006, doi: 10.1016/j.patrec.2005.10.010.
S. M. Lundberg and S.-I. Lee, “A Unified Approach to Interpreting Model Predictions,” in Advances in Neural Information Processing Systems (NeurIPS), Nov. 2017, pp. 4766–4777. [Online]. Available: http://arxiv.org/abs/1705.07874
A. Prashanthan, R. Roshan, and M. Maduranga, “RetenNet: A Deployable Machine Learning Pipeline with Explainable AI and Prescriptive Optimization for Customer Churn Management,” J. Futur. Artif. Intell. Technol., vol. 2, no. 2, pp. 182–201, Jun. 2025, doi: 10.62411/faith.3048-3719-110.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Patrick Ndabarishye, Ajay Kumar Singh

This work is licensed under a Creative Commons Attribution 4.0 International License.














