Fraudulent Credit Card Activity Detection Using Adaptive Boosting and Aggregate Voting
Abstract
With the rapid digitization of financial transactions, credit card fraud has emerged as a major concern, posing serious risks to both individual users and financial institutions. This research addresses the challenge of detecting fraudulent credit card activities in a timely and accurate manner through a hybrid ensemble learning approach. The proposed framework integrates Adaptive Boosting (AdaBoost) and Aggregate Majority Voting to form a highly robust fraud detection system. AdaBoost, known for its ability to enhance the predictive performance of weak classifiers, is utilized as the core component. Multiple shallow decision trees are iteratively trained, with each successive model focusing on instances previously misclassified, thereby refining the detection of complex fraud patterns. To further strengthen the model's stability and decision reliability, the outputs of these classifiers are combined via a majority voting mechanism, where the final decision is determined by consensus among the classifiers. Experimental evaluation, conducted on both benchmark and real-world credit card datasets, demonstrates that this hybrid system outperforms traditional individual classifiers, standalone AdaBoost, and other conventional detection methods in terms of precision, recall, F1-score, and Receiver Operating Characteristic (ROC) curve analysis. The hybrid system also shows strong resilience in noisy environments, making it a viable solution for practical fraud prevention applications
Downloads
References
R. Bolton and D. Hand, “Statistical fraud detection: A review,” Statistical Science, vol. 17, no. 3, pp. 235–255, 2002.
[2] A. Srivastava, A. Kundu, S. Sural, and A. Majumdar, “Credit card fraud detection using hidden Markov model,” IEEE Transactions on Dependable and Secure Computing, vol. 5, no. 1, pp. 37–48, Jan.–Mar. 2008.
[3] J. West and M. Bhattacharya, “Intelligent financial fraud detection: A comprehensive review,” Computers & Security, vol. 57, pp. 47–66, 2016.
[4] A. Dal Pozzolo, O. Caelen, R. A. Johnson, and G. ontempi, “Calibrating probability with undersampling for unbalanced classification,” in Proc. IEEE Symp. Series on Computational Intelligence, 2015, pp. 159–166.
[5] Bank Negara Malaysia, “Payment systems data,” 2016. [Online]. Available: https://www.bnm.gov.my
[6] K. Randhawa, C. K. Loo, M. Seera, C. P. Lim, and A. N. K. Nandi, “Credit card fraud detection using AdaBoost and majority voting,” IEEE Access, vol. 6, pp. 14277–14284, 2018.
[7] The Nilson Report, “Global credit card fraud losses,” Issue 1074, 2015.
[8] R. Jha, M. R. Islam, and A. Abdullah, “Cost of fraud to merchants: An analysis,” Journal of Financial Crime, vol. 24, no. 3, pp. 450–463, 2017.
[9] S. Bhattacharyya, S. Jha, K. Tharakunnel, and J. C. Westland, “Data mining for credit card fraud: A comparative study,” Decision Support Systems, vol. 50, no. 3, pp. 602–613, 2011.
[10] H. Abdallah, A. Maarof, and A. Zainal, “Fraud detection system: A survey,” Journal of Network and Computer Applications, vol. 68, pp. 90–113, 2016.
[11] X. Maes, V. Gestel, and D. Baesens, “Credit card fraud detection using Bayesian and neural networks,” in Proc. 1st Int. Conf. Computational Intelligence for Financial Engineering, 2009.
[12] M. Syeda, Y. Zhang, and Y. Pan, “Parallel granular neural networks for fast credit card fraud detection,” in Proc. IEEE Int. Conf. Fuzzy Systems, 2002, pp. 572–577.
[13] J. Han, M. Kamber, and J. Pei, Data Mining: Concepts and Techniques, 3rd ed. San Francisco, CA, USA: Morgan Kaufmann, 2011.
[14] I. Brown, “Statistical methods in credit card fraud detection,” Comput. Fraud Secur., vol. 2009, no. 9, pp. 13– 17, 2009.
[15] Y. Freund and R. Schapire, “A decision-theoretic eneralization of on-line learning and an application to boosting,” J. Comput. Syst. Sci., vol. 55, no. 1, pp. 119–139, 1997.
[16] T. G. Dietterich, “Ensemble methods in machine learning,” in Proc. Int. Workshop Multiple Classifier Systems, 2000, pp. 1–15.
[17] L. Breiman, “Bagging predictors,” Machine Learning, vol. 24, no. 2, pp. 123–140, 199