Evaluating Explainable Artificial Intelligence Methods for Interpretable Machine Learning Models in Large Scale Enterprise Data Analytics Systems
Keywords:
Explainable AI, Machine learning, LIME method, SHAP values, Enterprise systemsAbstract
Explainable Artificial Intelligence (XAI) has become a critical area of research within artificial intelligence, focusing on improving the transparency and interpretability of machine learning (ML) models, often referred to as "black-box" models. The need for XAI techniques arises from the inherent complexity of ML models, which can make their decision-making processes difficult for users to understand. This study investigates various XAI techniques, including LIME (Local Interpretable Model-agnostic Explanations) and SHAP (SHapley Additive exPlanations), to assess their impact on model interpretability without significantly compromising predictive performance. A comparative experimental design was used, applying these XAI methods to different ML models, including deep neural networks and ensemble methods, within large-scale enterprise data analytics systems. The results indicate that XAI methods significantly enhance model transparency and decision traceability, allowing users to understand the influence of individual features on predictions. While a slight reduction in predictive accuracy was observed, especially with simpler models, the trade-off between interpretability and performance was deemed acceptable, particularly in fields requiring transparency, such as healthcare, finance, and autonomous systems. The use of XAI in enterprise data systems has practical implications for fostering trust and enabling informed decision-making among stakeholders. Furthermore, the study discusses the challenges and limitations of applying XAI techniques, such as complexity, scalability, and model-specific limitations. Future research is suggested to focus on developing more scalable and efficient XAI methods, enhancing their applicability across various model types, and addressing the challenges of real-time applications. This will be crucial in ensuring the widespread adoption of XAI in critical domains, promoting the ethical use of AI while maintaining predictive accuracy.
References
[1] J. Henriques, T. Rocha, P. de Carvalho, C. Silva, and S. Paredes, “Interpretability and Explainability of Machine Learning Models: Achievements and Challenges,” in IFMBE Proceedings, 2024, pp. 81 – 94. doi: 10.1007/978-3-031-59216-4_9.
[2] A. Orlov, M. Ozhgibesov, and A. Agapov, “Enterprise machine learning and data platform,” in Saint Petersburg 2020 - Geosciences: Converting Knowledge into Resources, 2020. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85099534612&partnerID=40&md5=82882a49a08361aa105108f7c008560c
[3] M. A. Wajeed, K. M. Quadry, M. A. Mallik, and K. R. Khanna, “EXPLAINABLE MACHINE LEARNING FOR TEXT CLASSIFICATION: A NOVEL APPROACH TO TRANSPARENCY AND INTERPRETABILITY,” J. Theor. Appl. Inf. Technol., vol. 103, no. 3, pp. 907 – 915, 2025, [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-86000130472&partnerID=40&md5=41626c819996be79427f950e73334fe6
[4] M. Bakator et al., “Transforming Industry 5.0 Through Advanced Analytics and Machine Learning,” J. Soft Comput. Decis. Anal., vol. 3, no. 1, pp. 122 – 128, 2025, doi: 10.31181/jscda31202564.
[5] K. Hemachandran, S. Khanra, R. V Rodriguez, and J. R. Jaramillo, Machine Learning for Business Analytics: Real-Time Data Analysis for Decision-Making. 2022. doi: 10.4324/9781003206316.
[6] N. Kanchepu, Unveiling the Black Box: The Crucial Role of Interpretability in Machine Learning Models. 2025. doi: 10.2174/9789815305548125010004.
[7] R. Elshawi, Y. Sherif, M. Al-Mallah, and S. Sakr, “Interpretability in healthcare a comparative study of local machine learning interpretability techniques,” in Proceedings - IEEE Symposium on Computer-Based Medical Systems, 2019, pp. 275 – 280. doi: 10.1109/CBMS.2019.00065.
[8] X. Wang et al., Assessing learning outcomes with interpretable learner models. 2024. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85209599448&partnerID=40&md5=f9d9dcecfb3713ac1965d5f145407cb1
[9] M. Saqr and S. López-Pernas, Explainable Artificial Intelligence in Education: A Tutorial for Identifying the Variables that Matter. 2025. doi: 10.1007/978-3-031-95365-1_6.
[10] M. A. Shakir et al., “Developing Interpretable Models for Complex Decision-Making,” in Conference of Open Innovation Association, FRUCT, 2024, pp. 66 – 75. doi: 10.23919/FRUCT64283.2024.10749922.
[11] D. Gaspar, P. Silva, and C. Silva, “Explainable AI for Intrusion Detection Systems: LIME and SHAP Applicability on Multi-Layer Perceptron,” IEEE Access, vol. 12, pp. 30164 – 30175, 2024, doi: 10.1109/ACCESS.2024.3368377.
[12] Y. Swathi and M. Challa, “A Comparative Analysis of Explainable AI Techniques for Enhanced Model Interpretability,” in Proceedings - 2023 3rd International Conference on Pervasive Computing and Social Networking, ICPCSN 2023, 2023, pp. 229 – 234. doi: 10.1109/ICPCSN58827.2023.00043.
[13] B. Mohammed, “A Review on Explainable Artificial Intelligence Methods, Applications, and Challenges,” Indones. J. Electr. Eng. Informatics, vol. 11, no. 4, pp. 1007 – 1024, 2023, doi: 10.52549/ijeei.v11i4.5151.
[14] F. H. Yagin and A. Pinar, Explainable artificial intelligence in medical research: A synopsis for clinical practitioners—Comprehensive XAI methodologies. 2025. doi: 10.1016/B978-0-443-23979-3.00012-9.
[15] J. Jung, H. Lee, H. Jung, and H. Kim, “Essential properties and explanation effectiveness of explainable artificial intelligence in healthcare: A systematic review,” Heliyon, vol. 9, no. 5, 2023, doi: 10.1016/j.heliyon.2023.e16110.
[16] M. Thamer and Z. N. Sultani, “Explainable AI in the Medical Field: A Survey on Machine Learning Interpretability and Use Cases,” Al-Nahrain J. Sci., vol. 28, no. 4, pp. 188 – 206, 2025, doi: 10.22401/ANJS.28.4.15.
[17] B. P. Sheela and H. Girisha, “An Explainable Artificial Intelligence (XAI) Framework for Deep Learning Based Classification to Generate Textual Explanations on Predicted Images,” Int. J. Intell. Eng. Syst., vol. 17, no. 6, pp. 651 – 662, 2024, doi: 10.22266/ijies2024.1231.50.
[18] S. Verma, S. Avasthi, P. Mishra, M. Gupta, and G. Agarwal, XAI in healthcare: Black box to interpretable models. 2024. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-85192864780&partnerID=40&md5=efda84ca5dd8a898ea0c2ab68c9ff01a
[19] U. Khakurel and D. B. Rawat, “Evaluating Explainable Artificial Intelligence (XAI): Algorithmic Explanations for Transparency and Trustworthiness of ML Algorithms and AI Systems,” in Proceedings of SPIE - The International Society for Optical Engineering, 2022. doi: 10.1117/12.2620598.
[20] S. Roychowdhury, V. Lanfranchi, and S. Mazumdar, “Evaluating explanation performance for clinical decision support systems for non-imaging data: A systematic literature review,” Comput. Biol. Med., vol. 197, 2025, doi: 10.1016/j.compbiomed.2025.110944.
[21] L. Weber, S. Lapuschkin, A. Binder, and W. Samek, “Beyond explaining: Opportunities and challenges of XAI-based model improvement,” Inf. Fusion, vol. 92, pp. 154 – 176, 2023, doi: 10.1016/j.inffus.2022.11.013.
[22] A. Srinivasulu, C. V Ravikumar, S. Mastan, R. S. Vali, and S. Rajeyyagari, Unveiling the Transformative Infuence of Machine Learning Techniques on Business Data Analytics: A Comprehensive Exploration. 2025. doi: 10.1201/9781003640776-1.
[23] J. Narmadha, G. N. Kaavya, and S. D. Preethii, “Analysis on electricity generation forecasting system,” in Proceedings - 2017 IEEE International Conference on Electrical, Instrumentation and Communication Engineering, ICEICE 2017, 2017, pp. 1 – 5. doi: 10.1109/ICEICE.2017.8191901.
[24] A. Aleksieva-Petrova, V. Gancheva, and M. Petrov, “APTITUDE framework for learning data classification based on machine learning,” Int. J. Circuits, Syst. Signal Process., vol. 14, pp. 379 – 385, 2020, doi: 10.46300/9106.2020.14.51.
[25] V. Hassija et al., “Interpreting Black-Box Models: A Review on Explainable Artificial Intelligence,” Cognit. Comput., vol. 16, no. 1, pp. 45 – 74, 2024, doi: 10.1007/s12559-023-10179-8.
[26] C. Rudin, “Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead,” Nat. Mach. Intell., vol. 1, no. 5, pp. 206 – 215, 2019, doi: 10.1038/s42256-019-0048-x.
[27] M. Garouani, J. Mothe, A. Barhrhouj, and J. Aligon, “Investigating the Duality of Interpretability and Explainability in Machine Learning,” in Proceedings - International Conference on Tools with Artificial Intelligence, ICTAI, 2024, pp. 861 – 867. doi: 10.1109/ICTAI62512.2024.00125.
[28] Y.-W. Chen, S.-Y. Chien, and F. Yu, “An overview of XAI Algorithms,” in 2023 International Automatic Control Conference, CACS 2023, 2023. doi: 10.1109/CACS60074.2023.10326174.
[29] S. Bhatnagar and R. Agrawal, “Understanding explainable artificial intelligence techniques: a comparative analysis for practical application,” Bull. Electr. Eng. Informatics, vol. 13, no. 6, pp. 4451 – 4455, 2024, doi: 10.11591/eei.v13i6.8378.
[30] M. Kopzhasarova and D. Kozhamzharova, “Explainable AI (XAI): Techniques, applications, and challenges,” in CEUR Workshop Proceedings, 2025. [Online]. Available: https://www.scopus.com/inward/record.uri?eid=2-s2.0-105006886912&partnerID=40&md5=4007adb38600ba88e07d459c1ef6a021
[31] M. H. Azam, M. H. Hasan, N. Y. Murad, and E. A. Bt Patah, “Transparency in AI: A Review of Explainable Artificial Intelligence Techniques,” in 2024 8th International Conference on Computing, Communication, Control and Automation, ICCUBEA 2024, 2024. doi: 10.1109/ICCUBEA61740.2024.10774981.
[32] A. M. Salih et al., “A Perspective on Explainable Artificial Intelligence Methods: SHAP and LIME,” Adv. Intell. Syst., vol. 7, no. 1, 2025, doi: 10.1002/aisy.202400304.
[33] M. H. Albydhany and I. Al-Darraji, “Explainable AI for Visual Inspection: A Comparative Analysis and Review,” J. Vis. Exp., vol. 2025-October, no. 224, 2025, doi: 10.3791/69440.
[34] M. Panda and S. R. Mahanta, Explainable Artificial Intelligence for Healthcare Applications Using Random Forest Classifier with LIME and SHAP. 2024. doi: 10.1201/9781003442509-6.
[35] J. Duell, M. Seisenberger, and X. Fan, “A Comparison of Global Explanations Given on Electronic Health Records,” Lect. Notes Networks Syst., vol. 794, pp. 413 – 424, 2024, doi: 10.1007/978-3-031-44981-9_34.
[36] S. Kumar, N. Dhanda, and K. K. Gupta, “Exploring Explainable AI Methods for Decoding Machine Learning Model Decisions,” in Proceedings of the 2025 3rd International Conference on Inventive Computing and Informatics, ICICI 2025, 2025, pp. 1083 – 1089. doi: 10.1109/ICICI65870.2025.11069883.


