Comparative analysis of interpretable artificial intelligence methods
Article Sidebar
Issue Vol. 38 (2026)
-
Comparative analysis of Next.js and Astro frameworks
Patryk Gieda, Marek Miłosz1-5
-
Behavioral analysis of ransomware threats to ESXi Hypervisors: a machine learning-based predictive model
Upakar Bhatta6-10
-
Performance and usability evaluation of a VR virtual museum application
Kamil Gabrysiewicz, Krzysztof Dziedzic11-18
-
Comparative analysis of selected data visualization methods
Damian Węcławski, Radosław Tomczyk, Paweł Powroźnik19-25
-
A performance comparison of web programming interfaces: GraphQL, gRPC and Thrift
Piotr Rożek, Mariusz Dzieńkowski26-31
-
A review of security mechanisms in electronic payment systems
Omniah ALibrahim, Suhair Alshehri32-42
-
Clustering methods in machine learning
Bartłomiej Głuszczak, Paweł Powroźnik43-50
-
Comparative analysis of interpretable artificial intelligence methods
Aleksandra Kuszewska, Małgorzata Charytanowicz51-58
-
Comparative analysis of machine learning classifiers
Łukasz Krukowski, Grzegorz Kozieł59-65
-
Analysis of the impact of machine learning algorithms on the quality of generated sounds
Krzysztof Pedrycz, Mateusz Pikula66-72
-
Comparative analysis of the functionalities of applications supporting the self-control process of anticoagulation therapy
Marcin Furmaga, Vitalii Baida73-80
-
Application of machine learning for predicting Formula 1 race results
Sylwia Krzysztoń, Jakub Smołka81-86
-
Analysis of latency reduction and performance improvement methods in selected VR applications
Mateusz Czapczyński, Krzysztof Dziedzic87-94
-
Integrating deep learning image analysis into Web GIS applications: A Hybrid Flask - Spring Boot architecture for automated place detection
Medjon HYSENAJ95-101
Main Article Content
DOI
Sustainable Development Goals (SDG)
- Quality education
- Industry, Innovation, Technology and Infrastructure
Authors
Abstract
The aim of this article is to analyze and compare methods for explaining the results of artificial intelligence methods. Three methods were analyzed: Grad-CAM, SHAP, and LIME, evaluated in terms of their effectiveness on different data types. The analysis used five datasets: Iris, Wine Quality, Brain Tumor Dataset, PHCD, and WheatGrain. Two datasets are tabular, two are image, and one is mixed. SHAP and LIME were applied to tabular datasets, while all three methods were used for image data. Grad-CAM proved the fastest and most effective in locating key regions, while SHAP was slower but more accurate in pixel attribution, and LIME achieved the lowest precision. For tabular data, SHAP provided more accurate and consistent explanations than LIME, especially for high-dimensional datasets.
Keywords:
References
[1] Y. N. Harari, Nexus. Krótka historia informacji – od epoki kamienia do sztucznej inteligencji, Wydawnictwo Literackie, Kraków, 2024.
[2] Y. LeCun, Y. Bengio, G. Hinton, Deep learning, Nature 521(7553) (2015) 436–444, https://doi.org/10.1038/nature14539. DOI: https://doi.org/10.1038/nature14539
[3] D. Pedreschi, F. Giannotti, R. Guidotti, A. Monreale, S. Ruggieri, S. Turini, Meaningful Explanations of Black Box AI Decision Systems, Proceedings of the 33rd AAAI Conference on Artificial Intelligence (2019) 9780–9784, https://doi.org/10.1609/aaai.v33i01.33019780. DOI: https://doi.org/10.1609/aaai.v33i01.33019780
[4] Survey results on AI usage by developers, Stack Overflow, Developer Survey 2024 – AI, https://survey.stackoverflow.co/2024/ai#developer-tools-ai-acc, [22.09.2025].
[5] A. B. Arrieta, N. Díaz-Rodríguez, J. Del Ser, A. Bennetot, S. Tabik, A. Barbado, S. García, S. Gil-Lopez, D. Molina, R. Benjamins, R. Chatila, F. Herrera, Explainable Artificial Intelligence (XAI): Concepts, taxonomies, opportunities and challenges toward responsible AI, Information Fusion 58 (2020) 82–115, https://doi.org/10.1016/j.inffus.2019.12.012. DOI: https://doi.org/10.1016/j.inffus.2019.12.012
[6] S. Nazim, M. M. Alam, S. S. Rizvi, J. C. Mustapha, S. S. Hussain, M. M. Suud, Advancing malware imagery classification with explainable deep learning: A state-of-the-art approach using SHAP, LIME and Grad-CAM, PLoS ONE 20(5) (2025) e0318542, https://doi.org/10.1371/journal.pone.0318542. DOI: https://doi.org/10.1371/journal.pone.0318542
[7] Polish Handwritten Characters Database (PHCD), https://cs.pollub.pl/phcd/, [22.09.2025].
[8] WheatGrain Images, https://zenodo.org/records/15172506, [22.09.2025].
[9] Brain Tumor Dataest, https://www.kaggle.com/datasets/ preetviradiya/brian-tumor-dataset, [22.09.2025].
[10] Iris Dataset, https://www.kaggle.com/datasets/hima nshunakrani/iris-dataset, [22.09.2025].
[11] Wine Quality Dataset, https://www.kaggle.com/dat asets/yasserh/wine-quality-dataset, [22.09.2025].
[12] WheatGrain Features, https://zenodo.org/records/15172506, [22.09.2025].
[13] M. Charytanowicz, J. Niewczas, P. Kulczycki, P. A. Kowalski, S. Łukasik, Discrimination of Wheat Grain Varieties Using X-ray Images, In: E. Pietka, P. Badura, J. Kawa, W. Więcławek (eds), Information Technologies in Biomedicine. Advances in Intelligent Systems and Computing, Springer 471 (2016) 39–50, https://doi.org/10.1007/978-3-319-39796-2_4. DOI: https://doi.org/10.1007/978-3-319-39796-2_4
[14] R. R. Selvaraju, M. Cogswell, A. Das, R. Vedantam, D. Parikh, D. Batra, Grad-CAM: Visual Explanations from Deep Networks via Gradient-based Localization, International Journal of Computer Vision 128(2) (2020) 336–359, https://doi.org/10.1007/s11263-019-01228-7. DOI: https://doi.org/10.1007/s11263-019-01228-7
[15] V. Vimbi, N. Shaffi, M. Mahmud, Interpreting Artificial Intelligence Models: A Systematic Review on the Application of LIME and SHAP in Alzheimer’s Disease Detection, Brain Informatics 11 (2024) 10, https://doi.org/10.1186/s40708-024-00222-1. DOI: https://doi.org/10.1186/s40708-024-00222-1
[16] A. Salih, Z. Raisi-Estabragh, I. B. Galazzo, P. Radeva, S. E. Petersen, G. Menegaz, K. Lekadir, A Perspective on Explainable Artificial Intelligence Methods: SHAP and LIME, Advanced Intelligent Systems 7(1) (2024) 2400304, https://doi.org/10.1002/aisy.202400304. DOI: https://doi.org/10.1002/aisy.202400304
[17] A. Kuanar, A. Akbar, P. Sujata, D. Kar, Deep Neural Network (DNN) Modelling for Prediction of the Mode of Delivery, European Journal of Obstetrics & Gynecology and Reproductive Biology 297 (2024) 241–248, https://doi.org/10.1016/j.ejogrb.2024.04.012. DOI: https://doi.org/10.1016/j.ejogrb.2024.04.012
[18] S. J. Rigatti, Random Forest, Journal of the American Medical Informatics Association 24(6) (2017) 1015–1024, https://doi.org/10.1093/jamia/ocx133. DOI: https://doi.org/10.1093/jamia/ocx133
[19] G. Huang, Z. Liu, L. van der Maaten, K. Q. Weinberger, Densely Connected Convolutional Networks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2017) 2261–2269, https://doi.org/10.1109/CVPR.2017.243. DOI: https://doi.org/10.1109/CVPR.2017.243
[20] M. Sandler, A. Howard, M. Zhu, A. Zhmoginov, L.-C. Chen, MobileNetV2: Inverted Residuals and Linear Bottlenecks, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2018) 4510–4520, https://doi.org/10.1109/CVPR.2018.00474. DOI: https://doi.org/10.1109/CVPR.2018.00474
[21] M. Pawlicki, A. Pawlicka, F. Uccello, S. Szelest, S. D’Antonio, R. Kozik, M. Choraś, Evaluating the necessity of the multiple metrics for assessing explainable AI: A critical examination, Neurocomputing 602 (2024) 128282, https://doi.org/10.1016/j.neucom.2024.128282. DOI: https://doi.org/10.1016/j.neucom.2024.128282
[22] W. J. Yeo, W. van der Heever, R. Mao, E. Cambria, R. Satapathy, G. Mengaldo, A comprehensive review on financial explainable AI, Artificial Intelligence Review 58 (2025) 189, https://doi.org/10.1007/s10462-024-11077-7. DOI: https://doi.org/10.1007/s10462-024-11077-7
[23] D. S. Watson, L. Gultchin, A. Taly, L. Floridi, Local Explanations via Necessity and Sufficiency: Unifying Theory and Practice, Minds and Machines 32(3) (2022) 475–498, https://doi.org/10.1007/s11023-022-09598-7. DOI: https://doi.org/10.1007/s11023-022-09598-7
[24] M. Saarela, L. Geogieva, Robustness, Stability, and Fidelity of Explanations for a Deep Skin Cancer Classification Model, Applied Sciences 12(19) (2022) 9545, https://doi.org/10.3390/app12199545. DOI: https://doi.org/10.3390/app12199545
[25] L. Coroamă, A. Groza, Evaluation Metrics in Explainable Artificial Intelligence (XAI), Proceedings of the International Conference on Advanced Research in Technologies, Information, Innovation and Sustainability (ARTIIS) (2022) 401–413, https://doi.org/10.1007/978-3-031-20319-0_30. DOI: https://doi.org/10.1007/978-3-031-20319-0_30
Article Details
Abstract views: 7

