Clustering methods in machine learning
Article Sidebar
Issue Vol. 38 (2026)
-
Comparative analysis of Next.js and Astro frameworks
Patryk Gieda, Marek Miłosz1-5
-
Behavioral analysis of ransomware threats to ESXi Hypervisors: a machine learning-based predictive model
Upakar Bhatta6-10
-
Performance and usability evaluation of a VR virtual museum application
Kamil Gabrysiewicz, Krzysztof Dziedzic11-18
-
Comparative analysis of selected data visualization methods
Damian Węcławski, Radosław Tomczyk, Paweł Powroźnik19-25
-
A performance comparison of web programming interfaces: GraphQL, gRPC and Thrift
Piotr Rożek, Mariusz Dzieńkowski26-31
-
A review of security mechanisms in electronic payment systems
Omniah ALibrahim, Suhair Alshehri32-42
-
Clustering methods in machine learning
Bartłomiej Głuszczak, Paweł Powroźnik43-50
-
Comparative analysis of interpretable artificial intelligence methods
Aleksandra Kuszewska, Małgorzata Charytanowicz51-58
-
Comparative analysis of machine learning classifiers
Łukasz Krukowski, Grzegorz Kozieł59-65
-
Analysis of the impact of machine learning algorithms on the quality of generated sounds
Krzysztof Pedrycz, Mateusz Pikula66-72
-
Comparative analysis of the functionalities of applications supporting the self-control process of anticoagulation therapy
Marcin Furmaga, Vitalii Baida73-80
-
Application of machine learning for predicting Formula 1 race results
Sylwia Krzysztoń, Jakub Smołka81-86
-
Analysis of latency reduction and performance improvement methods in selected VR applications
Mateusz Czapczyński, Krzysztof Dziedzic87-94
-
Integrating deep learning image analysis into Web GIS applications: A Hybrid Flask - Spring Boot architecture for automated place detection
Medjon HYSENAJ95-101
Main Article Content
DOI
Sustainable Development Goals (SDG)
- Quality education
Authors
Abstract
Clustering is a fundamental task in machine learning for discovering hidden structures in unlabelled data. The article reviews key clustering methods, including centroid, density, hierarchy and model-based approaches. Their advantages, limitations and applications are analysed to provide a comprehensive overview of the state of clustering in machine learning. Their effectiveness is compared on the basis of selected metrics to evaluate the outcome of a given clustering. Recent developments and challenges, including scalability and interpretability problems, are also discussed.
Keywords:
References
[1] M.W. Berry, A. Mohamed, B.W. Yap, Supervised and Unsupervised Learning for Data Science, Springer, Cham, 2019. DOI: https://doi.org/10.1007/978-3-030-22475-2
[2] C.K. Reddy, C.C. Aggarwal, Data Clustering: Algorithms and Applications, CRC Press, Boca Raton, 2014.
[3] J.B. MacQueen, Some methods for classification and analysis of multivariate observations, Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Vol. 1: Statistics, University of California Press, Berkeley (1967) 281-297.
[4] L. Kaufman, P.J. Rousseeuw, Finding Groups in Data: An Introduction to Cluster Analysis, Wiley, New York, 1990. DOI: https://doi.org/10.1002/9780470316801
[5] M. Ester, H.-P. Kriegel, J. Sander, X. Xu, A density-based algorithm for discovering clusters in large spatial databases with noise, Proceedings of the 2nd International Conference on Knowledge Discovery & Data Mining (1996) 226-231, https://doi.org/10.5555/3001460.3001507.
[6] M. Suyal, S. Sharma, A review on analysis of K-means clustering machine learning algorithm based on unsupervised learning, Journal of Artificial Intelligence and Systems 6 (2024) 85-95, https://doi.org/10.33969/AIS.2024060106. DOI: https://doi.org/10.33969/AIS.2024060106
[7] R.R. Muhima, M. Kurniawan, O.T. Pambudi, A LOF K-means clustering on hotspot data, International Journal of Artificial Intelligence & Robotics 2(1) (2020) 29-33, https://doi.org/10.25139/ijair.v2i1.2634. DOI: https://doi.org/10.25139/ijair.v2i1.2634
[8] M. Ahmed, R. Seraj, S.M.S. Islam, The K-means algorithm: A comprehensive survey and performance evaluation, Electronics 9(8) (2020) 1295, https://doi.org/10.3390/electronics9081295. DOI: https://doi.org/10.3390/electronics9081295
[9] C. Larasvaty, S. Khomsah, R. Nisa, Cluster analysis of COVID-19 in Indonesia using K-means method, Journal of DINDA: Data Science, Information Technology, and Data Analytics 3(1) (2023) 36-39, https://doi.org/10.20895/dinda.v3i1.822. DOI: https://doi.org/10.20895/dinda.v3i1.822
[10] X. Huang, Z. Song, Clustering analysis on e-commerce transaction based on K-means clustering, Journal of Networks 9(2) (2014) 443-450, https://doi.org/10.4304/jnw.9.2.443-450. DOI: https://doi.org/10.4304/jnw.9.2.443-450
[11] W.L. Zhang, G.X. Li, W. Gao, The research of speech emotion recognition based on Gaussian mixture model, Applied Mechanics and Materials 668-669 (2014) 1126-1129, https://doi.org/10.4028/www.scientific.net/AMM.668-669.1126. DOI: https://doi.org/10.4028/www.scientific.net/AMM.668-669.1126
[12] A. Khoshkbarforoushha, R. Ranjan, P. Strazdins, Resource distribution estimation for data-intensive workloads: Give me my share & no one gets hurt!, In Advances in Service-Oriented and Cloud Computing (2016) 228-237. DOI: https://doi.org/10.1007/978-3-319-33313-7_17
[13] M. Hahsler, M. Piekenbrock, D. Doran, dbscan: Fast density-based clustering with R, Journal of Statistical Software 91(1) (2019) 1-30, https://doi.org/10.18637/jss.v091.i01. DOI: https://doi.org/10.18637/jss.v091.i01
[14] D. Comaniciu, P. Meer, Mean shift: A robust approach toward feature space analysis, IEEE Transactions on Pattern Analysis and Machine Intelligence 24(5) (2002) 603-619, https://doi.org/10.1109/34.1000236. DOI: https://doi.org/10.1109/34.1000236
[15] K.G. Derpanis, Mean Shift Clustering York University, 2005, https://www.cse.yorku.ca/~kosta/CompVis_Notes/mean_shift.pdf [16.09.2025].
[16] Y. Li, H. Jiang, J. Lu, X. Li, Z. Sun, M. Li, MR-BIRCH: A scalable MapReduce-based BIRCH clustering algorithm, Journal of Intelligent & Fuzzy Systems 40(3) (2021) 5295-5305, https://doi.org/10.3233/JIFS-202079. DOI: https://doi.org/10.3233/JIFS-202079
[17] T. Hastie, R. Tibshirani, J. Friedman, The Elements of Statistical Learning, 2nd ed., Springer, New York, 2009. DOI: https://doi.org/10.1007/978-0-387-84858-7
Article Details
Abstract views: 8

