
datasets. DBSCAN excelled at identifying clusters
of arbitrary shapes and densities, making it well-
suited for non-linear data like the concentric circles.
Nonetheless, its sensitivity to parameters, particularly
the neighborhood radius (eps) and minimum points
(min samples), remains a challenge that warrants
further exploration.
Hierarchical Clustering emerged as a flexible and
interpretable method, particularly through its dendro-
gram visualizations, which allow researchers to an-
alyze cluster structures at various levels of granular-
ity. However, its computational complexity limits its
applicability to larger datasets, making scalability an
area for improvement. The study utilized quantita-
tive performance metrics such as the Silhouette Score,
Adjusted Rand Index, and Calinski-Harabasz Index to
provide an objective evaluation of the clustering re-
sults. These metrics, combined with visual analysis,
offered a holistic understanding of algorithm perfor-
mance under structured and unstructured data condi-
tions.
In summary, this work highlights the importance
of selecting appropriate clustering techniques based
on the underlying data geometry and complexity.
While K-Means is effective for convex and uniform
datasets, DBSCAN and Hierarchical Clustering are
better suited for non-linear and irregular data struc-
tures. Future advancements in hybrid clustering
methods, adaptive parameter tuning, and kernelized
techniques can address the observed limitations and
enhance clustering robustness. This study lays the
groundwork for further exploration of clustering al-
gorithms in real-world scenarios, where data often ex-
hibit noise, complexity, and diverse geometries.
REFERENCES
Arthur, D. and Vassilvitskii, S. (2006). k-means++: The
advantages of careful seeding. In Stanford.
Cui, M. (2020). Introduction to the k-means clustering al-
gorithm based on the elbow method. In Accounting,
Auditing and Finance, volume 1, pages 5–8.
Dandekar, A., Zen, R. A., and Bressan, S. (2018). A com-
parative study of synthetic dataset generation tech-
niques. In Database and Expert Systems Applications,
pages 387–395.
Ester, M., Kriegel, H. P., Sander, J., and Xu, X. (1996).
A density-based algorithm for discovering clusters in
large spatial databases with noise. In Proceedings of
KDD’96, pages 226–231.
Halkidi, M., Batistakisand, Y., and Vazirgiannis, M. (2001).
On clustering validation techniques. Journal of intel-
ligent information systems, 17:107–145.
Hubert, L. and Arabie, P. (1985). Comparing partitions.
Journal of classification, 2:193–218.
Jain, A. K. (2010). Data clustering: 50 years beyond k-
means. Pattern recognition letters, 31(8):651–666.
Madhulatha, T. S. (2012). An overview on clustering meth-
ods. arXiv preprint arXiv:1205.1117.
Murtagh, F. and Contreras, P. (2012). Algorithms for hi-
erarchical clustering: an overview. Wiley Interdisci-
plinary Reviews: Data Mining and Knowledge Dis-
covery, 2(1):86–97.
Pedregosa, F., Varoquaux, G., Gramfort, A., Michel, V.,
Thirion, B., Grisel, O., and Duchesnay, E. (2011).
Scikit-learn: Machine learning in python. The Journal
of Machine Learning Research, 12:2825–2830.
Rousseeuw, P. J. (1987). Silhouettes: a graphical aid to
the interpretation and validation of cluster analysis.
Journal of computational and applied mathematics,
20:53–65.
Saxena, A., Prasad, M., Gupta, A., Bharill, N., et al. (2017).
A review of clustering techniques and developments.
Neurocomputing, 267:664–681.
Shutaywi, M. and Kachouie, N. N. (2021). Silhouette anal-
ysis for performance evaluation in machine learning
with applications to clustering. Entropy, 23(6):759.
Steinbach, M. and Kumar, V. (2003). Finding clusters of
different sizes, shapes, and densities in noisy, high di-
mensional data. In Proceedings of SIAM International
Conference on Data Mining, pages 47–58.
Xu, R. and Wunsch, D. (2005). Survey of clustering al-
gorithms. IEEE Transactions on neural networks,
16(3):645–678.
INCOFT 2025 - International Conference on Futuristic Technology
780