
www.rsisinternational.org
INTERNATIONAL JOURNAL OF LATEST TECHNOLOGY IN ENGINEERING,
MANAGEMENT & APPLIED SCIENCE (IJLTEMAS)
ISSN 2278-2540 | DOI: 10.51583/IJLTEMAS | Volume XV, Issue IV, April 2026
LITERATURE REVIEW
Martin et al (1996) proposed a clustering technique called DBSCAN. The method was developed in discovering
clusters of arbitrary shape because the shape of clusters in spatial datasets may be spherical, draw-out, linear,
elongated etc. The method DBSCAN is from the word Density based spatial clustering of applications with
Noise. The researchers use neighborhood and distance in the clustering of object for effective heuristic
parameters. The method was compared with CLARANS and it was concluded that DBSCAN performed better
in handling large dataset and also significantly more effective in discovering clusters of arbitrary shape.
Tian, Raghu and Miran (1996) proposed a clustering technique called BIRCH. The method was referred to as an
efficient clustering method that can be used for a very large dataset. The method birch is from the word balanced
iterative reducing and clustering using hierarchies. It was stated that the proposed method can handle noise which
is a situation where data points are not part of the underlying pattern or structure of the given set of observations.
The method was compared with an existing clustering method called CLARANS and it was concluded that Birch
is better than Clarans in handling large dataset.
Ding and He (2002) provide a comprehensive analysis and experiments on divisive and agglomerative clustering.
Similarities were used instead of the traditional distances. For divisive clustering, 4 new cluster selection criteria
was introduced, the average similarity, the cluster cohesion, avg-cohesion, and temporary objective. Based on
the 4 new cluster selection criteria introduced on internet newsgroups, it is shown that average similarity and
similarity-cohesion selection perform well. For agglomerative clustering, MinMax linkage was introduced and
compared with single-linkage, complete linkage and average linkage. MinMax linkage were found to be most
effective than other in merging clusters.
Almeida et al 2007 presented a new method of hierarchical cluster analysis capable of detecting outlier and
automatic clustering. This technique is based on agglomerative hierarchical clustering which is one of the most
frequent approaches in unsupervised clustering. The algorithm comprises of three steps namely; treatment of
outlier (outlier control), blocks identifier and group of blocks using similarity approach. The resulting
classification method was validated by comparing it with some of the traditional methods. It was observed that
the new method outperform some of the traditional methods in terms of clustering objects.
Basu and Murty (2013) developed am hierarchical clustering method called CUES. The method CUES is from
the Clustering Using Extensive Similarity. The method were developed using a new cluster distance measure to
identify two dissimilar clusters, it will never merger them but can stopped the algorithm if the distance between
two clusters became very high. The method was compared with existing methods (single-link hierarchical,
average–link hierarchical, bisecting k-means, buckshot, k nearest neighbor and it was found that CURE perform
significantly better than other existing clustering.
Yogita and Harish (2013) stated some improved hierarchical clustering algorithms like CURE, BIRCH,
CHEMELEON, Linkage, Leaders–Subleaders, Bisecting k-Means that overcome the limitations that exist in
pure hierarchical clustering algorithms. They went further to give some criteria on the basis of which one can
also determine the best among these mentioned algorithms.
May and Moe (2016) presented a modified agglomerative hierarchical clustering (MAHC) which is an
improvement on commonly used hierarchical clustering method. The modification was done to address the
shortcomings of AHC such as precision, measurement and dimensionality. Instead of frequency of occurrency,
the modified method used item based collation for the construction of distance matrix. It was concluded that
MAHC has shorter algorithm and grouped items in clusters better than the AHC method.
Pelin and Derya (2017) proposed a hierarchical clustering method called K- Linkage which evaluates the distance
between two clusters by calculating the average distance between k pairs of observations. The researchers
introduces two concepts k-min linkage considers k minimum (closet) pairs from points in the first cluster to
points in the second cluster, k-max linkage takes into account k maximum (farthest) pairs of observations. The
improved hierarchical clustering algorithm based on k-linkage was executed on five well-known benchmark