# Introduction lustering is the process of organizing data objects into a set of disjoint classes called Clusters. Clustering is an Unsupervised Clustering technique of Classification. Classification refers to a technique that assigns data objects to a set of classes. Unsupervised means that clustering does not depends upon predefined classes while clustering the data objects. Formally, given a set of dimensional points and a function that gives the distance between two points , we are required to compute cluster centers, such that the points falling in the same cluster are similar and points that are in different cluster are dissimilar. Most of the initial clustering techniques were developed by statistics or pattern recognition communities, where the goal was to cluster a modest number of data instances. However, within the data mining community, the focus has been on clustering large datasets [2]. Developing clustering algorithms to effectively and efficiently cluster rapidly growing datasets has been identified as an important challenge. A number of clustering algorithms have been proposed to solve clustering problems. One of the most popular clustering methods are K-Means, SOM, HCA. Their shortcomings are discussed below. The standard k-means algorithm needs to calculate the distance from the each data object to all Author : e-mail: er.amantoor@gmail.com the centers of k clusters when it executes the iteration each time, which takes up a lot of execution time especially for large-capacity databases. In K-Means algorithm initial cluster centers are produced arbitrary, it does not promise to produce the peculiar clustering results. Efficiency of original k-means algorithm is heavily rely on the initial centroid. Initial centroid also has an influence on the number of iterations required while running the original K-Means algorithm. Computational Complexity of K-Means algorithm is very high and does not provide high quality clusters when it comes to cluster High dimensional data set. Kohonon's SOMs are a type of unsupervised learning. The goal is to discover some underlying structure of the data. SOM algorithm is computationally expensive. Large quantity of good quality representative training data required. No generally accepted measure of 'quality' of a SOM e.g. Average quantization error (how well the data is classified). Every SOM is different therefore we must be careful what conclusions we draw from our results. SOM is non-deterministic and can and will produce different results in different run. Hierarchical clustering algorithms are either topdown or bottom-up. Bottom-up algorithms treat each document as a singleton cluster at the outset and then successively merge (or agglomerate) pairs of clusters until all clusters have been merged into a single cluster that contains all documents. Bottom-up hierarchical clustering is therefore called hierarchical agglomerative clustering or HAC. [6] Top-down clustering requires a method for splitting a cluster. It proceeds by splitting clusters recursively until individual documents are reached. This algorithm is sensitive to outliers and sometimes it is difficult to identify the correct number of clusters from Dendrogram. [7] Various methods have been proposed in literature but it have been analyzed that the K-Means, SOM, HCA fails to give optimum result when it comes to clustering high dimensional data set because their complexity tends to make things more difficult when number of dimensions are added. In data mining this problem is known as "Curse of Dimensionality". This research will deal the problem of high dimensionality and large data set. A large number of algorithms had been proposed till date, each of them address some specific requirement. There does not exist a single algorithm which can adequately handle all sorts of requirement. This makes a great challenge for the user to do selection among the available algorithm for specific task. To cope with this problem, a new algorithm is going to be proposed in this research that is named as "Advanced Clustering Algorithm". This paper is organized s follows. Section 2 presents an overview of ACA. Section 3 introduces proposed method. Section 4 describes about the time complexity of proposed method. Section 5 experimentally demonstrates the performance of proposed method. And the final Section 6 describes the conclusion. # II. # Advanced Clustering Algorithm Experimental results have shown Kohonon's SOM is superlative clustering algorithm among Kmeans, HCA [8]. For the shortcomings of the above SOM algorithm, this paper presents an Advanced Clustering Algorithm method. The main idea of algorithm is to set two simple data structures to retain the labels of cluster and the distance of all the data objects to the nearest cluster during the each iteration that can be used in next iteration. We calculate the distance between the current data object and the new cluster center, if the computed distance is smaller than or equal to the distance to the old center, the data object stays in it's cluster that was assigned to in previous iteration. Therefore, there is no need to calculate the distance from this data object to the other k-1 clustering centers, saving the calculative time to the k-1 cluster centers. Otherwise, we must calculate the distance from the current data object to all k cluster centers, and find the nearest cluster center and assign this point to the nearest cluster center. And then we separately record the label of nearest cluster center and the distance to it's center. Because in each iteration some data points still remain in the original cluster, it means that some parts of the data points will not be calculated, saving a total time of calculating the distance, thereby enhancing the efficiency of the algorithm. # III. # Proposed Algorithm The process of the Advanced Clustering algorithm is described as follows: Input: The number of desired clusters k, and a database D= {d 1 , d 2 , d n } containing n data objects. Output: A set of k clusters. 1. Draw multiple sub-samples {SI, S2, . . . ,Sj } from the original dataset. 2. Repeat step 3 for m=l to i 3. Apply combined approach for sub sample. IV. # Time Complexity This paper proposes an Advanced Clustering Algorithm, to obtain the initial cluster, time complexity of the advanced algorithm is O (nk). Here some data points remain in the original clusters, while the others move to other clusters. If the data point retains in the original cluster, this needs O (1), else O (k). With the convergence of clustering algorithm, the number of data points moved from their cluster will reduce. If half of the data points move from their cluster, the time complexity is O (nk/2). Hence the total time complexity is O (nk). While the standard k-means clustering algorithm require O(nkt). So the proposed algorithm in this paper can effectively improve the speed of clustering and reduce the computational complexity. But the Advanced kmeans algorithm requires the pre estimated the number of clusters, k, which is the same to the standard kmeans algorithm. If you want to get to the optimal solution, you must test the different value of k. # V. # Experimental Results This paper selects academic data set repository of machine learning databases to test the efficiency of the advanced algorithm and the standard algorithms. Two simulated experiments have been carried out to demonstrate the performance of the Advanced in this paper. This algorithm has also been applied to the clustering of real datasets. In two experiments, time taken for each experiment is computed. The same data set is given as input to the standard algorithm and the Advanced Clustering Algorithm. Experiments compare Advanced Clustering Algorithm with the standard algorithm in terms of the total execution time of clusters and their accuracy. Experimental operating system is Window 8, program language is java. This paper uses academic activities as the test datasets and gives a brief description of the datasets used in experiment evaluation. Table 1 shows some characteristics of the datasets. # Conclusion SOM algorithm is a typical clustering algorithm and it is widely used for clustering large sets of data. This paper elaborates Advanced Clustering Algorithm and analyses the shortcomings of the standard kmeans, SOM and HAC clustering algorithm. Because the computational complexity of the standard algorithm is objectionably high owing to the need to reassign the data points a number of times during every iteration, which makes the efficiency of standard clustering is not high. This paper presents a simple and efficient way for assigning data points to clusters. The proposed method ![Global Journal of Computer Science and TechnologyVolume XIV Issue II Version I](image-2.png "C") 56![Choose minimum of minimum distance from cluster center criteria Now apply new calculation again on dataset S for k1 clusters. 7. Combine two nearest clusters into one cluster and recalculate the new cluster center for the combined cluster until the number of clusters reduces into k.](image-3.png "4. Compute centroid 5 . 6 .") 1![Figure 1 : Display data set according to class attributes](image-4.png "Figure 1 :") 23![Figure 2 : Display All Attributes](image-5.png "Figure 2 :Figure 3 :") 1DatasetNumber ofNumber ofattributesrecordsAcademicActivities 2Clustering AlgorithmParameterSOMK-HACECAMeansError Rate 0.8189 0.8456 0.8379 0.3672Execution297128113411000TimemsmsmsmsAccessingFastSlowSlowVeryTimefastNumber of6664Clusters © 2014 Global Journals Inc. (US)Cluster analysis method is one of the main analytical methods in data mining; this method of clustering algorithm will influence the clustering results directly. This paper proposes an Advanced Clustering Algorithm in order to solve this question, requiring a simple data structure to store some information[1] in every iteration, which is to be used in the next iteration. The Advanced Clustering Algorithm method avoids computing the distance of each data object to the cluster centers repeat, saving the running time. Experimental results show that the Advanced Clustering Algorithm method can effectively improve the speed of clustering and accuracy, reducing the computational complexity of the traditional algorithm. This paper includes Advanced Clustering Algorithm (ACA) and describes the experimental results and conclusions through experimenting with academic data sets.Abstract- © 2014 Global Journals Inc. (US)An Advanced Clustering Algorithm (ACA) for Clustering Large Data Set to Achieve High Dimensionality ## Global Journals Inc. (US) Guidelines Handbook 2014 www.GlobalJournals.org * A New Algorithm to Get the Initial Centroids FYuan ZHMeng HZhang DongC Proc. of the 3rd International Conference on Machine Learning and Cybernetics of the 3rd International Conference on Machine Learning and Cybernetics August 2004 * Clustering algorithms Research SunJigui LiuJie ZhaoLianyu Journal of Software 19 1 January 2008 * Research on Modified kmeans Data Cluster Algorithm SunShibao ;I SQin Keyun CPJacobs Bean Computer Engineering 33 13 July 2007 Fine particles, thin films and exchange anisotropy * UCI Repository of Machine Learning Databases CMerz PMurphy * An efficient enhanced k-means clustering algorithm A MFahim A MSalem FTorkey Journal of Zhejiang University Science A 10 July 2006 * A grid-based density isoline clustering algorithm YCZhao SongJGdilc Proc. of theInternet Conf. on Info-Net YXZhong SCui YYang of theInternet Conf. on Info-NetBeijing IEEE Press 2001 140 * A Survey paper on recent clustering approaches in data mining AmarpreetAmanpreet Kaur Toor Singh November 2013 3 * Analysis of Clustering Algorithm based on Number of Clusters, error rate, Computation Time and Map Topology on large Data Set AmarpreetAmanpreet Kaur Toor Singh November-December 2013 2 * Extensions to the k-means algorithm for clustering large data sets with categorical values ZHuang Data Mining and Knowledge Discovery 2 1998 * Improving the Accuracy and Efficiency of the k-means Clustering Algorithm KANazeer MPSebastian Proceeding of the World Congress on Engineering eeding of the World Congress on Engineering July 2009 1 * Partitionalvs hierarchical clustering using a minimum grammar compexity approach AlnFred JmnLeitão Proc. of the SSPR & SPR 2000 of the SSPR & SPR 2000 2000.193?202 1876 * Hempel's raven paradox: A positive approach to cluster analysis RGelbard ISpiegler Computers and Operations Research 27 4 2000 * A fast clustering algorithm to cluster very large categorical data sets in data mining ZHuang Proc. of the SIGMOD Workshop on Research Issues on Data Mining and Knowledge Discovery of the SIGMOD Workshop on Research Issues on Data Mining and Knowledge DiscoveryTucson 1997 * Nearest-Neighbor in data clustering: Incorporating local information into global optimization CDing XHe K- Proc. of the ACM Symp. on Applied Computing of the ACM Symp. on Applied ComputingNicosia ACM Press 2004. 2004 584 * An efficient data clustering method for very large databases TZhang RRam Akrishnan .Birch Proc. of the HVJagadish ISMumick of the 1996 * An algorithm for clustering spatial-temporal data DBirant AKut St-Dbscan ACM SIGMOD Int'l Conf. on Management of Data Montreal ACM Press 1996. 2007 103