; Enhancing K-Means Algorithm with Semi-Unsupervised Centroid Selection Method
Learning Center
Plans & pricing Sign in
Sign Out
Your Federal Quarterly Tax Payments are due April 15th Get Help Now >>

Enhancing K-Means Algorithm with Semi-Unsupervised Centroid Selection Method


The International Journal of Computer Science and Information Security (IJCSIS) is a well-established publication venue on novel research in computer science and information security. The year 2010 has been very eventful and encouraging for all IJCSIS authors/researchers and IJCSIS technical committee, as we see more and more interest in IJCSIS research publications. IJCSIS is now empowered by over thousands of academics, researchers, authors/reviewers/students and research organizations. Reaching this milestone would not have been possible without the support, feedback, and continuous engagement of our authors and reviewers. Field coverage includes: security infrastructures, network security: Internet security, content protection, cryptography, steganography and formal methods in information security; multimedia systems, software, information systems, intelligent systems, web services, data mining, wireless communication, networking and technologies, innovation technology and management. ( See monthly Call for Papers) We are grateful to our reviewers for providing valuable comments. IJCSIS December 2010 issue (Vol. 8, No. 9) has paper acceptance rate of nearly 35%. We wish everyone a successful scientific research year on 2011. Available at http://sites.google.com/site/ijcsis/ IJCSIS Vol. 8, No. 9, December 2010 Edition ISSN 1947-5500 � IJCSIS, USA.

More Info
  • pg 1
									                                                               (IJCSIS) International Journal of Computer Science and Information Security,
                                                               Vol. 8, No. 9, December 2010

             Enhancing K-Means Algorithm with
         Semi-Unsupervised Centroid Selection Method
                                             R. Shanmugasundaram and Dr. S. Sukumaran
                                                                                  number of clusters (K), which are represented by their
  Abstract— The k-means algorithm is one of the frequently used                   centroids.
clustering methods in data mining, due to its performance in                         The K-means algorithm is as follows:
clustering massive data sets. The final clustering result of the kmeans           1. Select initial centers of the K clusters. Repeat the steps 2
clustering algorithm is based on the correctness of the initial                   through 3 until the cluster membership stabilizes.
centroids, which are selected randomly. The original k-means
                                                                                  2. Generate a new partition by assigning each the data to its
algorithm converges to local minimum, not the global optimum. The
                                                                                  closest cluster centers.
k-means clustering performance can be enhanced if the initial cluster
centers are found. To find the initial cluster centers a series of                3. Compute new cluster centers as centroids of the clusters.
procedure is performed. Data in a cell is partitioned using a cutting                Though K-means is simple and can be used for a wide
plane that divides cell in two smaller cells. The plane is perpendicular          variety of data types, it is quite sensitive to initial positions of
to the data axis with very high variance and is intended to minimize              cluster centers. The final cluster centroids may not be optimal
the sum squared errors of the two cells as much as possible, while at             ones as the algorithm can converge to local optimal solutions.
the same time keep the two cells far apart as possible. Cells are                 An empty cluster can be attained if no points are allocated to
partitioned one at a time until the number of cells equals to the                 the cluster during the assignment step. Therefore, it is
predefined number of clusters, K. The centers of the K cells become
                                                                                  important for K-means to have good initial cluster centers [15,
the initial cluster centers for K-means. In this paper, an efficient
method for computing initial centroids is proposed. A Semi
                                                                                  16]. In this paper a Semi-Unsupervised Selection Method
Unsupervised Centroid Selection Method is used to compute the                     (SCSM) is presented. The organization of this paper is as
initial centroids. Gene dataset is used to experiment the proposed                follows. In the next section, the literature survey is presented.
approach of data clustering using initial centroids. The experimental             In Section III, efficient semi-unsupervised centroid selection
results illustrate that the proposed method is vey much apt for the               algorithm is presented. The experimental results and are
gene clustering applications.                                                     presented in Section IV. Section V concludes the paper.
   Index Terms— Clustering algorithm, K-means algorithm, Data
partitioning, initial cluster centers, semi-unsupervised gene selection.                             II.     LITERATURE SURVEY

                         I.     INTRODUCTION                                         Clustering statistical data has been studied from early time
                                                                                  and lots of advanced models as well as algorithms have been

C    LUSTERING, or unsupervised classification, will be
     considered as a mixture of problem where the aim is to
partition a set of data object into a predefined number of
                                                                                  proposed. This section of the paper provides a view on the
                                                                                  related research work in the field of clustering that may assist
                                                                                  the researchers.
clusters [13]. Number of clusters might be established by                            Bradley and Fayyad together in [2] put forth a technique for
means of the cluster validity criterion or described by user.                     refining initial points for clustering algorithms, in particular k-
Clustering problems are broadly used in many applications,                        means clustering algorithm. They presented a fast and efficient
such as customer segmentation, classification, and trend                          algorithm for refining an initial starting point for a general
analysis. For example, consider that customers purchased a                        class of clustering algorithms. The iterative techniques that are
retail database records containing items. A clustering method                     more sensitive to initial starting conditions were used in most
could group the customers in such a way that customers with                       of the clustering algorithms like K-means, and EM normally
similar buying patterns are in the same cluster. Several real-                    converges to one local minima. They implemented this
word applications deal with high dimensional data. It is                          iterative technique for refining the initial condition which
always a challenge for clustering algorithms because of the                       allows the algorithm to converge to a better local minimum
manual processing is practically not possible. A high quality                     value. The refined initial point is used to evaluate the
computer-based clustering removes the unimportant features                        performance of K-means algorithm in clustering the given
and replaces the original set by a smaller representative set of                  data set. The results illustrated that the refinement run time is
data objects.                                                                     significantly lower than the time required to cluster the full
   K-means is a well known prototype-based [14], partitioning                     database. In addition, the method is scalable and can be
clustering technique that attempts to find a user-specified                       coupled with a scalable clustering algorithm to concentrate on
                                                                                  the large-scale clustering problems especially in case of data
   R. Shanmugasundram, Associate Professor, Department of Computer                   Yang et al. in [3] proposed an efficient data clustering
Science, Erode Arts & Science College, Erode, India.                              algorithm. It is well known that K-means (KM) algorithm is
   Dr. S. Sukumaran, Associate Professor, Department of Computer Science,         one of the most popular clustering techniques because it is
Erode Arts and Science College, Erode, India.

                                                                            337                                http://sites.google.com/site/ijcsis/
                                                                                                               ISSN 1947-5500
                                                        (IJCSIS) International Journal of Computer Science and Information Security,
                                                        Vol. 8, No. 9, December 2010

unproblematic to implement and works rapid in most                           7. Merge the data items which have the same pattern string
situations. But the sensitivity of KM algorithm to initialization         Pt yielding K′ clusters. The centroids of the K′ clusters are
makes it easily trapped in local optima. K-Harmonic Means                 computed. If K′ > K, apply Merge- DBMSDC (Density based
(KHM) clustering resolves the problem of initialization faced             Multi Scale Data Condensation) algorithm [6] to merge these
by KM algorithm. Even then KHM also easily runs into local                K′ clusters into K clusters.
optima. PSO algorithm is a global optimization technique. A                  8. Find the centroids of K clusters and use the centroid as
hybrid data clustering algorithm based on the PSO and KHM                 initial centers for clustering the original dataset using K
(PSOKHM) was proposed by Yang et al. in [3]. This hybrid                  Means.
data clustering algorithm utilizes the advantages of both the                Although the mentioned initialization algorithms can help
algorithms. Therefore the PSOKHM algorithm not only helps                 finding good initial centers for some extent, they are quite
the KHM clustering run off from local optima but also                     complex and some use the K-Means algorithm as part of their
conquer the inadequacy of the slow convergence speed of the               algorithms, which still need to use the random method for
PSO algorithm. They conducted experiments to compare the                  cluster center initialization. The proposed approach for finding
hybrid data clustering algorithm with that of PSO and KHM                 initial cluster centroid is presented in the following section.
clustering on seven different data sets. The results of the
experiments show that PSOKHM was simply superior to the                                         III.     METHODOLOGY
other two clustering algorithms.
                                                                          3.1.      Initial Cluster Centers Deriving from Data
   Huang in [4] put forth a technique that enhances the
implementation of K-Means algorithm to various data sets.
                                                                            The algorithm follows a novel approach that performs data
Generally, the efficiency of K-Means algorithm in clustering
                                                                          partitioning along the data axis with the highest variance. The
the data sets is high. The restriction for implementing K-
                                                                          approach has been used successfully for color quantization [7].
Means algorithm to cluster real world data which contains
                                                                          The data partitioning tries to divide data space into small cells
categorical value is because of the fact that it was mostly
                                                                          or clusters where intercluster distances are large as possible
employed to numerical values. They presented two algorithms
                                                                          and intracluster distances are small as possible.
which extend the k-means algorithm to categorical domains
and domains with mixed numeric and categorical values. The
k-modes algorithm uses a trouble-free matching dissimilarity
measure to deal with categorical objects, replaces the means of
clusters with modes, and uses a frequency-based method to
modernize modes in the clustering process to decrease the
clustering cost function. The k-prototypes algorithm, from the
definition of a combined dissimilarity measure, further
integrates the k-means and k-modes algorithms to allow for
clustering objects described by mixed numeric and categorical
attributes. The experiments were conducted on well known
                                                                          Fig. 1 Diagram of ten data points in 2D, sorted by its X value, with an
soybean disease and credit approval data sets to demonstrate                              ordering number for each data point
the clustering performance of the two algorithms.
   Kluger [5] first proposed spectral biclustering for                       For instance, consider Fig. 1. Suppose ten data points in 2D
processing gene expression data. But Kluger’s focus is mainly             data space are given.
on unsupervised clustering, not on gene selection.                           The goal is to partition the ten data points in Fig. 1 into two
   There are some present works related to the finding                    disjoint cells where sum of the total clustering errors of the
initialization centroids.                                                 two cells is minimal, see Fig. 2. Suppose a cutting plane
   1. Compute mean (μj) and standard deviation (σ j) for every            perpendicular to X-axis will be used to partition the data. Let
jth attribute values.                                                     C1 and C2 be the first cell and the second cell respectively and
   2. Compute percentile Z1, Z2,…, Zk corresponding to area                  and     be the cell centroids of the first cell and the second
under the normal curve from – ∞ to (2s-1)/2k, s=1, 2, … ,k                cell, respectively. The total clustering error of the first cell is
(clusters).                                                               thus computed by:
   3. Compute attribute values xs =zs σj+μj corresponding to                                                                          (1)
these percentiles using mean and standard deviation of the                                              ,
   4. Perform the K-means to cluster data based on jth attribute            and the total clustering error of the second cell is thus
values using xs as initial centers and assign cluster labels to           computed by:
every data.                                                                                         ,                          (2)
   5. Repeat the steps of 3-4 for all attributes (l).
   6. For every data item t create the string of the class labels
Pt = (P1, P2,…, Pl) where Pj is the class label of t when using
the jth attribute values for step 4 clustering.

                                                                    338                                  http://sites.google.com/site/ijcsis/
                                                                                                         ISSN 1947-5500
                                                                   (IJCSIS) International Journal of Computer Science and Information Security,
                                                                   Vol. 8, No. 9, December 2010

   where ci is the ith data in a cell. As a result, the sum of total                                               ,                                 (5)
clustering errors of both cells are minimal (as shown in Fig.
2.)                                                                               The same argument is also true for the second cell. The total
                                                                                  clustering error of second cell can be minimized by reducing
                                                                                  the total discrepancies between all data in second cell to m,
                                                                                  which is computed by:
                                                                                                             ,                           (6)

                                                                                     where d(ci,cm) is the distance between m and each data in
                                                                                  each cell. Therefore the problem to minimize the sum of total
                                                                                  clustering errors of both cells can be transformed into the
                                                                                  problem to minimize the sum of total clustering error of all
                                                                                  data in the two cells to m.
                                                                                     The relationship between the total clustering error and the
  Fig. 2 Diagram of partitioning a cell of ten data points into two
                                                                                  clustering point may is illustrated in Fig. 4, where the
smaller cells, a solid line represents the intercluster distance and dash
                lines represent the intracluster distance
                                                                                  horizontal-axis represents the partitioning point that runs from
                                                                                  1 to n where n is the total number of data points and the
                                                                                  vertical-axis represents the total clustering error. When m=0,
                                                                                  the total clustering error of second cell equals to the total
                                                                                  clustering error of all data points while the total clustering
                                                                                  error of first cell is zero. On the other hand, when m=n, the
                                                                                  total clustering error of the first cell equals to the total
                                                                                  clustering error of all data points, while the total clustering
                                                                                  error of the second cell is zero.

 Fig. 3Illustration of partitioning the ten data points into two smaller
    cells using m as a partitioning point. A solid line in the square
represents the distance between the cell centroid and a data in cell, a
dash line represents the distance between m and data in each cell and
   a solid dash line represents the distance between m and the data               Fig. 4 Graphs depict the total clustering error, lines 1 and 2 represent
                          centroid in each cell                                   the total clustering error of the first cell and second cell, respectively,
                                                                                    Line 3 represents a summation of the total clustering errors of the
                                                                                                          first and the second cells
   The partition could be done using a cutting plane that passes
through m. Thus
                                                                                     A parabola curve shown in Fig. 4 represents a summation of
                                                                                  the total clustering error of the first cell and the second cell,
           ,             ,              ,        
                                                                                  represented by the dash line 2. Note that the lowest point of
                                                                                  the parabola curve is the optimal clustering point (m). At this
               ,             ,              ,                      (3)
                                                                                  point, the summation of total clustering error of the first cell
                                                                                  and the second cell are minimum.
          (as shown in Fig. 3). Thus
                                                                                     Since time complexity of locating the optimal point m is
                                                                                  O(n2), the distances between adjacent data is used along the X-
           ,                     ,                  ,       .| |                  axis to find the approximated point of n but with time of O(n).
                                                                                     Let             ,         be the squared Euclidean distance of
           ,                     ,                  ,       .| |    (4)           adjacent data points along the X-axis.
                                                                                     If i is in the first cell then      ,     ∑      . On the one
                                                                                  hand, if i is in the second cell then         ,      ∑        (as
   m is called as the partitioning data point where |C1| and |C2|                 shown in Fig. 5).
are the numbers of data points in cluster C1 and C2
respectively. The total clustering error of the first cell can be
minimized by reducing the total discrepancies between all data
in first cell to m, which is computed by:

                                                                            339                                   http://sites.google.com/site/ijcsis/
                                                                                                                  ISSN 1947-5500
                                                                (IJCSIS) International Journal of Computer Science and Information Security,
                                                                Vol. 8, No. 9, December 2010

                                                                                    3. Compute variance of each attribute of cell c. Choose an
                                                                                 attribute axis with the highest variance as the principal axis for
                                                                                    4. Compute squared Euclidean distances between adjacent
                                                                                 data along the data axis with the highest variance
                                                                                      ,       and compute the             ∑
                                                                                    5. Compute centroid distance of cell c:

                                                                                    Where dsumi is the summation of distances between the
    Fig. 5 Illustration of ten data points, a solid line represents the          adjacent data.
   distance between adjacent data along the X-axis and a dash line                  6. Divide cell c into two smaller cells. The partition
         represents the distance between m and any data point                    boundary is the plane perpendicular to the principal axis and
                                                                                 passes through a point m whose dsumi approximately equals
   The task of approximating the optimal point (m) in 2D is                      to centroidDist. The sorted linked lists of cell c are scanned
thus replaced by finding m in one-dimensional line as shown                      and divided into two for the two smaller cells accordingly
in Fig. 6.                                                                          7. Calculate Delta clustering error for c as the total
                                                                                 clustering error before partition minus total clustering error of
                                                                                 its two sub cells and insert the cell into an empty Max heap
Fig. 6 Illustration of the ten data points on a one-dimensional line and         with Delta clustering error as a key.
                              the relevant Dj                                       8. Delete a max cell from Max heap and assign it as a
                                                                                 current cell.
   The point (m) is therefore a centroid on the one dimensional                     9. For each of the two sub cells of c, which is not empty,
line (as shown in Fig. 6), which yields                                          perform step 3 - 7 on the sub cell.
                                                         (7)                        10. Repeat steps 8 - 9. Until the number of cells (Size of
                     ,                ,                                          heap) reaches K.
                                                                                    11. Use centroids of cells in max heap as the initial cluster
Let             ∑         and a centroidDist can be computed                     centers for K-means clustering
                               ∑                          (8)                       The above presented algorithms for finding the initialization
                                                                                 centroids do not provide a better result. Thus an efficient
                                                                                 method is proposed for obtaining the initial cluster centroids.
   It is probable to choose either the X-axis or Y-axis as the                   The proposed approach is well suited to cluster the gene
principal axis for data partitioning. However, data axis with                    dataset. So the proposed method is explained on the basis of
the highest variance will be chosen as the principal axis for                    genes.
data partitioning. The reason is to make the inter distance
between the centers of the two cells as large as possible while                  3.2.      Proposed Methodology
the sum of total clustering errors of the two cells are reduced                     The proposed method is Semi-Unsupervised Centroid
from that of the original cell. To partition the given data into k               Selection method. The proposed algorithm finds the initial
cells, it is started with a cell containing all given data and                   cluster centroids for the microarray gene dataset. The steps
partition the cell into two cells. Later on the next cell is                     involved in this procedure are as follows.
selected to be partitioned that yields the largest reduction of                     Spectral biclustering [10-12] can be carried out in the
total clustering errors (or Delta clustering error). This can be                 following three steps: data normalization, Bistochastization,
described as Total clustering error of the original cell – the                   and seeded region growing clustering. The raw data in many
sum of Total clustering errors of the two sub cells of the                       cancer gene-expression datasets can be arranged in one matrix.
original cell. This is done so that every time a partition on a                  In this matrix, denoted by, the rows and columns represent the
cell is performed, the partition will help reduce the sum of                     genes and the different conditions (e.g., different patients),
total clustering errors for all cells, as much as possible.                      respectively. Then the data normalization is performed as
   The partitioning algorithm can be used now to partition a                     follows. Take the logarithm of the expression data. Carry out
given set of data into k cells. The centers of the cells can then                five to ten cycles of subtracting either the mean or median of
be used as good initial cluster centers for the K-means                          the rows (genes) and columns (conditions) and then perform
algorithm. Following are the steps of the initial centroid                       five to ten cycles of row-column normalization. Since gene
predicting algorithm.                                                            expression microarray experiments can generate data sets with
   1. Let cell c contain the entire data set.                                    multiple missing values, the k-nearest neighbor (KNN)
   2. Sort all data in the cell c in ascending order on each                     algorithm is used to fill those missing values.
attribute value and links data by a linked list for each attribute.                 Define A       1/m ∑ A   to be the average of ith row,
                                                                                 A      1/n ∑ A   to be the average of th column, and

                                                                           340                               http://sites.google.com/site/ijcsis/
                                                                                                             ISSN 1947-5500
                                                         (IJCSIS) International Journal of Computer Science and Information Security,
                                                         Vol. 8, No. 9, December 2010

A..     1/mn ∑   ∑ A   to be the average of the whole                      compute the correlation (similarity) between each gene profile
matrix, where m is the number of genes and n the number of                 (e.g.,) and the eigenvectors .   ,       1,2 … … … , as
    Bistochastization may be done as follows. First, a matrix of                                                                               (9)
                                                                                  ,                     ,          1,2 … … .  ,
interactions is defined K      K by K       A     A A . j A..                           || || . || ||
Then the singular value decomposition (SVD) of the matrix K                                             1,2, … … .
is computed as given by U V T , where              is a diagonal
matrix of the same dimension as K and with nonnegative                        Where ||. || means vector 2—norms. Seen from (10), a
diagonal elements in decreasing order, U and V are m m                     large absolute of ,  indicates a strong correlation (similarity)
and n n orthonormal column matrices. The th column of the                  between ith gene and jth eigenvector. Therefore, genes can be
matrix V is denoted by v and v  Therefore, a scatter plot of               ranked as the absolute correlation values | , | for each
experimental conditions of the two best class partitioning                 eigenvector. For the eigenvector the top genes can be
eigenvectors v andv is obtained. The v   and v are often                   preselected, denoted by Gj, according to the corresponding
chosen as the eigenvectors corresponding to the largest and the            | , | value for    1,2, … … . . The value l can be empirically
second largest eigenvalues, respectively. The main reason is               determined. Thus, for each eigenvector of         , … . a set of
that they can capture most of the variance in the data and                 genes with largest values of the Cosine Measure is obtained
provide the optimal partition of different experimental                    which are taken as the initial cluster centroids in the proposed
conditions. In general, an s-dimensional scatter plot can be               clustering technique.
obtained by using eigenvectors v , v , … . v   (with largest
eigenvalues).                                                                              IV.     EXPERIMENTAL RESULTS
    Define P     v , v , … . v which has a dimension of n s.
The rows of matrix P stand for different conditions, which will              The proposed SCSM method is experimented using two
be clustered using SRG. Seeded region growing clustering is                microarray data sets: the lymphoma data set and the liver
carried out as follows. It begins with some seeds (initial state           cancer data set.
of the clusters). At each step of the algorithm, it is considered
                                                                                                         TABLE I
all as-yet unallocated samples, which border with at least one              GENE IDS (CLIDS) AND GENE NAMES IN THE TWO MICROARRAY DATA SETS
of the regions. Among them one sample, which has the                         Data set      Gene ID/              Gene Name               Gene Rank
minimum difference from its adjoining cluster, is allocated to                              CLID                                         G1    G2
its most similar adjoining cluster. With the result of clustering,
                                                                           Lymphoma         GENE              *CD63 antigen               3     /
the distinct types of cancer data can be predicted with very
                                                                                            1622X               (melanoma 1
high accuracy. In the next section, such clustering result is
used to select the best gene combinations or explained as the
best initial centroids.
                                                                                             GENE              *FGR tyrosine              /          3
                                                                                             2328X                 Kinase;
3.2.1. Semi-Unsupervised Centroid Selection (SCSM)
   The proposed semi-unsupervised centroid selection method
                                                                                             GENE             *mosaic protein             /          4
includes two steps: gene ranking and gene combination
                                                                                             3343X             LR11=hybrid;
                                                                                                              Receptor gp250
   As stated above, the best class partitioning eigenvectors is
obtained .Now these eigenvectors , , … .           are used to
rank and preselect genes.
                                                                              Liver         IMAGE:             116682 ECM1                7          /
   The proposed semi-unsupervised centroid selection method
                                                                              Cancer         301122         extracellular matrix
is based on the following two assumptions.
                                                                                                            protein 1 Hs.81071
   •      The genes which are most relevant to the cancer
should capture most variance in the data.
   •      Since , , … . may reveal the most variance in
                                                                              The lymphoma microarray data has three subtypes of
the data, the genes “similar” to , , … . should be relevant
                                                                           cancer, i.e., CLL, FL, and DLCL. The dataset is obtained from
to the cancer
                                                                           [8]. When applying the proposed method to this data set, the
                                                                           clustering result with two best partition eigenvectors is
  The gene ranking and preselecting process can be
                                                                           obtained. Seen from cluster results the three classes are
summarized as follows. After defining the ith gene
                                                                           correctly divided. Then two sets of l=20 genes are selected
profile      , ,…      , cosine measure is used to

                                                                     341                                    http://sites.google.com/site/ijcsis/
                                                                                                            ISSN 1947-5500
                                                        (IJCSIS) International Journal of Computer Science and Information Security,
                                                        Vol. 8, No. 9, December 2010

                                                              TABLE II
                                                COMPARISON OF GENERALIZATION ABILITY

                                                            Number of genes
                  Data set               Method                                                    Test Rate (%)           (p1,p2)
                                        k-means                  4026                                 100±0             (0, 0.9937)
                Lymphoma            Existing Method               81                                  100±0             (0, 0.9937)
                                         SCSM                    2±0                                99.92±0.37           (NA, NA)
                                        k-means                  1648                               98.10±0.11          (0, 0.9973)
               Liver Cancer         Existing Method               23                                98.10±0.11          (0, 0.9973)
                                         SCSM                    1±0                                98.70±0.08           (NA, NA)

   according to |Ri,1| and |Ri,2| respectively. (Here set have to
be two.) From the two sets of 20 genes each, the two-gene                    same gene selection result is obtained for each data set, but
combinations is chosen that can best divide the lymphoma                  slightly different classification accuracies. The p-values for
data. Two pairs of genes have been found: 1) Gene 1622X and               both numbers of genes and classification accuracies is
Gene 2328X, and 2) Gene 1622X and Gene 3343X, which                       calculated for both data sets in Table II, which showed that the
perfectly divide the lymphoma data. Since the results are                 differences between the numbers of genes used in our method
similar to each other, only the result of one group is shown.             and other methods are statistically significant, whereas the
Gene ID and gene names of the selecting genes in the                      differences between the classification accuracies between the
lymphoma data set are shown in Table I, where the group and               proposed method and other methods are not statistically
the rank of genes are also shown.                                         significant.
   The method is applied to the liver cancer data with two
classes, i.e., nontumor liver and HCC. The lung cancer data is                                90
                                                                                                                              DPDA (Existing)
obtained from [9]. The clustering result with the two best                                                                    SCSM (proposed)
partition eigenvectors is obtained. From the results it can be                                85
                                                                               Accuracy (%)

seen that there are three samples misclassified and the
clustering accuracy is 98.1%. Actually, it can set so that the                                80
scatter plot is on a single axis. Then top 20 genes are selected
with the largest. From the top 20 genes, it is found one gene
that can divide the liver cancer data well with accuracy of                                   70
98.7%. The result and gene name of selecting gene in liver
cancer data set are shown in Table I.                                                         65
                                                                                                       Lymphoma                Liver Cancer
4.1.      Comparison with results                                                                                   Dataset
   The paired t-test method is used to show the statistical
difference between our results and other published results. In            Figure 7: Comparison of classification accuracy among the proposed
general, given two paired sets and of measured values, the                          and existing technique for two different datasets.
paired t-test can be employed to compute a –value between
and determines whether they differ from each other in a                      The Figure 7 shows that the DPDA-K-Means Algorithm
statistically significant way under the assumptions that the              with Initial Cluster Centers Derived from Data Partitioning
paired differences are independent and identically normally               along the Data Axis with the Highest Variance method
distributed. The -value is defined as follows:                            produces result with less percentage of accuracy than the
                                          1                               proposed clustering with SCSM. The classification accuracy
                                                                          of the proposed method is very high than all the existing
                                                                          method. The result also shows that the proposed method is
                                                                          suitable only for the gene clustering and when the proposed
   Where,              ,              and ,  are the mean                 method used to cluster the other data it produces a less
values for and , respectively. Hence, all p∈[0,1], with a high -          percentage of accuracy.
value indicating statistically insignificant differences and a               The figure 8 shows the comparison of clustering time
low -value indicating statistically significant differences               among the DPDA-K-Means Algorithm with Initial Cluster
between and .                                                             Centers Derived from Data Partitioning along the Data Axis
   The order of cancer subtypes are shuffled and carried out              with the Highest Variance method and the proposed clustering
the experiments 20 times for each data set. Each time the                 with SCSM.

                                                                    342                                      http://sites.google.com/site/ijcsis/
                                                                                                             ISSN 1947-5500
                                                                      (IJCSIS) International Journal of Computer Science and Information Security,
                                                                      Vol. 8, No. 9, December 2010

                    130                                                                  [3]    F. Yang, T. Sun, and C. Zhang, “An efficient hybrid data clustering
                           DPDA (Existing)                                                      method based on K-harmonic means and Particle Swarm Optimization,”
                           SCSM (Proposed)                                                      An International Journal on Expert Systems with Applications, vol. 36,
                                                                                                no. 6, pp. 9847-9852, 2009.
      Time in sec

                                                                                         [4]    Zhexue Huang, “Extensions to the k-Means Algorithm for Clustering
                    110                                                                         Large Data Sets with Categorical Values,” Journal on Data Mining and
                                                                                                Knowledge Discovery, Springer, vol. 2, no. 3, pp. 283-304, 1998.
                    100                                                                  [5]    Y. Kluger, R. Basri, J. T. Chang, and M. Gerstein, “Spectral biclustering
                                                                                                of microarray cancer data: co-clustering genes and conditions,” Genome
                     90                                                                         Res., vol. 13, pp. 703–716, 2003.
                                                                                         [6]    P. Fränti and J. Kivijärvi, “Randomised Local Search Algorithm for the
                     80                                                                         Clustering Problem”. Pattern Analysis and Applications, Volume 3,
                                                                                                Issue 4, pages 358 – 369, 2000.
                            Lymphoma                 Liver Cancer                        [7]    M. Halkidi, Y. Batistakis and M. Vazirgiannis, “Cluster Validity
                                           Dataset                                              Methods: part I”. In Proceedings of the ACM SIGMOD International
                                                                                                Conference on Management of Data, Volume 31, Issue 2, pages 40 – 45,
Figure 8: Comparison of classification time among the proposed and                              June 2002.
                                                                                         [8]    Alizedeh AA, Eisen MB, Davis RE, Ma C, Lossos IS, Rosenwald A,
            existing technique for two different datasets.
                                                                                                Boldrick JC, Sabet H, Tran T, Yu X, et al.: Distinct types of diffuse
                                                                                                large B-cell lymphoma identified by gene expression profiling.
   From the graph it can be easily said that the proposed                                [9]    Hong, Z.Q. and Yang, J.Y. "Optimal Discriminant Plane for a Small
method takes slightly more time to cluster the gene data than                                   Number of Samples and Design Method of Classifier on the Plane",
                                                                                                Pattern Recognition, Vol. 24, No. 4, pp. 317-324, 1991.
the existing method. Even the clustering time taken is more,
                                                                                         [10]   Manjunath Aradhya, Francesco Masulli, and Stefano Rovetta
clustering accuracy is very high. Thus the proposed system                                      “Biclustering of Microarray Data based on Modular Singular Value
can be used for the gene clustering.                                                            Decomposition”, Proceedings of CIBB 2009.
                                                                                         [11]   LIU Wei, “A Parallel Algorithm for Gene Expressing Data
                                                                                                Biclustering”, journal of computers, vol. 3, no. 10, october 2008
                                V.      CONCLUSION
                                                                                         [12]   Kenneth Bryan, P´adraig Cunningham and Nadia Bolshakova,
   The most commonly used efficient clustering technique is                                     “Biclustering of Expression Data Using Simulated Annealing”, This
                                                                                                research was sponsored by Science oundation Ireland under grant
k-means clustering. Initial starting points those computed                                      number SFI-02/IN1/I111.
randomly by K-means often make the clustering results                                    [13]   A. K. Jain, M. N. Murty and P. J. Flynn, “Data Clustering: A Review”,
reaching the local optima. So to overcome this disadvantage a                                   ACM Computing Surveys, Vol. 31, No. 3, September 1999
new technique is proposed. Semi-Unsupervised Centroid                                    [14]   Shai Ben-David, David Pal, and Hans Ulrich Simon, “Stability of k-
                                                                                                Means Clustering”.
Selection method is used with the present clustering approach                            [15]   Madhu Yedla, Srinivasa Rao Pathakota and T. M. Srinivasa, “Enhancing
in the proposed system to compute the initial centroids for the                                 K- means Clustering Algorithm with Improved Initial Center”, Vol. 1,
k-means algorithm. The experiments for this proposed                                            121-125, 2010.
approach is conducted on the microarray gene database. The                               [16]   A. M. Fahim, A. M. Salem, F. A. Torkey and M. A. Ramadan, “An
                                                                                                Efficient enhanced k-means clustering algorithm,” journal of Zhejiang
data sets used are lymphoma and the liver cancer data set. The                                  University, 10(7): 16261633, 2006.
accuracy of the proposed approach is compared with the
existing technique called the DPDA. The results are obtained
and tabulated. It is clearly observed from the results that, the
proposed approach shows significant performance. In the
lymphoma data set, the accuracy of the proposed approach is
about 87%. The accuracy of the DPDA approach is very less
(i.e.) 75%. Similarly for the liver cancer data set, the accuracy
of the proposed approach is about 81% which is also higher
than the existing approach. Moreover, time taken for
classification of the proposed approach is more or less similar
to the DPDA approach. The time taken for classification by
the proposed approach in lymphoma and liver cancer data sets
are 115 and 130 seconds respectively which is almost similar
to the existing approach. Thus the proposed approach provides
the best classification accuracy within a short time interval.


[1]        Guangsheng Feng, Huiqiang Wang, Qian Zhao, and Ying Liang, “A
           Novel Clustering Algorithm for Prefix-Coded Data Stream Based upon
           Median-Tree,” IEEE, International Conference on Internet Computing in
           Science and Engineering, ICICSE '08, pp. 79-84, 2008.
[2]        P. S. Bradley, and U. M. Fayyad, “Refining Initial Points for K-Means
           Clustering,” ACM, Proceedings of the 15th International Conference on
           Machine Learning, pp. 91-99, 1998.

                                                                                   343                                      http://sites.google.com/site/ijcsis/
                                                                                                                            ISSN 1947-5500

To top