kmed: Distance-Based K-Medoids

Weksi Budiaji

2019-06-14

Abstract

The kmed vignette consists of four sequantial parts of distance-based (k-medoids) cluster analysis. The first part is defining the distance. It has numerical, binary, categorical, and mixed distances. The next part is applying a clustering algorithm in the pre-defined distance. There are four k-medoids presented, namely the simple and fast k-medoids, k-medoids, ranked k-medoids, and increasing number of clusters in k-medoids. After the clustering result is obtained, a validation step is required. The cluster validation applies internal and relative criteria. The last part is visualizing the cluster result in a biplot or marked barplot.

1. Introduction

The kmed package is designed to analyse k-medoids based clustering. The features include:

2. Distance Computation

2.A. Numerical variables (distNumeric)

The distNumeric function can be applied to calculate numerical distances. There are four distance options, namely Manhattan weighted by range (mrw), squared Euclidean weighted by range (ser), squared Euclidean weighted by squared range (ser.2), squared Euclidean weighted by variance (sev), and unweighted squared Euclidean (se). The distNumeric function provides method in which the desired distance method can be selected. The default method is mrw.

The distance computation in a numerical variable data set is performed in the iris data set. An example of manual calculation of the numerical distances is applied for the first and second objects only to introduce what the distNumeric function does.

##   Sepal.Length Sepal.Width Petal.Length Petal.Width Species
## 1          5.1         3.5          1.4         0.2  setosa
## 2          4.9         3.0          1.4         0.2  setosa
## 3          4.7         3.2          1.3         0.2  setosa

2.A.1. Manhattan weighted by range (method = "mrw")

By applying the distNumeric function with method = "mrw", the distance among objects in the iris data set can be obtained.

##           1         2         3
## 1 0.0000000 0.2638889 0.2530603
## 2 0.2638889 0.0000000 0.1558380
## 3 0.2530603 0.1558380 0.0000000

The Manhattan weighted by range distance between objects 1 and 2 is 0.2638889. To calculate this distance, the range of each variable is computed.

## Sepal.Length  Sepal.Width Petal.Length  Petal.Width 
##          3.6          2.4          5.9          2.4

Then, the distance between objects 1 and 2 is

## [1] 0.2638889

which is based on the data

##   Sepal.Length Sepal.Width Petal.Length Petal.Width
## 1          5.1         3.5          1.4         0.2
## 2          4.9         3.0          1.4         0.2

(Back to Intoduction)

2.A.2. squared Euclidean weighted by range (method = "ser")

##            1          2          3
## 1 0.00000000 0.11527778 0.08363936
## 2 0.11527778 0.00000000 0.02947269
## 3 0.08363936 0.02947269 0.00000000

The squared Euclidean weighted by range distance between objects 1 and 2 is 0.11527778. It is obtained by

## [1] 0.1152778

(Back to Intoduction)

2.A.3. squared Euclidean weighted by squared range (method = "ser.2")

##            1          2          3
## 1 0.00000000 0.04648920 0.02825795
## 2 0.04648920 0.00000000 0.01031814
## 3 0.02825795 0.01031814 0.00000000

The squared Euclidean weighted by squared range distance between objects 1 and 2 is 0.04648920 that is computed by

## [1] 0.0464892

where the data are

##   Sepal.Length Sepal.Width Petal.Length Petal.Width
## 1          5.1         3.5          1.4         0.2
## 2          4.9         3.0          1.4         0.2

(Back to Intoduction)

2.A.4. squared Euclidean weighted by variance (method = "sev")

##           1         2         3
## 1 0.0000000 1.3742671 0.7102849
## 2 1.3742671 0.0000000 0.2720932
## 3 0.7102849 0.2720932 0.0000000

The squared Euclidean weighted by variance distance between objects 1 and 2 is 1.3742671. To compute this distance, the variance of each variable is calculated.

## Sepal.Length  Sepal.Width Petal.Length  Petal.Width 
##    0.6856935    0.1899794    3.1162779    0.5810063

Then, the distance between objects 1 and 2 is

## [1] 1.374267

(Back to Intoduction)

2.A.5. squared Euclidean (method = "se")

##      1    2    3
## 1 0.00 0.29 0.26
## 2 0.29 0.00 0.09
## 3 0.26 0.09 0.00

The squared Euclidean distance between objects 1 and 2 is 0.29. It is computed by

## [1] 0.29

(Back to Intoduction)

2.B. Binary or Categorical variables

There are two functions to calculate the binary and categorical variables. The first is matching to compute the simple matching distance and the second is cooccur to calculate the co-occurrence distance. To introduce what these functions do, the bin data set is generated.

2.B.1. Simple matching (matching)

The matching function calculates the simple matching distance between two data sets. If the two data sets are identical, the functions calculates the distance among objects within the data set. The simple matching distance is equal to the proportion of the mis-match categories.

##   x y
## 1 1 1
## 2 1 2
## 3 2 2
## 4 2 2
##     1   2   3   4
## 1 0.0 0.5 1.0 1.0
## 2 0.5 0.0 0.5 0.5
## 3 1.0 0.5 0.0 0.0
## 4 1.0 0.5 0.0 0.0

As an example of the simple matching distance, the distance between objects 1 and 2 is calculated by

## [1] 0.5

The distance between objects 1 and 2, which is 0.5, is produced from one mis-match and one match categories from the two variables (x and y) in the bin data set. When x1 is equal to x2, for instance, the score is 0. Meanwile, if x1 is not equal to x2, the score is 1. These scores are also valid in the y variable. Hence, the distance between objects 1 and 2 is (0+1)/2 that is equal to 1/2.

(Back to Intoduction)

2.B.2. Co-occurrence distance (cooccur)

The co-ocurrence distance (Ahmad and Dey 2007; Harikumar and PV 2015) can be calculated via the cooccur function. To calculate the distance between objects, the distribution of the variables are taken into consideration. Compared to the simple matching distance, the co-occurrence distance redefines the score of match and mis-match categories such that they are unnecessary to be 0 and 1, respectively. Due to relying on the distribution of all inclusion variables, the co-occurence distance of a data set with a single variable is absent.

The co-occurrence distance of the bin data set is

##           1         2        3        4
## 1 0.0000000 0.6666667 1.166667 1.166667
## 2 0.6666667 0.0000000 0.500000 0.500000
## 3 1.1666667 0.5000000 0.000000 0.000000
## 4 1.1666667 0.5000000 0.000000 0.000000

To show how co-occurrence distance is calculated, the distance between objects 1 and 2 is presented.

##   x y
## 1 1 1
## 2 1 2
## 3 2 2
## 4 2 2

Step 1 Creating cross tabulations

##    
##     1 2
##   1 1 1
##   2 0 2
##    
##     1 2
##   1 1 0
##   2 1 2

Step 2 Calculating the column proportions of each cross tabulation

##    
##     1         2
##   1 1 0.3333333
##   2 0 0.6666667
##    
##       1 2
##   1 0.5 0
##   2 0.5 1

Step 3 Finding the maximum values for each row of the proportion

##         1         2 
## 1.0000000 0.6666667
##   1   2 
## 0.5 1.0

Step 4 Defining the scores of each variable

The score is obtained by a summation of the maximum value subtracted and divided by a constant. The constant has a value depending on the number of inclusion variables. For the bin data set, the constant is 1 because both x and y variables are only depended on one other variable, i.e. x depends on the distribution of y and y relies on the distribution of x.

## [1] 0.6666667
## [1] 0.5

It can be implied that the score for mis-match categories are 0.5 and 0.67 in the x and y variables, respectively. Note that the score for match categories is alwalys 0. Thus, the distance between objects 1 and 2 is 0+0.6666667 = 0.6666667 and between objects 1 and 3 is 0.5+0.6666667 = 1.1666667

(Back to Intoduction)

2.C. Mixed variables (distmix)

There are six available distance methods for a mixed variable data set. The distmix function calculates mixed variable distance in which it requires column id of each class of variables. The mixdata data set is generated to describe each method in the distmix function.

2.C.1 Gower (method = "gower")

The method = "gower" in the distmix function calculates the Gower (1971) distance. The original Gower distance allows missing values, while it is not allowed in the distmix function.

##   num1 num2 bin1 bin2 cat1 cat2
## 1  1.4  0.2    1    1    1    3
## 2  1.4  0.2    1    2    3    1
## 3  4.7  1.4    2    2    2    2
## 4  4.5  1.5    2    2    1    2

The Gower distance of the mixdata data set is

##           1         2         3         4
## 1 0.0000000 0.5000000 0.9871795 0.8232323
## 2 0.5000000 0.0000000 0.8205128 0.8232323
## 3 0.9871795 0.8205128 0.0000000 0.1895882
## 4 0.8232323 0.8232323 0.1895882 0.0000000

As an example, the distance between objects 3 and 4 is presented. The range of each numerical variables is necessary.

## num1 num2 
##  3.3  1.3

The Gower distance calculates the Gower similarity first. In the Gower similarity, the mis-match categories in the binary/ categorical variables are scored 0 and the match categories are 1. Meanwhile, in the numerical variables, 1 is subtracted by a ratio between the absolute difference and its range. Then, the Gower similarity can be weighted by the number of variables. Thus, the Gower similarity between objects 3 and 4 is

## [1] 0.8104118

The Gower distance is obtained by subtracting 1 with the Gower similarity. The distance between objects 3 and 4 is then

## [1] 0.1895882

(Back to Intoduction)

2.C.2 Wishart (method = "wishart")

The Wishart (2003) distance can be calculated via method = "wishart". Although it allows missing values, it is again illegitimate in the distmix function. The Wishart distance for the mixdata is

##           1         2         3         4
## 1 0.0000000 0.7071068 1.2871280 1.2277616
## 2 0.7071068 0.0000000 1.2206686 1.2277616
## 3 1.2871280 1.2206686 0.0000000 0.4144946
## 4 1.2277616 1.2277616 0.4144946 0.0000000

To calculate the Wishart distance, the variance of each numerical variable is required. It weighs the squared difference of a numerical variable.

##   num1   num2 
## 3.4200 0.5225

Meanwhile, the mis-match categories in the binary/ categorical variables are scored 1 and the match categories are 0. Then, all score of the variables is added and squared rooted. Thus, the distance between objects 3 and 4 is

## [1] 0.4144946

(Back to Intoduction)

2.C.3 Podani (method = "podani")

The method = "podani" in the distmix function calculates the Podani (1999) distance. Similar to The Gower and Wishart distances, it allows missing values, yet it is not allowed in the distmix function. The Podani distance for the mixdata is

##          1        2        3        4
## 1 0.000000 1.732051 2.419105 2.209629
## 2 1.732051 0.000000 2.202742 2.209629
## 3 2.419105 2.202742 0.000000 1.004784
## 4 2.209629 2.209629 1.004784 0.000000

The Podani and Wishart distances are similar. They are different in the denumerator for the numerical variables. Instead of a variance, the Podani distance applies the squared range for a numerical variable. Unlike the Gower and Podani distances, the number of variables as a weight is absent in the Podani distance. Hence, the distance between objects 3 and 4 is

## [1] 1.004784

which is based on data

##   num1 num2 bin1 bin2 cat1 cat2
## 3  4.7  1.4    2    2    2    2
## 4  4.5  1.5    2    2    1    2

(Back to Intoduction)

2.C.4 Huang (method = "huang")

The method = "huang" in the distmix function calculates the Huang (1997) distance. The Huang distance of the mixdata data set is

##           1         2         3         4
## 1  0.000000  3.858249 17.474332 15.158249
## 2  3.858249  0.000000 16.188249 15.158249
## 3 17.474332 16.188249  0.000000  1.336083
## 4 15.158249 15.158249  1.336083  0.000000

The average standard deviation of the numerical variables is required to calculate the Huang distance. This measure weighs the binary/ categorical variables.

## [1] 1.286083

While the squared difference of the numerical variables is calculated, the mis-match categories are scored 1 and the match categories are 0 in the binary/ categorical variables. Thus, the distance between objects 3 and 4 is

## [1] 1.336083

(Back to Intoduction)

2.C.5 Harikumar and PV (method = "harikumar")

The Harikumar and PV (2015) distance can be calculated via method = "harikumar". The Harikumar and PV distance for the mixdata is

##     1   2   3   4
## 1 0.0 3.0 7.5 6.9
## 2 3.0 0.0 7.5 7.4
## 3 7.5 7.5 0.0 0.8
## 4 6.9 7.4 0.8 0.0

The Harikumar and PV distance requires an absolute difference in the numerical variables and unweighted simple matching, i.e. Hamming distance, in the binary variables. For the categorical variables, it applies co-occurrence distance. The co-occurence distance in the categorical variables is (for manual calculation see co-occurrence subsection)

##     1 2   3   4
## 1 0.0 2 1.0 0.5
## 2 2.0 0 2.0 2.0
## 3 1.0 2 0.0 0.5
## 4 0.5 2 0.5 0.0

Hence, the distance between objects 1 and 3 is

## [1] 0.8

where the data are

##   num1 num2 bin1 bin2 cat1 cat2
## 3  4.7  1.4    2    2    2    2
## 4  4.5  1.5    2    2    1    2

(Back to Intoduction)

2.C.6 Ahmad and Dey (method = "ahmad")

The method = "ahmad" in the distmix function calculates the Ahmad and Dey (2007) distance. The Ahmad and Dey distance of the mixdata data set is

##          1        2        3        4
## 1  0.00000  4.45679 20.04605 16.48827
## 2  4.45679  0.00000 16.33000 15.30000
## 3 20.04605 16.33000  0.00000  0.30000
## 4 16.48827 15.30000  0.30000  0.00000

The Ahmad and dey distance requires a squared difference in the numerical variables and co-occurrence distance for both the binary and categorical variables. The co-occurrence distance in the mixdata data set is

##          1        2        3        4
## 1 0.000000 2.111111 2.777778 2.277778
## 2 2.111111 0.000000 2.000000 2.000000
## 3 2.777778 2.000000 0.000000 0.500000
## 4 2.277778 2.000000 0.500000 0.000000

Thus, the distance between objects 2 and 3 is

## [1] 16.33

which is based on the data

##   num1 num2 bin1 bin2 cat1 cat2
## 2  1.4  0.2    1    2    3    1
## 3  4.7  1.4    2    2    2    2

(Back to Intoduction)

3. K-medoids algorithms

There are some k-medoids algorithms available in this package. They are the simple and fast k-medoids (fastkmed), k-medoids, ranked k-medoids (rankkmed), and increasing number of clusters k-medoids (inckmed). All algorithms have a list of results, namely the cluster membership, id medoids, and distance of all objects to their medoid.

In this section, the algorithms are applied in the iris data set by applying the mrw distance (see Manhattan weighted by range). The number of clusters in this data set is 3.

3.A. Simple and fast k-medoids algorithm (fastkmed)

The simple and fast k-medoid (SFKM) algorithm has been proposed by Park and Jun (2009). The fastkmed function runs this algorithm to cluster the objects. The compulsory inputs are a distance matrix or distance object and a number of clusters. Hence, the SFKM algorithm for the iris data set is

## $cluster
##   1   2   3   4   5   6   7   8   9  10  11  12  13  14  15  16  17  18 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  19  20  21  22  23  24  25  26  27  28  29  30  31  32  33  34  35  36 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  37  38  39  40  41  42  43  44  45  46  47  48  49  50  51  52  53  54 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   3   3   3   2 
##  55  56  57  58  59  60  61  62  63  64  65  66  67  68  69  70  71  72 
##   3   2   3   2   2   2   2   2   2   2   2   3   2   2   2   2   3   2 
##  73  74  75  76  77  78  79  80  81  82  83  84  85  86  87  88  89  90 
##   2   2   2   3   3   3   2   2   2   2   2   2   2   2   3   2   2   2 
##  91  92  93  94  95  96  97  98  99 100 101 102 103 104 105 106 107 108 
##   2   2   2   2   2   2   2   2   2   2   3   3   3   3   3   3   2   3 
## 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 
##   3   3   3   3   3   3   3   3   3   3   3   2   3   3   3   3   3   3 
## 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 
##   3   3   3   3   3   3   3   3   2   3   3   3   3   3   3   3   3   3 
## 145 146 147 148 149 150 
##   3   3   3   3   3   3 
## 
## $medoid
## [1]   8  95 148
## 
## $minimum_distance
## [1] 48.76718

Then, a classification table can be obtained.

##    
##     setosa versicolor virginica
##   1     50          0         0
##   2      0         39         3
##   3      0         11        47

Applying the SFKM algorithm in iris data set with the Manhattan weighted by range, the misclassification rate is

## [1] 0.09333333

(Back to Intoduction)

3.B. K-medoids algorithm

Reynolds et al. (2006) has been proposed a k-medoids (KM) algorithm. It is similar to the SFKM such that the fastkmed can be applied. The difference is in the initial medoid selection where the KM selects the initial medoid randomly. Thus, the KM algorithm for the iris data set by setting the init is

## [1] 40 56 85
## $cluster
##   1   2   3   4   5   6   7   8   9  10  11  12  13  14  15  16  17  18 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  19  20  21  22  23  24  25  26  27  28  29  30  31  32  33  34  35  36 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  37  38  39  40  41  42  43  44  45  46  47  48  49  50  51  52  53  54 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   3   3   3   2 
##  55  56  57  58  59  60  61  62  63  64  65  66  67  68  69  70  71  72 
##   2   2   3   2   2   2   2   2   2   2   2   3   2   2   2   2   3   2 
##  73  74  75  76  77  78  79  80  81  82  83  84  85  86  87  88  89  90 
##   2   2   2   3   2   3   2   2   2   2   2   2   2   2   3   2   2   2 
##  91  92  93  94  95  96  97  98  99 100 101 102 103 104 105 106 107 108 
##   2   2   2   2   2   2   2   2   2   2   3   3   3   3   3   3   2   3 
## 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 
##   3   3   3   3   3   3   3   3   3   3   3   2   3   3   3   3   3   3 
## 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 
##   3   3   3   3   3   3   3   3   2   3   3   3   3   3   3   3   3   3 
## 145 146 147 148 149 150 
##   3   3   3   3   3   3 
## 
## $medoid
## [1]   8 100 148
## 
## $minimum_distance
## [1] 48.8411

The classification table of the KM algorithm is

##    
##     setosa versicolor virginica
##   1     50          0         0
##   2      0         41         3
##   3      0          9        47

with the misclassification rate

## [1] 0.08

Compared to the SFKM algorithm, which has 9.33% misclassification, the misclassification of the KM algorithm is slightly better (8%).

(Back to Intoduction)

3.C. Rank k-medoids algorithm (rankkmed)

A rank k-medoids (RKM) has been proposed by Zadegan, Mirzaie, and Sadoughi (2013). The rankkmed function runs the RKM algorithm. The m argument is introduced to calculate a hostility score. The m indicates how many closest objects is selected. The selected objects as initial medoids in the RKM is randomly assigned. The RKM algorithm for the iris data set by setting m = 10 is then

## $cluster
##   1   2   3   4   5   6   7   8   9  10  11  12  13  14  15  16  17  18 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  19  20  21  22  23  24  25  26  27  28  29  30  31  32  33  34  35  36 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1   1 
##  37  38  39  40  41  42  43  44  45  46  47  48  49  50  51  52  53  54 
##   1   1   1   1   1   1   1   1   1   1   1   1   1   1   2   2   2   2 
##  55  56  57  58  59  60  61  62  63  64  65  66  67  68  69  70  71  72 
##   2   2   2   2   2   2   2   2   2   2   2   2   2   2   2   2   3   2 
##  73  74  75  76  77  78  79  80  81  82  83  84  85  86  87  88  89  90 
##   2   2   2   2   2   3   2   2   2   2   2   3   2   2   2   2   2   2 
##  91  92  93  94  95  96  97  98  99 100 101 102 103 104 105 106 107 108 
##   2   2   2   2   2   2   2   2   2   2   3   3   3   3   3   3   3   3 
## 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 
##   3   3   3   3   3   3   3   3   3   3   3   2   3   3   3   3   3   3 
## 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 
##   3   3   3   3   3   3   3   2   2   3   3   3   3   3   3   3   3   3 
## 145 146 147 148 149 150 
##   3   3   3   3   3   3 
## 
## $medoid
## [1] "50"  "92"  "128"
## 
## $minimum_distance
## [1] 56.71822

Then, a classification table is attained by

##    
##     setosa versicolor virginica
##   1     50          0         0
##   2      0         47         3
##   3      0          3        47

The misclassification proportion is

## [1] 0.04

With 4% misclassification rate, the RKM algorithm is the best among the three previous algorithms.

(Back to Intoduction)

3.D. Increasing number of clusters k-medoids algorithm (inckmed)

Yu et al. (2018) has been proposed an increasing number of clusters k-medoids (INCKM) algorithm. This algorithm is implemented in the inckmed function. The alpha argument indicates a stretch factor to select the initial medoids. The SFKM, KM and INCKM are similar algorithm with a different way to select the initial medoids. The INCKM algorithm of the iris data set with alpha = 1.1 is

## $cluster
##   1   2   3   4   5   6   7   8   9  10  11  12  13  14  15  16  17  18 
##   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3 
##  19  20  21  22  23  24  25  26  27  28  29  30  31  32  33  34  35  36 
##   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3   3 
##  37  38  39  40  41  42  43  44  45  46  47  48  49  50  51  52  53  54 
##   3   3   3   3   3   3   3   3   3   3   3   3   3   3   2   2   2   1 
##  55  56  57  58  59  60  61  62  63  64  65  66  67  68  69  70  71  72 
##   1   1   2   1   1   1   1   1   1   1   1   2   1   1   1   1   2   1 
##  73  74  75  76  77  78  79  80  81  82  83  84  85  86  87  88  89  90 
##   1   1   1   2   1   2   1   1   1   1   1   1   1   1   2   1   1   1 
##  91  92  93  94  95  96  97  98  99 100 101 102 103 104 105 106 107 108 
##   1   1   1   1   1   1   1   1   1   1   2   2   2   2   2   2   1   2 
## 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 
##   2   2   2   2   2   2   2   2   2   2   2   1   2   2   2   2   2   2 
## 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 
##   2   2   2   2   2   2   2   2   1   2   2   2   2   2   2   2   2   2 
## 145 146 147 148 149 150 
##   2   2   2   2   2   2 
## 
## $medoid
## [1] 100 148   8
## 
## $minimum_distance
## [1] 48.8411

Then, the classification table can be attained.

##    
##     setosa versicolor virginica
##   1      0         41         3
##   2      0          9        47
##   3     50          0         0

The misclassification rate is

## [1] 0.08

The algorithm has 8% misclassification rate such that the RKM algorithm performs the best among the four algorithms in the iris data set with the mrw distance.

(Back to Intoduction)

4. Cluster validation

The clustering algorithm result has to be validated. There are two types of validation implemented in the kmed package. They are internal and relative criteria validations.

4.A. Internal criteria

4.A.1. Silhouette (sil)

Rousseeuw (1987) has proposed a silhouette index as an internal measure of validation. It is based on the average distance of objects within a cluster and between the nearest cluster. The sil function calculates the silhouette index of clustering result. The arguments are a distance matrix or distance object, id medoids, and cluster membership. It produce a list of silhouette indices and sihouette plots.

The silhouette index and plot of the best clustering result of iris data set via RKM is presented.

The silhouette index of each object can be obtained by

##    silhouette cluster
## 49 0.78952089       1
## 50 0.82084673       1
## 51 0.07607567       2
## 52 0.22234719       2

Then, the plot is presented by

(Back to Intoduction)

4.A.2. Centroid-based shadow value (csv)

An other way to measure internal validation with its corresponding plot is by presenting the centroid-based shadow value (Leisch 2010). The csv function calculates and plots the shadow value of each object, which is based on the first and second closest medoids. The centroid of the original version of the csv is replaced by medoids in the csv function to adapt the k-medoids algorithm.

The required arguments in the csv function is identical to the silhouette (sil) function. Thus, the shadow value and plot of the best clustering result of iris data set via RKM can be obtained by

The shadow values of objects 49 to 52, for instance, are presented by

##      shadval cluster
## 49 0.2955819       1
## 50 0.0000000       1
## 51 0.7923323       2
## 52 0.7705314       2

The shadow value plot is also produced.

(Back to Intoduction)

4.B. Relative criteria

The relative criteria evaluate a clustering algorithm result by applying re-sampling strategy. Thus, a bootstrap strategy can be applied. It is expected that the result of the cluster bootstraping is robust over all replications (Dolnicar and Leisch 2010). There are three steps to validate the cluster result via the boostraping strategy.

Step 1 Creating a matrix of bootstrap replicates

To create a matrix of bootstrap replicates, the clustboot function can be applied. There are five arguments in the clustboot function with the algorithm argument being the most important. The algorithm argument is the argument for a clustering algorithm that a user wants to evaluate. It has to be a function. When the RKM of iris data set is validated, for instance, the RKM function, which is required as an input in the algorithm argument, is

When a function is created, it has to have two input arguments. They are x (a distance matrix) and nclust (a number of clusters). The output, on the other hand, is a vector of cluster membership (res$cluster). Thus, the matrix of bootstrap replicates can be produced by

with the objects 1 to 4 on the first to fifth replications being

##      [,1] [,2] [,3] [,4] [,5]
## [1,]    0    0    0    1    1
## [2,]    2    0    1    1    1
## [3,]    2    0    0    0    0
## [4,]    1    0    0    1    0

The rkmbootstrap is a matrix of bootrstrap replicates with a dimension of 150 x 50, i.e. n x b, where n is the number of objects and b is the number of bootstrap replicates. Note that the default evaluated algorithm is the SFKM algorithm such that if a user ignores the algorithm argument, the matrix of bootstrap replicates can still be produced. However, it misleads because it does not evaluate the user’s own algorithm.

Step 2 Transforming the bootstrap matrix into a consensus matrix

The matrix of bootstrap replicates produced by the clustboot in the step 1 can be transformed into a consensus matrix with a dimension of n x n via the consensusmatrix function. An element of the consensus matrix in row i dan column j is an agreement value between objects i and j to be in the same cluster when they are taken as a sample at the same time (Monti et al. 2003).

However, it requires an algorithm to order the objects in such a way that objects in the same cluster are close to each other. The consensusmatrix function has the reorder argument to comply this task. It is similar to the algorithm argument in the clustboot function in the step 1 where the reorder has to be a function that has two arguments and a vector of output.

Transforming the rkmbootstrap into a consensus matrix via the ward linkage algorithm to oder the objects, for example, can obtained by

The first to fourth rows and columns can be displayed as

##           1         1         1         1
## 1 1.0000000 0.9583333 0.9130435 0.8421053
## 1 0.9583333 1.0000000 1.0000000 0.9565217
## 1 0.9130435 1.0000000 1.0000000 0.9500000
## 1 0.8421053 0.9565217 0.9500000 1.0000000

Step 3 Visualizing the consensus matrix in a heatmap

The ordered consensus matrix in the step 2 can be visualized in a heatmap applying the clustheatmap function. The agreement indices in the consensus matrix can be transformed via a non-linear transformation (Hahsler and Hornik 2011). Thus, the consensusrkm can visualize into

(Back to Intoduction)

5. Cluster visualization

A cluster visualization of the clustering result can enhance the data structure understanding. The biplot and marked barplot are presented to visualize the clustering result.

A. Biplot

The pcabiplot function can be applied to plot a clustering result from a numerical data set. The numerical data set has to be converted into a principle component object via the prcomp function. The x and y axes in the plot can be replaced by any component of the principle components. The colour of the objects can be adjusted based on the cluster membership by supplying a vector of membership in the colobj argument.

The iris data set can be plotted in a pca biplot with the colour objects based on the RKM algorithm result.

The second principle component can be replaced by the third principle component.

(Back to Intoduction)

B. Marked barplot

A marked barplot has been proposed by Dolnicar and Leisch (2014); Leisch (2008) where the mark indicates a significant difference between the cluster’s mean and population’s mean in each variable. The barplot function creates a barplot of each cluster with a particular significant level. The layout of the barplot is set in the nc argument.

The barplot of iris data set partitioned by the RKM algorithm is

while the layout is changed into 2 columns and the alpha is set into 1%, it becomes

(Back to Intoduction)


References

Ahmad, A., and L. Dey. 2007. “A K-Mean Clustering Algorithm for Mixed Numeric and Categorical Data.” Data and Knowledge Engineering 63 (November). Elsevier:503–27. https://doi.org/10.1016/j.datak.2007.03.016.

Dolnicar, S., and F. Leisch. 2010. “Evaluation of Structure and Reproducibility of Cluster Solutions Using the Bootstrap.” Marketing Letters 21 (March). Springer:83–101. https://doi.org/10.1007/s11002-009-9083-4.

———. 2014. “Using Graphical Statistics to Better Understand Market Segmentation Solutions.” International Journal of Market Research 56 (March). Sage:207–30. https://doi.org/10.2501/IJMR-2013-073.

Gower, J. 1971. “A General Coefficient of Similarity and Some of Its Properties.” Biometrics 27 (December). International Biometric Society:857–71. https://doi.org/10.2307/2528823.

Hahsler, M., and K. Hornik. 2011. “Consensus Clustering; Dissimilarity Plots; A Visual Exploration Tool for Partitional Clustering.” Journal of Computational and Graphical Statistics 20 (August):335–54. https://doi.org/10.1198/jcgs.2010.09139.

Harikumar, S., and S. PV. 2015. “K-Medoid Clustering for Heterogeneous Data Sets.” JProcedia Computer Science 70. Elsevier:226–37. https://doi.org/10.1016/j.procs.2015.10.077.

Huang, Z. 1997. “Clustering Large Data Sets with Mixed Numeric and Categorical Values.” In The First Pacific-Asia Conference on Knowledge Discovery and Data Mining, 21–34.

Leisch, F. 2008. “Visualizing Cluster Analysis and Finite Mixture Models.” In Handbook of Data Visualization, 561–87. Springer Verlag. https://doi.org/10.1007/978-3-540-33037-0_22.

———. 2010. “Neighborhood Graphs, Stripes and Shadow Plots for Cluster Visualization.” Statistics and Computing 20 (October). Springer:457–69. https://doi.org/10.1007/s11222-009-9137-8.

Monti, S., P. Tamayo, J. Mesirov, and T. Golub. 2003. “Consensus Clustering; A Resampling-Based Method for Class Discovery and Visualization of Gene Expression Microarray Data.” Machine Learning 52 (July). Springer:91–118. https://doi.org/10.1023/A:1023949509487.

Park, H., and C. Jun. 2009. “A Simple and Fast Algorithm for K-Medoids Clustering.” Expert Systems with Applications 36 (2). Elsevier:3336–41. https://doi.org/10.1016/j.eswa.2008.01.039.

Podani, J. 1999. “Extending Gower’s General Coefficient of Similarity to Ordinal Characters.” Taxon 48 (May). Wiley:331–40. https://doi.org/10.2307/1224438.

Reynolds, A. P., G. Richards, B. De La Iglesia, and V. J. Rayward-Smith. 2006. “Clustering Rules; A Comparison of Partitioning and Hierarchical Clustering Algorithms.” Journal of Mathematical Modelling and Algorithms 5 (March). Springer:475–504. https://doi.org/10.1007/s10852-005-9022-1.

Rousseeuw, P. J. 1987. “A Graphical Aid to the Interpretation and Validation of Cluster Analysis.” Journal of Computational and Applied Mathematics 20 (November). Elsevier:53–65. https://doi.org/10.1016/0377-0427(87)90125-7.

Wishart, D. 2003. “K-Means Clustering with Outlier Detection, Mixed Variables and Missing Values.” In Exploratory Data Analysis in Empirical Research; Proceedings of the 25th Annual Conference of the Gesellschaft Fur Klassifikation E.v., University of Munich, March 14-16, 2001, 27:216–26. Springer Berlin Heidelberg. https://doi.org/10.1007/978-3-642-55721-7_23.

Yu, D., G. Liu, M. Guo, and X. Liu. 2018. “An Improved K-Medoids Algorithm Based on Step Increasing and Optimizing Medoids.” Expert Systems with Applications 92 (February). Elsevier:464–73. https://doi.org/10.1016/j.eswa.2017.09.052.

Zadegan, S.M.R, M. Mirzaie, and F. Sadoughi. 2013. “Ranked K-Medoids A Fast and Accurate Rank-Based Partitioning Algorithm for Clustering Large Datasets.” Knowledge-Based Systems 39 (February). Elsevier:133–43. https://doi.org/10.1016/j.knosys.2012.10.012.