Abstract

Image segmentation is an important problem that has received significant attention in the literature. Over the last few decades, a lot of algorithms were developed to solve image segmentation problem; prominent amongst these are the thresholding algorithms. However, the computational time complexity of thresholding exponentially increases with increasing number of desired thresholds. A wealth of alternative algorithms, notably those based on particle swarm optimization and evolutionary metaheuristics, were proposed to tackle the intrinsic challenges of thresholding. In codicil, clustering based algorithms were developed as multidimensional extensions of thresholding. While these algorithms have demonstrated successful results for fewer thresholds, their computational costs for a large number of thresholds are still a limiting factor. We propose a new clustering algorithm based on linear partitioning of the pixel intensity set and between-cluster variance criterion function for multilevel image segmentation. The results of testing the proposed algorithm on real images from Berkeley Segmentation Dataset and Benchmark show that the algorithm is comparable with state-of-the-art multilevel segmentation algorithms and consistently produces high quality results. The attractive properties of the algorithm are its simplicity, generalization to a large number of clusters, and computational cost effectiveness.

1. Introduction

Image segmentation remains an important problem in the fields of digital image processing, pattern analysis, image understanding, and computer vision. It is the process of subdivision of an image into homogeneous and disjoint sets sharing similar properties such as intensity, color, and contour [1]. Applications of image segmentation cut across disciplines as diverse as pattern recognition, image compression, content-based image retrieval, image processing, microscopic imaging, automatic image analysis, image and video understanding, video security, computer-aided surgery, automated medical diagnosis, human-machine interface, moving object tracking, image enhancement, biometric access control, computer vision, deoxyribonucleic acid sequencing, automatic vehicle recognition, optical character recognition, remote sensing, land cover classification, and machine learning [210]. Image segmentation is not a trivial problem because the majority of images are affected by factors such as noise content, occlusion, weak object boundary, inhomogeneous object region, weak contrast, nonuniformity of illumination, and reflectance [4, 11]. Different algorithms have been proposed over the years for image segmentation and two important properties commonly employed to categorize various segmentation algorithms are homogeneity and discontinuity of pixel properties. Edge-based methods are based on pixel discontinuity properties and they segment images by detecting pixels with rapid transition in intensities between regions [12, 13]. The image segmentation methods based on the property of homogeneity include thresholding, clustering, and region growing as well as region splitting and merging [14, 15]. These methods segment images based on the predefined objective functions and characteristics such as texture, shape, color, intensity, and other domain specific features [11, 16].

Image thresholding is very popular because it offers intrinsic benefits such as compact storage space, fast execution speed, simplicity, low computational cost, and real-time applicability [17, 18]. Thresholding attempts to identify and extract an object from its background on the basis of the distribution of gray levels or texture in the image object [19]. The prime objective of gray level image thresholding is to divide a gray level image into a -number of predetermined partitions or clusters based on different thresholds [20]. Image thresholding can be appositely classified into parametric and nonparametric methods. In parametric approach, a statistical model and a histogram are employed to obtain a set of parameters that control the fitness of the model. In the nonparametric approach, thresholds are chosen through the optimization of certain objective functions such as between-cluster variance, within-cluster variance, two-dimensional entropy, and cross entropy [21, 22]. Nonparametric thresholding can be realized either globally or locally. In the global implementation, object and background pixels are discriminated by comparing them with a chosen threshold and binary partitioning is used to segment the image. On the other hand, local thresholding finds a local threshold by examining the intensity values of the local neighborhood of each pixel. The Otsu algorithm [21] is one of the most widely used for global nonparametric thresholding. It provides a threshold by maximizing the between-cluster variance in the histogram of a gray level image. The algorithm is not complicated because it assumes the image histogram to be bimodal, meaning that there are two clusters or one threshold. However, bimodal thresholding cannot always achieve satisfactory results when the histogram of the image gray level is nonbimodal [20, 23]. Despite the position of Otsu that extending his algorithm to multilevel thresholding is a straightforward problem, he affirmed that “the selected thresholds become less credible as the number of levels increases” [21]. This is because the choice of appropriate thresholds is germane; making this choice manually or automatically is often complicated and several iterations are required to compute the zeroth and the first-order moments of multiple clusters [24].

Clustering techniques were described in [25] as multidimensional extensions of thresholding concepts and they are reputed in the literature to be very popular for image segmentation [26]. Clustering is an unsupervised learning task in which a finite set of clusters is identified so as to classify the pixels in a digital image. In clustering, the number of clusters is known a priori and image pixels are grouped into appropriate clusters based on the principle of intracluster similarity maximization or intercluster similarity minimization [27]. Clustering based segmentation algorithms can be divided into two broad categories, which are hard clustering and soft clustering. Hard clustering methods are used for datasets with sharp boundaries between clusters and a pixel belongs to only one cluster. -means clustering is one of the most prevalent hard clustering algorithms because of its implementation simplicity and low computational costs [28]. The conventional -means algorithm assigns each pixel in an image to the respective clusters on the basis of the minimum Euclidean distance. Meanwhile, one of the shortcomings of this algorithm is that poor assignment of pixels often occurs in situations where the pixel has the same minimum Euclidean distance to two or more clusters. An improper initialization procedure also gets the cluster centers in the conventional -means algorithm trapped in local minimal or stuck to the initial values so that they are unable to represent any group of data effectively. These scenarios often result in some clusters becoming dead centers, a situation in which clusters have no members [29].

In this study, we develop a new multilevel image segmentation method named Pixel Intensity Clustering Algorithm (PICA). Firstly, we utilized the set of image pixel intensities, the number of desired clusters, and a linear partitioning scheme to perform the initialization of cluster centroids. This initialization strategy is a notable departure from the use of randomization to determine initial cluster centroids, which is the custom in conventional clustering based segmentation algorithms. Our initialization scheme fulfills an important purpose of eliminating the incidence of dead centers, which is a common problem in conventional clustering based image segmentation algorithms [29]. Secondly, we adopted Otsu’s between-cluster variance criterion function [21] frequently used in multilevel thresholding to place pixel intensities into suitable clusters. Thirdly, in order to obtain the segmented image output, the input image is reconfigured based on the outcome of the clustering procedures in the preceding steps. The reconfiguration process uses the input image as a reference to generate the segmented image; the pixel intensity at each spatial location in the output image is assigned the centroid of the cluster to which the pixel intensity in the input image belongs. The number of pixel intensities in an input image histogram is generally far smaller than the size of the image; therefore, the use of pixel intensities of the image histogram (here called pixel intensity set) both for centroid initialization and for pixel clustering is a strong strategy for the computational cost reduction in PICA. The use of this approach obviously implies that the time complexity of PICA is proportionate to the size of the input image. PICA does not utilize spatial information as widely suggested [30], yet it achieves remarkable performance in segmentation result and computational time. Moreover, PICA satisfies survivability and recoverability principles because it can work for any number of clusters from 2 up to the maximum number of pixel intensities in the input image without breaking down. It also satisfies the principle of simplicity because it does not have to perform complex mathematical operations to achieve the desired result.

2. Relevant Literature

In recent years, a number of improved Otsu algorithms for multilevel thresholding were credited with the capability to reduce computational inefficiency. For instance, a Two-Stage Multithreshold Otsu (TSMO) algorithm was developed to improve the performance of the original Otsu algorithm [31]. The principle employed in TSMO for finding multilevel thresholds of an image is similar to that of the original Otsu method. It utilizes the statistical groups (16, 32, 64) with each group containing (=) gray levels to determine thresholds in two stages by applying Otsu criterion twice. Low classification errors were reported in the application of TSMO for test images with five thresholds. However, as reported, the TSMO algorithm cannot always guarantee a satisfactory performance for many complex images. The TSMO algorithm was later extended by Huang et al. [23], wherein they used the valley estimation scheme for automatic cluster determination. The experimental results of their study showed that the speed of computation is about 19000 times faster than the original Otsu algorithm when the number of clusters is seven. However, real-time performance was achieved when the number of clusters is fewer than six or five thresholds.

Besides the TSMO algorithm and its extensions, other algorithms have utilized different objective functions for multilevel thresholding. For example, Dong et al. [18] developed a linear iterative algorithm to determine thresholds that minimize a weighted sum-of-squared error objective function. The technique employed is mathematically tantamount to the Otsu algorithm, but 200 times faster in computational time. The algorithm may predictably not produce satisfactory results in large clusters because it was not extended to segment images beyond three clusters. Metaheuristic methods, such as Genetic Algorithm (GA), Artificial Bee Colony (ABC), Differential Evolution (DE), electromagnetism optimization (EMO), Bat Algorithm (BA), particle swarm optimization (PSO), Darwinian PSO (DPSO), and Fractional-Order DPSO (FODPSO) [1, 7, 17, 3236], have also been applied for multilevel thresholding. One of the best algorithms known amongst these metaheuristic algorithms is the PSO. It was illustrated in [37] that the PSO-based segmentation method acted better than other metaheuristic methods in terms of precision, robustness of results, and runtime. Nevertheless, a general problem with the PSO and similar optimization methods is that they may be trapped in local optimum points which consequently imply that the algorithm may work for some problems but may fail in others. DPSO and FODPSO were later proposed to solve this problem and it was proved that the FODPSO is faster than the PSO and more efficient than the DPSO [38, 39]. However, since all metaheuristic algorithms are random and stochastic, the outputs of FODPSO algorithm are not always the same in each run and, for large data, the efficiency of the method is constrained to a great extent [36].

There are several improvements of the conventional -means algorithm in the literature, which include moving -means (MKM), adaptive moving -means (AMKM), adaptive fuzzy moving -means (AFMKM), and enhanced moving -means (EMKM) [29, 40, 41]. The MKM, AMKM, and AFMKM methods homogeneously segment images, but they are still extremely sensitive to the poor initialization problem which is the major defect of clustering based segmentation methods. The EMKM, which is an improved version of AMKM, is nevertheless less sensitive to the initialization problem. However, the literature is replete with the fact that MKM, AMKM, AFMKM, and even EMKM cannot satisfactorily differentiate between dead centers and clusters with zero intracluster variance. Consequently, these methods fail to adequately distribute the pixels in an image into the appropriate clusters [29]. Soft clustering was later introduced to solve some of the inherent problems in hard clustering methods. Soft clustering algorithms are used when there are no hard boundaries amongst objects in an image and they introduce fuzziness by partly allocating pixels to all clusters with different degrees of membership. Examples of soft clustering methods are fuzzy -means (FCM), Gustafson-Kessel, Gaussian mixture decomposition, and fuzzy -varieties, but one of the most popular of all is the FCM algorithm [28, 29]. When compared with its hard clustering counterpart, conventional FCM is able to preserve more information from the original image. Nevertheless, the conventional FCM algorithm does not consider spatial information in the original image and because of this negligence, the algorithm is very sensitive to noise and outliers in the image. This shortcoming has led to the development of several other algorithms [42] that incorporated spatial information to enhance the conventional FCM algorithm. It is however reported in the literature that the computational time of many of these enhanced FCM algorithms is dependent on the size of the image and, therefore, the larger the image size, the more the segmentation time [42].

The foregoing review of existing studies generally shows that both thresholding and clustering based multilevel image segmentation algorithms in the literature are laden with research gaps [6, 26]. These gaps served as a strong motivation for this study.

3. Methodology

3.1. Proposed Algorithm

There are three main steps involved in PICA, initialization of cluster centroids, allocation of pixel intensities into clusters, and computation of output image. PICA fulfills the essential properties of clustering, which implies that, given an input image, the algorithm finds clusters such that the similarity of the pixel intensities in the same cluster , , is high while pixel intensities from different clusters are highly dissimilar and the clusters fulfill the following desirable properties [43]:(a), ; every cluster has at least one pixel intensity and there is no dead center syndrome.(b), ; a pixel intensity cannot belong to more than one cluster.The formulation of PICA proceeds as follows, assuming inputs to the algorithm are 2-dimensional grayscale image and the number of desired clusters is specified.

Step 1 (initialization of cluster centroids). The initialization of cluster centroids begins with the estimate of cluster weight. Cluster weight is the cumulative probabilities of pixel intensities in a cluster. If the input image is represented in pixel intensities , the number of pixels at intensity level is denoted by and the total number of pixels equals . The occurrence probability for a given pixel intensity is given by

The linear partitioning of a set of image pixel intensities is a useful strategy to guess the initial cluster centroids. Let , , be the set of image pixel intensities such that for every element the condition is satisfied. The purpose of the linear partitioning strategy is to guess the initial cluster centroids from the set . This set partitioning strategy provides a mechanism to avoid the incidence of dead centers when guessing the initial cluster centroids. The dead center syndrome is a common problem often associated with clustering methods [30, 4345]. In codicil, the set partitioning strategy provides an easy generalization of our algorithm to the case of number of clusters with lower computational costs. The generalization of multilevel image segmentation portrays the recoverability property of our algorithm. The set partitioning strategy of PICA replaces the random selection strategy, which is the usual practice in the conventional clustering algorithms [29].

Supposing the pixel intensity is to be selected as the initial centroid of the cluster, , the weight of this cluster can be estimated as follows: The variable is interpreted in this work as the cluster label, which is a number that associates a given image pixel intensity with its cluster. This implies that two or more image pixel intensities belonging to the same cluster would have the same cluster label for the purpose of easy grouping. The index of the pixel intensity is obtained as follows:The cluster centroid (), which is widely known as the center of gravity or center of mass of the cluster, is defined asThe parameter is defined as follows:

Step 2 (allocation of pixel intensities into clusters). The remaining number of image pixel intensities in not already allocated in clusters in Step 1 has to be allocated to suitable clusters based on the maximization of between-cluster variance. The cluster labels are then assigned to the allocated image pixel intensities and cluster centroids are updated accordingly. The maximization of the generalized between-cluster variance in allocating pixel intensities to suitable clusters, as used in PICA, is a departure from the minimization of Euclidean distance objective function that has hitherto been the norm in the clustering based image segmentation methods. In conventional clustering based image segmentation methods, cases in which a pixel has the same minimum Euclidean distance to two or more clusters have been reported to result in poor allocation of pixel intensities, which consequently lead to poor segmentation result [29, 46]. The between-cluster variance is defined as the sum of weighted squared distances (variances) between cluster centroids and grand or global centroid [22]:where the parameter is the global centroid of the image pixel intensities and it can be estimated asIf is the pixel intensity at the spatial location in ( is the height and is the width) of the input image, the global centroid can also be estimated directly from the image data, instead of from the image histogram as follows:The weight and centroid of the cluster that the pixel intensity , , is allocated to have to be updated according to the following update rules:This update mechanism ensures that the weights of all clusters add up to unity, following the probability theory; that is,

Step 3 (computation of output image). The output image is generated by assigning to the pixel intensity at the spatial location () in the output image the cluster centroid of the corresponding pixel intensity at the spatial location () in the input image. Given a pixel intensity at location () from an input image, the cluster label of this pixel intensity has to be determined. Using this cluster label, the centroid of the cluster that this pixel was allocated to can be determined. The implementation of PICA could be compactly outlined based on the above description as in Algorithm 1.

Input: grayscale image, number of clusters.
Output: grayscale image.
Let , represents the set of image pixel intensities and assuming BCV is the between cluster variance.
(1) for all    do
(2)  
(3)  estimate the th cluster centroid using the pixel intensity
(4)  assign cluster label to the pixel intensity
(5)  swap the pixel intensity with the pixel intensity
(6) end for
(7) for all    do
(8)  for all    do
(9)     tentatively allocate pixel intensity to the th cluster
(10)    BCV =    (variance of the th cluster centroid)
(11)    for all    and   not equal to   do
(12)         BCV = BCV +    (variance of the th cluster centroid)
(13)    end for
(14)  end for
(15)  permanently allocate to the cluster that gives maximum BCV
(16)  assign cluster label to the allocated pixel intensity
(17)  update cluster centroid
(18) end for
(19) compute output image
(20) stop

3.2. Estimated Time Complexity

The analysis of time complexity of PICA can be performed for each step of the algorithm. In Step 1, the number of instructions can be executed in , which is linear with respect to the number of pixel in the input image. The total number of instructions executed in Step 1 is the number of instructions to create an image histogram, compute probabilities of pixel intensities, determine set of image pixel intensities, compute global centroid, estimate initial cluster weight, determine initial cluster centroids, assign cluster labels to pixel intensities, and swap pixel intensities. All of these instructions can be executed in time complexity of , provided . The number of instructions required to compute the output image in Step 3 also can be executed in . The instructions in Step 2 are only executed whenever and they can be executed in . The time complexity of PICA can therefore be approximated by .

The computational time complexity of PICA is compared to that of conventional -means algorithm and optimized -means algorithm to strengthen the illustration of the computational speed advantage of the algorithm. We decided to select these algorithms because the conventional -means algorithm was proven to have a shorter processing time with complexity of and also the optimized -means algorithm was reputed to have time complexity of [29, 47]. These parameters, (which corresponds to ), , (which corresponds to ), , and , are, respectively, the number of pixels in an input image, the number of attribute dimensions, the number of clusters, the number of iterations, and the number of intensity values of the conflict pixels to be assigned to their respective clusters. PICA outperforms these algorithms in computational time complexity as parameter values become large. In fact, practical experience shows that KM, PSO, DPSO, and FDPSO took longer time to process and broke down when on the selected images from Berkeley Segmentation Dataset and Benchmark.

4. Experiments and Results

The experimental setup for this study and the numerical results obtained for the proposed algorithm are reported. Furthermore, empirical results of segmentation using the proposed algorithm were compared both qualitatively and quantitatively with state-of-the-art algorithms. The following subsections contain details of the experimental setup, the numerical results, and the comparative results.

4.1. Experimental Setup

All the experiments in this study were performed on an Intel Core i7 processor @ 3.5 GHz with 64 GB of RAM running the Windows XP operating system. The algorithm was implemented in C/C++. The program is very compact, well-structured, and easy to follow, which is one of the bases of the intrinsic simplicity of the algorithm. The input image to the program was in grayscale, but the program can as well process color images by processing each RGB channel separately and then combine the results. The ITU-R recommendation (ITU-R BT.709-5) was applied to convert a color image to a grayscale image before the program was executed. We collected twelve images from the Berkeley Segmentation Dataset and Benchmark with identification numbers “45077,” “157055,” “78098,” “42049,” “253027,” “169012,” “210088,” “208001,” “10081,” “155060,” “25098,” and “35070” [48].

4.2. Numerical Results

In this section, we report the results of the computations carried out by the proposed algorithm to determine some salient parameters, which are initial cluster centroids, updated cluster centroids, and cluster weights. The values of these parameters for the twelve images selected from the Berkeley Segmentation Dataset and Benchmark are shown in Tables 1, 2, and 3. In particular, Table 1 shows the initial cluster centroids computed for each of the twelve images from clusters 3 to 12 using (1) to (5). The values that are reported in Table 1 clearly indicate orderly sequences of numbers across the different clusters of many of the test images. This result illustrates the merit of the linear partitioning scheme for the initialization step of PICA. It can also be seen in Table 1 that the more the number of clusters, the better the spread of the initial cluster centroids towards the highest pixel intensity value in the input image. This apparently explains why better segmentation is achieved at a higher number of clusters for the proposed PICA because of the better representation of the pixel intensity in the input image.

Table 2 shows the final updated cluster centroids for clusters 3 to 12 for all the test images used in this study. These values were computed based on (9). The cluster centroid update procedure based on these equations can be deemed successful as reflected in the visual result of the segmentation. For instance, in Table 2, the initial centroids of the image “45077” for four clusters, which were 9, 69, 129, and 189, were updated to the final values of 39, 74, 109, and 159, showing that an update was actually performed. Similar results were obtained for all the other test images used in this study as shown in Table 2. The updated cluster centroids finally computed by the proposed algorithm tend towards the higher pixel intensity value in the input as the number of clusters is increased. This significantly contributes to the enhancement of the segmentation result in higher clusters. These values of the updated cluster centroids are ultimately used to compute the output images in the final step of PICA.

Table 3 shows the weights of clusters from 3 to 12 for all the test images in this study. It can be noted in Table 3 for all the images that cluster weights for a particular image add up to unity. This is in perfect agreement with the probability theory in (10). In codicil, it can be observed in Table 3 that no cluster weight for all the test images is equal to zero. This implies that there is no possibility for the algorithm to break down at any number of clusters and there is no incidence of empty clusters, which can result in the dead centers syndrome using the proposed algorithm.

4.3. Comparative Results

In this section, we adopted both qualitative (subjective) and quantitative (objective) evaluations to compare the segmentation results of the proposed algorithm with KM, PSO, DPSO, and FODPSO segmentation algorithms. These existing algorithms were selected because KM clustering is reported to be one of the most prevalent clustering algorithms due to its implementation simplicity and low computational costs. The PSO-based segmentation methods (PSO, DPSO, and FODPSO) were also shown in the literature to act better than other metaheuristic methods in terms of precision, robustness of results, and runtime [28, 37].

4.3.1. Qualitative

Qualitative evaluation on the basis of human visual perception is a widely used strategy in the literature to carry out performance assessment of image segmentation algorithms [29]. The original grayscale images selected from the Berkeley Segmentation Dataset and Benchmark for evaluations are labelled as (a) in Figures 112 while the segmentation outputs by the conventional KM, PSO, DPSO, FODPSO, and the proposed algorithm are labelled as (b) to (f), respectively. The number of clusters for segmenting each of the test images varied from 3 to 14. Worthy of note is the fact that, based on visual appeal, the segmentation outputs of all the conventional algorithms and our proposed algorithm are only fairly acceptable at cluster/level 3. In general, there were improvements in the visual appeal of the outputs of the conventional algorithms and the proposed algorithm as we increased the number of clusters/levels from 4 up to 14. However, it can be observed that there are several white patches in the outputs of the KM algorithm as conspicuously illustrated in Figures 2, 3, 4, 5, 7, 8, and 11. The segmentation outputs of PSO, DPSO, and FODPSO algorithms also have a lot of black patches as noticeably shown in Figures 2, 3, 4, 5, 6, 7, 8, 11, and 12. Conversely, the output of the proposed algorithm is shown across the entire test images and from clusters 4 to 14 to be more visually appealing because no white or black patches are found in any of the results. This qualitative evaluation clearly shows that our algorithm achieved more homogeneous segmentation outputs than the other comparative algorithms in this study. Quantitative evaluation is further carried out for the selected test images using the conventional algorithms and the proposed algorithm.

4.3.2. Quantitative

Quantitative evaluation objectively computes the performance of segmentation algorithms with appropriate similarity metrics and it is not subject to human errors like in the qualitative evaluation method. Consequently, for the quantitative aspect of this study, we employed some metrics to compare the proposed algorithm with the other algorithms. These metrics are the Jaccard index () and the Structural Similarity Measure (SSIM). The maximum Jaccard index and SSIM of 1 are only achieved if the two images are identical. These similarity metrics (Jaccard and SSIM) can be mathematically defined in terms of two images and .

Given that denotes the cardinality of image , which is a count of the number of elements that are present in , the Jaccard similarity index between the two images and is given as [49]where denotes the intersection between the two images and represents all elements that are in both images, while denotes the union between the two images and represents all elements that are in either of them.

The SSIM index was proposed to predict human preferences in image quality assessment. To compute SSIM, the means (, ), the variances (, ), and the cross-covariance () of images and are first computed and are based on these values, and the SSIM between images and is given as [50]where and are small positive constants, is the mean of , and is the variance.

The values obtained for these metrics in (11) and (12) when applied to the test images for clusters 3 to 14 are shown in Table 4. As shown in the table, both the Jaccard index and the SSIM index for the proposed algorithm improved progressively from cluster 3 up to cluster 14 for the different test images. In addition, the results of the two metrics for the proposed algorithm are better than the values obtained for the other algorithms compared from clusters/levels 3 up to 14 across all the test images (Table 4). These results further consolidate the qualitative comparison result, which earlier established that our proposed algorithm performs better than KM, PSO, DPSO, and FODPSO. The computed evaluation metrics also indicate that the values of the SSIM indices are consistently higher than the values of Jaccard indices for all the test images from clusters/levels 3 up to 14. For instance, at cluster 3 in Table 4, the Jaccard index for the proposed algorithm is 0.8639 while that of the SSIM index is 0.8979 and at cluster 14, the value of the Jaccard index is 0.9603 while that of the SSIM index is 0.9965. A similar trend was observed for all the conventional algorithms, even though the values obtained from the two metrics were lower for the conventional algorithms than the proposed algorithm. Apparently, the SSIM index value of 0.9965 obtained for the proposed algorithm at cluster 14 is a better quantitative representation of the visual appeal of the segmented image than the Jaccard index value of 0.9603. We can therefore safely infer that, based on the evaluations carried out in this study, the SSIM index is a better measure of similarity between an input image and a segmented image than the Jaccard index. Hence, the SSIM evaluation metric is highly promising for many other post-image segmentation processing tasks such as object features extraction and content retrieval.

5. Conclusion

In this study, a new clustering algorithm was proposed and applied to multilevel image segmentation problem. The objectives of the proposed algorithm were to improve the initialization of cluster centroids and the allocation of pixel intensities into clusters. This is to eliminate the dead center problem and inappropriate pixel allocation to clusters as commonly encountered in many of the existing clustering based multilevel segmentation methods. The simple mathematical concepts utilized in achieving the objectives are linear partitioning of the pixel intensity set and between-cluster variance criterion function. The new multilevel segmentation algorithm is conceptually simple, robust, and computationally cost effective. The algorithm was compared with KM, PSO, DPSO, and FODPSO on 12 real images from the Berkeley Segmentation Dataset and Benchmark. The qualitative results show that the proposed algorithm consistently produced patch-free segmentation, unlike the other existing algorithms. The quantitative results using the Jaccard index and the SSIM show that the segmentation results of our algorithm improve with increasing number of clusters across different images without a corresponding increase in computational cost, unlike the other algorithms. Our experiments further showed that when the number of clusters is equal to the number of pixel intensities , ( and ), the value of SSIM becomes 1, which is a confirmation of the survivability and recoverability capabilities of the algorithm. On the contrary, the other comparative algorithms in this study either break down or become unbearably slow for . This study opens new perspective on multilevel image segmentation because multilevel thresholding algorithms were previously considered suitable for solving multilevel image segmentation problem. Future work will involve a search for better initialization schemes and pixel intensities allocation concepts to further enhance the performance of the algorithm.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgment

E. Adetiba is on postdoctoral fellowship at the ICT and Society (ICTAS) Research Group, Durban University of Technology, South Africa, funded by the Durban University of Technology Research and Development Directorate. He is on postdoctoral research leave from the Department of Electrical & Information Engineering, College of Engineering, Covenant University, Ota, Ogun State, Nigeria.