Research Article  Open Access
Wenqi Hua, Lingfei Mo, "Clustering Ensemble Model Based on SelfOrganizing Map Network", Computational Intelligence and Neuroscience, vol. 2020, Article ID 2971565, 11 pages, 2020. https://doi.org/10.1155/2020/2971565
Clustering Ensemble Model Based on SelfOrganizing Map Network
Abstract
This paper proposes a clustering ensemble method that introduces cascade structure into the selforganizing map (SOM) to solve the problem of the poor performance of a single clusterer. Cascaded SOM is an extension of classical SOM combined with the cascaded structure. The method combines the outputs of multiple SOM networks in a cascaded manner using them as an input to another SOM network. It also utilizes the characteristic of highdimensional data insensitivity to changes in the values of a small number of dimensions to achieve the effect of ignoring part of the SOM network error output. Since the initial parameters of the SOM network and the sample training order are randomly generated, the model does not need to provide different training samples for each SOM network to generate a differentiated SOM clusterer. After testing on several classical datasets, the experimental results show that the model can effectively improve the accuracy of pattern recognition by 4%∼10%.
1. Introduction
Advances in network techniques provide more access to data, especially unlabeled data, which is a significant factor for the fast development of the clustering algorithm. Nowadays, clustering has an important role in areas like pattern recognition, image processing, recommendation system, and data mining. Also, the rising of ensemble learning naturally imports ensemble learning into the clustering, which is called clustering ensemble [1]. It is able, for the clustering ensemble model, to combine results of multiple clusterers while enhancing the performance compared with that of a single clusterer. Meanwhile, the clustering ensemble is robust facing noise or isolated points in input space.
A clustering ensemble model called cascaded selforganizing map (cascaded SOM) is proposed in this paper. Cascaded SOM realizes the ensemble function by learning responses of different clusterers to make a final decision. When learning the responses, the model is learning highdimensional data, which is insensitive to the change of values in a few dimensions. The characteristic of highdimensional data makes the model able to ignore the error of several clusterers and output the correct result.
Selforganizing map (SOM) was proposed by Kohonen in 1990 [2]. SOM is a competitive network that consists of an input layer and an output layer (competitive layer). SOM represents the distribution characteristics of input samples on the competitive layer using topographically ordered nodes and, in doing so, achieves clustering through dimensionality reduction. Due to its intrinsic nonlinear mapping capabilities on its lowdimensional neural surface, SOM is distinguished from other widely used clustering algorithms. Such advantage makes SOM a tool of visualizing nonlinear relations of data, topologybased cluster analysis, vector quantization, and projection of multidimensional data. Due to its versatility, SOM has been applied in areas like pattern recognition, imagetext processing, data mining, genomics, and medical diagnostics.
Since SOM was proposed, plenty of researchers have put effort to improve it and proposed several variant algorithms. GrowingSOM [3] improves the performance of SOM in incremental learning by replacing the original competitive layer with a variablescale one; TSSOM [4] changes the competitive layer from single layer into multiple layers with tree structure, whose recursive characteristic can help to reduce training time; GSOM [5] introduces genetic algorithm into the parameter initialization step of SOM to generate better parameter, compared with the original randomized parameter initialization method; similar to GSOM, particle swarm optimization (PSO) can also be applied in the parameter initialization step [6]. Moreover, there is research that compares SOM with an uncommon grid structure and a common grid structure, proving that SOM with uncommon grid structure performs better with data having special distribution [7]. During the past decade, researchers came out with different methods to improve SOM. For example, DASOM introduces denoising autoencoder to reduce the noise in input space [8]; constrained SOM preserves the topology structure by blocking the input space [9]; robust SSGSOM introduces HQ method into semisupervised growingSOM to improve network robustness [10]. PLSOM uses adaptive learning rate instead of learning rate reduction through training, which can focus on new patterns [11]. PLSOM2 solves the problem that the large error at the early period will lead to a small learning rate in the later period [12], as well as reducing the influence of the noise. Inspired by PLSOM, researchers use the eigenvalue of the autocorrelation matrix of the input vector to control the learning rate, reaching a faster converging speed [13]. In classic SOM, there might be several useless neurons between different patterns, reducing competitive layer utilization. A segmentation method based on the distribution of neurons is proposed to raise the competitive layer utilization of 1D SOM and get a more stable network structure [14].
Each neuron in the SOM competitive layer represents a prototype of one pattern in input space which is more representative than any input that inspires this neuron. Such characteristic is applied in several types of researches. Facing situation where samples of one pattern are far less than those of other patterns, SOM oversampling (SOMO) uses SOM to train prototype of minority samples as the input of SMOTE algorithm, providing oversampling for minority samples as well as avoiding the oversampled data being representative [15]. Also, in the real world, there should be neither samples nor prototypes in certain areas in the input space. Classic SOM is not able to handle this problem. Forbidden region SOM (FRSOM) set forbidden regions according to prior knowledge [16]. During training, FRSOM considers the distance differently; that is, the path which represents the distance between two points must not go across the forbidden area. This strategy ensures that no prototype will be set in the forbidden region, improving the clustering result.
Ensemble learning uses multiple learners to figure out a single question, and the answer to the question is given by combining all the outputs of the learners, which can improve the result [17]. When introduced to the clustering algorithm, ensemble learning can reduce the effect caused by noise, leading to higher robustness. The clustering ensemble is usually divided into two parts, which are cluster generation and cluster fusion. Specifically, there are n samples in the input space X, learned by M clusterers, and a fusion algorithm combines all the clusterers to get the final learner P (see in Figure 1).
Therefore, clustering ensemble algorithms can be divided into those which focus on clusterer generation and those which focus on clusterer fusion. As for clusterer generation, there are algorithms which generate different clusterers by different clustering algorithms [1], or by the same clustering algorithm with different parameters [18, 19]. Also, dataset generation methods are used to generate different dataset from the original dataset to train different clusterers [20–22]. As for clusterer fusion, there is also plenty of research. An algorithm resamples different samples into the same cluster [23]; another one uses graph segmentation algorithm to solve clustering ensemble task [24]; furthermore, a mixed model is proposed [25]; Dempster–Shafer evidence theory is also introduced to clustering ensemble area [26]. Apart from all these, some researchers classify the sample into a transition sample and core sample and then use the core sample as the cluster segmentation basis [27, 28] (see Figure 2). Besides these two aspects, prior knowledge is also applied to improve the clustering ensemble result [29, 30].
Most research about the clustering ensemble introduces a new algorithm into the ensemble step to correct the result of certain subclusterers and combine results of all clusterers. According to research into using selforganizing map quantization error to indicate the singlepixel change in large random patterns [31], the selforganizing map cannot detect the small shift of single pixel. However, this study indicates that the selforganizing map itself can ignore a small shift in highdimensional data, and such characteristic makes the selforganizing map able to be introduced to the ensemble step.
2. Materials and Methods
The SOM network can create a map between highdimensional data and lowdimensional data, with original topology structure reserved. As a nonlinear dimension reduction algorithm, SOM trains competition layer neurons which represent prototypes of pattern in highdimensional space to correspond with the topology structure of the original data.
The network structure of SOM is shown in Figure 3. The network consists of k neurons where , and each neuron has a weight vector whose dimension equals that of input space. Two values need to be calculated when SOM is being trained. The first is , which is the topology distance of two neurons on the competitive layer; the second is , the Euclidean distance between the input vector and the weight vector. During the training period, the target is to reduce the Euclidean distance of the weight vectors belonging to neurons that have small topology distance.
When the SOM is trained on dataset D, for each input x_{i} there will be a winner neuron whose weight vector’s Euclidean distance from the input vector is the smallest in the competition layer. Then, the weight vectors of both the winner neuron and its neighboring neurons are updated as shown in Algorithm 1, where η is the learning rate and is a neighborhood function which decreases by the increasing of .

Each round of training will force the weight vectors of the winner neuron and its neighboring neurons to move towards the input vector x_{i}. Along with the iteration time t growing, the value of decreases, leading to the separation of the neurons in the competition layer. Finally, the neurons will be spontaneously gathered around different clusters. Since one or several clusters can represent a pattern in the input dataset, the Euclidean distance of weight vectors can represent the distance between different patterns, i.e., reserve the data topology structure.
Cascaded SOM combines the output of multiple competition layers as the input of another SOM model, leading to the ensemble of multiple SOM models, which can improve the performance of the learning model. Figure 4 shows the structure of cascaded SOM (2layer cascaded SOM as an example).
Since the competition layer of the SOM network indicates the topology structure of input patterns, the cascaded SOM needs to retain this characteristic, which cannot be provided by the original output of the SOM network. Inspired by the winner neuron of the same input pattern appearing in the same area, in order to reach the target, the output of the competition layer is changed into a onehot vector to send the location of the winner neuron to the next layer. After the processing, the output data can be used as the training sample of the next layer, whose training step is exactly the same as shown in Algorithm 1.
The conversion of the output of the competition layer increases the dimension of the input space of the next layer, and the change of the winner of a single competition layer only changes the value on two vectors. When calculating Euclidean distance, the result would only be slightly influenced by the difference in several dimensions, which improves the robustness of the model.
3. Results and Discussion
To measure the improvement brought by the proposed cascaded SOM to the classic SOM, several experiments are conducted on MNIST handwritten digit dataset, USPS handwritten digit dataset, and Fashion MNIST dataset (Figure 5). Data in MNIST and Fashion MNIST are 2dimensional arrays with 28 28 elements, while those in USPS are 2dimensional arrays with 16 16 elements. In the experiments, the 2dimensional array is transferred to 1dimensional array, which means that the array is flattened. The data in the MNIST and the Fashion MNIST are transferred into 1 784 arrays, and the data in USPS are transferred into 1 256 arrays. The rest of this section demonstrates the design of these experiments and discusses the results of the experiments.
(a)
(b)
(c)
In the experiments, a twolayer cascaded SOM network is used. Iteration time of training, network number of the first layer, the difference between training samples, and the network size are chosen to test their influence on the performance of the proposed algorithm. The accuracy of the clustering is utilized to evaluate the result of the clustering:where N is the total number of the samples, is the set of the clusters, is the real label of the sample, and is the category of ; compares and and then outputs 1 if they are the same and 0 if they are not.
When keeping the iteration time of the second layer and the network number of the first layer fixed, the accuracy of the firstlayer SOM networks as well as the cascaded SOM network and the iteration time of the firstlayer SOM networks are positively correlated before the firstlayer SOM networks are converged (Figure 6). After that, the accuracy of the firstlayer SOM networks has no obvious improvement. Compared with the first layer, the accuracy of the cascaded SOM network has a stable improvement (5% on MNIST and USPS, 4% on Fashion MNIST).
(a)
(b)
(c)
When the parameters of the firstlayer network are fixed while the iteration time of the secondlayer SOM network increases, the improvement of the accuracy is shown in Figure 6. As a result of the incomplete training, the accuracy has not been improved but has dropped sharply when the iteration time is only one. After several rounds of training, the accuracy improvement of the cascaded SOM network tends to a fixed value (5.7% on MNIST, 5% on USPS, 4.5% on Fashion MNIST).
The core of cascaded SOM is the exploitation of the insensitivity to changes in the values of a small number of dimensions. Therefore, the ratio of numbers of the valuechanged dimension and the total dimension can influence the performance of the cascaded SOM network. The number of the firstlayer networks is utilized to control the ratio mentioned above. Also, the firstlayer networks should be different from each other to avoid misclassification towards the same input, which can be ensured using different initialized parameters or different training samples.
When all the firstlayer SOM networks are trained with the same training dataset, and the difference of the parameter is provided by randomly initialized weight vector, as shown in Figure 7, the improvement of the accuracy reaches a limit if the number of the firstlayer networks increases to 13.
While the difference among firstlayer networks is provided by both different training datasets and randomly initialized weight vector, the extra difference leads to better generalization ability. As a result, the upper limit of the accuracy is higher than the cascaded SOM network with the firstlayer SOM network trained with the same dataset. However, when the number of the firstlayer SOM networks is small, the generalization ability is not high enough to reduce the accuracy brought by the firstlayer training dataset being different from the secondlayer training dataset. Therefore, as shown in Figure 8, the cascaded SOM network with the firstlayer SOM network trained with the same dataset has higher accuracy than that with firstlayer SOM network trained with the different datasets when the number of firstlayer SOM networks is small.
(a)
(b)
(c)
Meanwhile, the size of the firstlayer SOM network can influence the performance of the SOM network. The next experiment is to figure out the effect of the size of the firstlayer SOM network on accuracy. In one round of experiments, all the firstlayer and secondlayer SOM networks have the same size, the shape of those SOM networks is square, and the number of firstlayer SOM networks is 10.
A bigger SOM network leads to a better performance of the cascaded SOM network (Figure 9). The reason is that a bigger SOM network is beneficial to separate different clusters and reduce the number of the intercluster neurons, which is hard to judge which cluster it belongs to. This benefit can work on both the first layer and the second layer, leading to an increasing accuracy of both firstlayer SOM network and cascaded SOM network. Contrary to the performance of the cascaded SOM network, there is a decrease in the accuracy of the firstlayer SOM network when the size of the network keeps increasing. This phenomenon appears because each neuron can gain less knowledge if the network size keeps increasing, which raises the noiseknowledge ratio and reduces accuracy. Nevertheless, when it comes to the cascaded SOM network, larger network size means higherdimension input of the secondlayer SOM network, which improves the antinoise ability of the network, making the accuracy of the cascaded SOM network keep improving instead.
(a)
(b)
(c)
Table 1 shows the performance of the cascaded SOM network with 3 different parameter settings on MNIST, USPS, and Fashion MNIST dataset. The accuracy of the cascaded SOM is positively correlated with the iteration time of the firstlayer SOM network, the iteration time of the secondlayer SOM network, the number of the firstlayer SOM networks, and the size of SOM network. Due to the difference among the data distribution and the data complexity of different datasets, the performance of the cascaded SOM network is slightly different, but, in general, the cascaded SOM network has better performance than the classic SOM network.
 
A: iteration number of the first layer: 30; iteration number of the second layer: 30; number of the firstlayer networks: 10; network size: 10 × 10. B: iteration number of the first layer: 50; iteration number of the second layer: 50; number of the firstlayer networks: 10; network size: 15 × 15. C: iteration number of the first layer: 50; iteration number of the second layer: 50; number of the firstlayer networks: 20; network size: 20 × 20. 
All the experiments above use the 2layer structure. Table 2 shows the performance of the 3layer cascaded SOM network. The additional third competition layer can indeed provide very limited promotion to the accuracy based on the 2layer cascaded SOM network. This indicates that the additional competition layer cannot deal with the lack of difference in the secondlayer SOM network trained based on the same firstlayer SOM network. Such a problem can be solved by increasing the difference between both firstlayer and secondlayer SOM networks, but the extra cost of training time should also be considered.

The experiments on MNIST, USPS, and Fashion MNIST indicate that the proposed cascaded SOM network can stably improve the performance of the SOM network. At the same time, like any other ensemble model, the difference among sublearners raises the generalization ability, which can result from randomly initialized weight vector or different training datasets. Though there is evidence that adding an extra layer can further improve the performance of the cascaded SOM network, the extra time cost must be considered.
The cascaded SOM network proved to be an effective algorithm, compared to ensemble classic SOM network, in improving the model. The rest of this section will compare the proposed algorithm and other clustering ensemble algorithms.
In the following experiment, ARI (adjusted Rand index) and NMI (normalized mutual information) are utilized as the evaluation indices of the clustering ensemble algorithm, which are defined as follows:where n is the total number of samples, is the number of objects in the intersection of clusters and , is the clustering result and is the true class numbers of the samples, n_{i} and n_{j} are the numbers of samples in clusters and , respectively, and is the binomial coefficient . The maximum value of ARI is equal to 1. The maximum value of NMI is equal to 1.
Eight realworld datasets from the UCI Machine Learning Repository were used, including Iris, Wine, Thyroid, Multiple Features (Mfeatures), and Ionosphere dataset. Table 3 shows the details of these datasets. BCW has an attribute with missing values in some objects, which is removed. The second attribute in the Ionosphere dataset is also removed.

Six other clustering ensemble algorithms are chosen, which are COAverage (CoAssociation method using the average linkage method) [32], ONCEAverage (ObjectNeighborhood Clustering Ensemble using the average linkage method) [33], DSCE (DualSimilarity Clustering Ensemble method) [34], ACE (Adaptive Clustering Ensemble) [35], DICLENS (Division Clustering Ensemble) [36], and MCLA (MetaClustering Algorithm) [1].
The results of the experiments are shown in Tables 4 and 5. The proposed algorithm has the best performance on several datasets (Wine, Mfeatures, and Glass) and has a relatively good performance on the other datasets too.
 
The bold values mean the best ARI value of the different algorithms for each dataset. 
 
The bold values mean the best NMI value of the different algorithms for each dataset. 
4. Conclusions
In this paper, a new clustering ensemble approach is investigated. The proposed method is based on introducing a cascaded structure to the selforganizing map. To test the clustering ensemble ability of the algorithm, experiments are conducted on MNIST, USPS, and Fashion MNIST dataset. The experiments on those datasets indicate that the proposed cascaded SOM network has the following characteristics:(1)It can stably improve the performance of the SOM network.(2)The difference among sublearners, which can result from randomly initialized weight vector or different training datasets, raises the generalization ability.(3)The 2layer cascaded SOM network can improve the performance of the SOM network, while the 3layer cascaded SOM network can only slightly increase the accuracy compared to that of a 2layer one, yet the extra time cost must be considered.
The proposed algorithm is proved to have advantages on several datasets, compared with other clustering ensemble algorithms.
Also, the proposed algorithm can still be improved. For now, all the methods, except using different training datasets, to improve the performance of the cascaded SOM are increasing the data dimension, which increases the training time cost at the same time. Therefore, the network can hardly reach its theoretical limit in practical application due to time cost control.
Data Availability
The data used to support the findings of this study are included in the article.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Authors’ Contributions
Wenqi Hua and Lingfei Mo contributed equally to this work.
Acknowledgments
The authors would like to thank the Future X Lab of Southeast University, China, for providing equipment for conducting the experiments. This study was funded by the National Science Foundation of China (61603091, MultiDimensions Based Physical Activity Assessment for the Human Daily Life).
References
 A. Strehl and J. Ghosh, “Cluster ensembles—A knowledge reuse framework for combining multiple partitions,” Journal of Machine Learning Research, vol. 3, no. 3, pp. 583–617, 2003. View at: Google Scholar
 T. Kohonen, “The selforganizing map,” Neurocomputing, vol. 21, no. 1–3, pp. 1–6, 1998. View at: Publisher Site  Google Scholar
 B. Fritzke, “Growing cell structures–A selforganizing network for unsupervised and supervised learning,” Neural Networks, vol. 7, no. 9, pp. 1441–1460, 1994. View at: Publisher Site  Google Scholar
 P. Koikkalainen, “Tree structured selforganizing maps,” Kohonen Maps, Elsevier Science BV, Amsterdam, Netherlands, 1999. View at: Publisher Site  Google Scholar
 D. Polani, “On the optimization of selforganizing maps by genetic algorithms,” Kohonen Maps, Elsevier Science BV, Amsterdam, Netherlands, 1999. View at: Publisher Site  Google Scholar
 Y. W. Jing and Li F. Li, “SOM and PSO based alerts clustering in intrusion detection system,” Applied Mechanics and Materials, vol. 401–403, pp. 1453–1457, 2013. View at: Publisher Site  Google Scholar
 E. LópezRubio and A. Díaz Ramos, “Grid topologies for the selforganizing map,” Neural Networks, vol. 56, pp. 35–48, 2014. View at: Publisher Site  Google Scholar
 C. Ferles, Y. Papanikolaou, and K. J. Naidoo, “Denoising autoencoder selforganizing map (DASOM),” Neural Networks, vol. 105, pp. 112–131, 2018. View at: Publisher Site  Google Scholar
 J. AzorinLopez, M. SavalCalvo, A. FusterGuillo, J. GarciaRodriguez, and H. MoraMora, “Constrained selforganizing feature map to preserve feature extraction topology,” Neural Computing and Applications, vol. 28, no. S1, pp. 439–459, 2017. View at: Publisher Site  Google Scholar
 A. Mehrizi, H. Sadoghi Yazdi, and A. H. Taherinia, “Robust semisupervised growing selforganizing map,” Expert Systems with Applications, vol. 105, pp. 23–33, 2018. View at: Publisher Site  Google Scholar
 E. Berglund and J. Sitte, “The parameterless selforganizing map algorithm,” IEEE Transactions on Neural Networks, vol. 17, no. 2, pp. 305–316, 2006. View at: Publisher Site  Google Scholar
 E. Berglund, “Improved PLSOM algorithm,” Applied Intelligence, vol. 32, no. 1, pp. 122–130, 2010. View at: Publisher Site  Google Scholar
 A. A. Hameed, B. Karlik, M. S. Salman, and G. Eleyan, “Robust adaptive learning approach to selforganizing maps,” KnowledgeBased Systems, vol. 171, pp. 25–36, 2019. View at: Publisher Site  Google Scholar
 M. B. Gorzałczany and R. Filip, “Generalized selforganizing maps for automatic determination of the number of clusters and their multiprototypes in cluster analysis,” IEEE Transactions on Neural Networks and Learning Systems, vol. 29, no. 7, pp. 2833–2845, 2017. View at: Publisher Site  Google Scholar
 G. Douzas and F. Bacao, “Selforganizing map oversampling (SOMO) for imbalanced data set learning,” Expert Systems with Applications, vol. 82, pp. 40–52, 2017. View at: Publisher Site  Google Scholar
 A. D. Ramos, E. LópezRubio, and E. J. Palomo, “The forbidden region selforganizing map neural network,” IEEE Transactions on Neural Networks and Learning Systems, vol. 31, no. 1, pp. 201–211, 2019. View at: Publisher Site  Google Scholar
 A. P. Topchy, M. H. C. Law, A. K. Jain, and A. L. Fred, “Analysis of consensus partition in cluster ensemble,” in Proceedings of the Fourth IEEE International Conference on Data Mining (ICDM’04), Brighton, UK, November 2004. View at: Publisher Site  Google Scholar
 M. C. Naldi, A. C. P. L. F. Carvalho, and R. J. G. B. Campello, “Cluster ensemble selection based on relative validity indexes,” Data Mining and Knowledge Discovery, vol. 27, no. 2, pp. 259–289, 2013. View at: Publisher Site  Google Scholar
 H. Alizadeh, B. MinaeiBidgoli, and H. Parvin, “Cluster ensemble selection based on a new cluster stability measure1,” Intelligent Data Analysis, vol. 18, no. 3, pp. 389–408, 2014. View at: Publisher Site  Google Scholar
 Z. Yu, J. You, H.S. Wong, and G. Han, “From cluster ensemble to structure ensemble,” Information Sciences, vol. 198, pp. 81–99, 2012. View at: Publisher Site  Google Scholar
 E. Rashedi and A. Mirzaei, “A hierarchical clusterer ensemble method based on boosting theory,” KnowledgeBased Systems, vol. 45, pp. 83–93, 2013. View at: Publisher Site  Google Scholar
 Y. Hong, S. Kwong, H. Wang, and Q. Ren, “Resamplingbased selective clustering ensembles,” Pattern Recognition Letters, vol. 30, no. 3, pp. 298–305, 2009. View at: Publisher Site  Google Scholar
 B. MinaeiBidgoli, A. Topchy, and W. F. Punch, “Ensembles of partitions via data resampling,” in Proceedings of the International Conference on Information Technology: Coding and Computing, 2004. Proceedings. ITCC 2004, vol. 2, Las Vegas, NV, USA, 2004. View at: Publisher Site  Google Scholar
 Z. Yu, P. Luo, J. You et al., “Incremental semisupervised clustering ensemble for high dimensional data clustering,” IEEE Transactions on Knowledge and Data Engineering, vol. 28, no. 3, pp. 701–714, 2015. View at: Publisher Site  Google Scholar
 A. Topchy, A. K. Jain, and W. Punch, “A mixture model for clustering ensembles,” in Proceedings of the 2004 SIAM International Conference on Data Mining, Lake Buena Vista, FL, USA, April 2004. View at: Publisher Site  Google Scholar
 F. Li, Y. Qian, J. Wang, and J. Liang, “Multigranulation information fusion: A DempsterShafer evidence theorybased clustering ensemble method,” Information Sciences, vol. 378, pp. 389–409, 2017. View at: Publisher Site  Google Scholar
 F. Li, Y. Qian, J. Wang, C. Dang, and L. Jing, “Clustering ensemble based on sample’s stability,” Artificial Intelligence, vol. 273, pp. 37–55, 2019. View at: Publisher Site  Google Scholar
 D. Huang, J.H. Lai, and C.D. Wang, “Robust ensemble clustering using probability trajectories,” IEEE Transactions on Knowledge and Data Engineering, vol. 28, no. 5, pp. 1312–1326, 2015. View at: Publisher Site  Google Scholar
 F. Yang, T. Li, Q. Zhou, and H. Xiao, “Cluster ensemble selection with constraints,” Neurocomputing, vol. 235, pp. 59–70, 2017. View at: Publisher Site  Google Scholar
 W. Ye, H. Wang, S. Yan, T. Li, and Y. Yang, “Nonnegative matrix factorization for clustering ensemble based on dark knowledge,” KnowledgeBased Systems, vol. 163, pp. 624–631, 2019. View at: Publisher Site  Google Scholar
 J. M. Wandeto and B. DrespLangley, “The quantization error in a selforganizing map as a contrast and colour specific indicator of singlepixel change in large random patterns,” Neural Networks, vol. 119, pp. 273–285, 2019. View at: Publisher Site  Google Scholar
 A. L. N. Fred and A. K. Jain, “Combining multiple clusterings using evidence accumulation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 6, pp. 835–850, 2005. View at: Publisher Site  Google Scholar
 T. Alqurashi and W. Wang, “Objectneighbourhood clustering ensemble method,” in Proceedings of the International Conference on Intelligent Data Engineering & Automated Learning, Salamanca, Spain, September 2014. View at: Google Scholar
 T. Alqurashi and W. Wang, “A new consensus function based on dualsimilarity measurements for clustering ensemble,” in Proceeding of the IEEE International Conference on Data Science and Advanced Analytics (DSAA), pp. 1–9, Paris, France, October 2015. View at: Google Scholar
 T. Alqurashi and W. Wang, “Clustering ensemble method,” International Journal of Machine Learning and Cybernetics, vol. 10, no. 6, pp. 1227–1246, 2019. View at: Publisher Site  Google Scholar
 S. Mimaroglu and E. Aksehirli, “DICLENS: Divisive clustering ensemble with automatic cluster number,” IEEE/ACM Transactions on Computational Biology and Bioinformatics, vol. 9, no. 2, pp. 408–420, 2012. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2020 Wenqi Hua and Lingfei Mo. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.