Research Article  Open Access
ZhiFei Lai, HuiFang Deng, "Multiscale HighLevel Feature Fusion for Histopathological Image Classification", Computational and Mathematical Methods in Medicine, vol. 2017, Article ID 7521846, 6 pages, 2017. https://doi.org/10.1155/2017/7521846
Multiscale HighLevel Feature Fusion for Histopathological Image Classification
Abstract
Histopathological image classification is one of the most important steps for disease diagnosis. We proposed a method for multiclass histopathological image classification based on deep convolutional neural network referred to as coding network. It can gain better representation for the histopathological image than only using coding network. The main process is that training a deep convolutional neural network is to extract highlevel feature and fuse two convolutional layers’ highlevel feature as multiscale highlevel feature. In order to gain better performance and high efficiency, we would employ sparse autoencoder (SAE) and principal components analysis (PCA) to reduce the dimensionality of multiscale highlevel feature. We evaluate the proposed method on a real histopathological image dataset. Our results suggest that the proposed method is effective and outperforms the coding network.
1. Introduction
Medical image classification is one of the most important steps for disease diagnosis. In this paper, we would focus on the histopathological image classification task, which is a subset of medical image classification and can provide useful hints for doctor’s disease diagnosis. Xu et al. [1] proposed a classification method based on a deep convolutional neural networks (DCNNs) to learn highlevel feature to classify epithelial and stromal tissues. In their study, it is a binary classification task and it directly uses DCNN to complete the classification task rather than multiscale features. CruzRoa et al. [2] proposed a method to automate detection of of invasive ductal carcinoma in whole slide images of breast cancer with convolutional neural networks (CNNs). In [3], the authors used stacked sparse autoencoder (SSAE) to detect nuclei on breast cancer histopathological images where SSAE can learn discriminative highlevel features. In [4], Esteva et al. proposed the method using DCNN to classify the skin cancer which can achieve dermatologistlevel diagnosis and they evaluated the results on a dataset of 129,450 clinical images. It turns out that if the dataset is big enough, training the DCNN to deal with classification task usually works very well. Many current works [5–7] directly employ the DCNN to gain highlevel feature classifying medical image but seldom incorporating the highlevel feature solved this challenge problem; that is, their model just designed different convolutional layers and maxpooling layer and connected the softmax layer in the end as classifier. However, it is pointed out that simply using DCNN to classify is not enough to gain a better performance and high efficiency especially in case of the limited histopathological image dataset. A convolutional autoencoder (CAE) based algorithm was proposed in [8], which encoded the image through CAE at first and then added the preextracted VGG features to a semisupervised CNN. Experimental results on a dataset of 2078 images indicated the proposed method can reduce the error rate of attribute and shape classification by 21.54% and 15.07%, respectively. Shu et al. [9] developed the Deep Transfer Networks (DTNs) to address the insufficient training examples challenge. They pretrained two SSAEs for text and image, respectively, and then continued to train SSAEs for shared representations called weakly shared DTNs. It proved the effectiveness of the model on NUSWIDE dataset [10]. In the approach presented in [11], CNNs were accelerated with distributed GPUS via hybrid parallelism strategy called “Wheel” which transmitted most of the parameters in one server to reduce transmit data time. In addition, it fully run each GPU to reduce the idle time.
In order to improve the efficiency and accuracy of histopathological image classification, we propose a novel method based on the DCNN referred as coding network which extract twolayer highlevel feature as multiscale highlevel feature. This method is inspired by [12] that used DCNN to predict 10000 classes. But the difference is that our method fuses the highlevel features to feed into another classifier and would put the multiscale feature into sparse autoencoder to reduce the dimensionality.
The remainder of this paper is organized as follows. We will describe the detailed procedure of extracting multiscale highlevel feature and give an effective algorithm to reduce dimensionality of multiscale highlevel feature in Section 2. In Section 3, we report our experiment in SDT dataset and the analysis of the model. Finally, we conclude the paper and give the future work in Section 4.
2. The Methodology of Proposed Model
2.1. Coding Network and Multiscale HighLevel Feature Extraction
Our coding network contains six convolutional layers which follows by normalization layer called local response normalization reference by [13]. Before feeding into the softmax layer, it contains one fullconnected layer and dropout [14] layer that includes setting to zero the output of each hidden neuron with probability 0.5. The normalization layer and dropout layer are important to coding network owing to improving the overall accuracy where it may get discriminative highlevel feature. When training the coding network, the input is a fixedsize RGB image. Table 1 shows the detailed configure of coding network. The convolutional layer employs a filter with a receptive field of , , and and a 1 pixel stride and 0 pixel padding. The pooling layer is performed over a 5 × 5 pixel window, with stride 2.

In extracting multiscale highlevel feature, coding network is intent to extract highlevel features. We extract the 6th convolutional layer and the fullconnected layer combined as multiscale highlevel features. The 6th convolutional layer is fixed to 2048 referred to as cfr, while the dimension of fullconnected layer is 256 referred to as ffr. It directly fuses the cfr and ffr into one feature vector that can continue to feed into another softmax classifier to classify the histopathological image.
2.2. The Dimensionality Reduction of Multiscale HighLevel Feature
We reduce the dimensionality of multiscale highlevel feature vector based on the following considerations:(1)The extracted multiscale highlevel feature which combines twolayer highlevel features of coding network would lead to considerable computational complexity because of their high dimensions.(2)In addition, owing to the two different highlevel features extracted from the coding network, it may inevitably bring about obvious correlation between them.
Therefore, we would use sparse autoencoder (SAE) to reduce multiscale highlevel features because sparse autoencoder could gain more discriminative features due to cut the correlation between highlevel features. We employ representation of the extracted multiscale highlevel feature. Sparse autoencoder is the threelayer neural network which is trying to train a target function , where is the output of sparse autoencoder and the second layer contains the sparse representation of original multiscale highlevel . It can be defined as follows: In (1), is the activation function that we apply here ; is the weight matrix and is the bias. It is worth noting that the dimensionality of is lower than . In addition, sparse autoencoder minimized the cost function as follows: where is the total number of pathological images. However, in order to achieve the purpose of reducing dimensionality, we have to impose the sparse constraints overall cost function. Here, it gives the penalty term KullbackLeibler (KL) divergence to cost function. It can be defined by where is the dimensionality of ; a sparsity parameter that is closed to zero; is the average activation of hidden unit that can be defined as where denotes the activation of sparse autoencoder hidden unit when the network is given a specific input . And is given by Therefore, the original cost function is substituted by controls the weight of the sparsity penalty term. Minimizing the new cost function, we could gain the sparse representation of multiscale highlevel features.
Principal components analysis (PCA) is a classical data dimensionality reduction algorithm in unsupervised feature learning area. A large amount of works [15–17] was based on PCA to solve the problem of high dimensionality, in which the feature will be decomposed as the linearly independent eigenvectors that choose the principal eigenvectors of the original feature. It is an effective method to reduce data dimensionality. Therefore, in order to demonstrate the effectiveness of SAE, it may use PCA algorithm for comparison to reduce the multiscale highlevel feature dimensionality.
3. Experiment and Evaluation
3.1. Dataset Description and Experiment Setting
A real skin biopsy image dataset called SDT which contains 6 classes of skin disease images is used to evaluate the overall performance of the proposed method. The dataset is composed of 2019 images where each image is an RGB image of size . Table 2 summarizes the 6 classes in the dataset and employs T1, T2, T3, T4, T5, and T6 to label each category. In order to efficiently reduce the overfitting problem of the proposed method, we manually enlarge the SDT dataset through extracting random image patches from the original image. Then we resize the image patches to . This not only saved the main information of medical image but also reduced the running time of our algorithm. The configuration of our experiment can be seen as follows. The dataset is divided into 3 parts: training set, validation set, and test set with a ratio 7 : 1 : 2. In addition, we trained our model on the extracted patches by 10fold crossvalidation. For the coding network, it had taken 45 epoches to get the network convergence. It trained the coding network using stochastic gradient descent with the batch size of 100 images, momentum of 0.9. Furthermore, the learning rate was initialized at 0.01, where it would adjust manually through training. The strategy is referenced by [13] which is to divide the learning rate by 10 when the validation error rate stopped improving.

3.2. Accuracy Analysis
In this section, it is necessary to compare the coding network with the multiscale highlevel features for demonstrating the effectiveness and efficiency of proposed algorithm. Furthermore, multiscale feature + SAE (MSAE) will be compared with multiscale feature + PCA (MPCA) to verify the efficiency of dimensionality reduction. Quantitative evaluation is shown in Table 3 in which the overall accuracy reaches 86.2%, 92.6%, and 95.3%, respectively. From Table 3 it can be found that MSAE owns the best algorithm accuracy that outperforms the other two algorithms. In addition, it is obvious that MSAE and MPCA can gain perform better than coding network which demonstrates the efficiency of multiscale highlevel features and the effectiveness of dimensionality reduction. All experiments are implemented in Matlab using MatConvNet package [18]. And all experiments are conducted on a computer with i56500 3.2 GHz CPU, 32 G main memory, and GTX1060 GPU.

In Table 4, it reports the experimental results which shows the comparison of algorithm accuracy in each category. We can see that no matter which algorithm is, it would gain better accuracy than overall performance in some cases. The unbalanced problem would account for this phenomenon. The more the samples of a category lie in the dataset, the better the accuracy can achieve. In addition, from Table 4 we clearly know that MSAE outperforms the coding network and MPCA in each category which confirms the effectiveness of our proposed method again.

Figure 1 shows the confusion matrix of different algorithms. In confusion matrix, the mint green cells represent the number and percentage of correct predictions made by the algorithm; the pink shaded cells exhibit the number and percentage of incorrect predictions. At the last row of confusion matrix, it gives the precision of each category, while it also gains the recall of each class in the last column of confusion matrix. The last diagonal element of confusion matrix represents the overall accuracy of algorithm. From Figure 1, MSAE can achieve better precision than MPCA and coding network in each class. The performance of our proposed model in minority class can clearly improve. In addition, it would draw a conclusion that T2 has the familiar structure with T5. Relative number samples of T2 that are predicted to T5 would account for this phenomenon. In order to better evaluate of the algorithms, we follow [19] using the receiver operating characteristic (ROC) of different classes as evaluation criteria. Figure 2 compares the ROC curve of different classes. In addition, it computes the area under the curve (AUC) to more intuitive comparison. Table 5 displays the mean AUC of different algorithms. It is obvious that the AUC of our proposed model can achieve 0.9912 which proves to be better than coding network and MPCA with 0.9617 and 0.9855, respectively.

(a) The confusion matrix of coding network
(b) The confusion matrix of MPCA
(c) The confusion matrix of MSAE
4. Conclusion
We propose a multiclass histopathological image classification method which is based on the multiscale features. This method trains the coding network to extract highlevel feature and combines one convolutional layer feature and fullconnected feature as multiscale highlevel feature. In order to solve the problem of high dimensionality and speed up the running time of the algorithm, we use SAE and PCA to reduce the dimensionality of multiscale feature. At last, we evaluate the method on the dataset SDT and the results show that MSAE outperforms PSAE and coding network. In our future work, we will train serval DCNNs to ensemble learning that can extract highlevel features from these DCNNS. Specifically, we will design one deep learning framework referred to as multiscale highlevel framework which gain the features. Meanwhile, fusing these highlevel features feeds into the classifier to classify the histopathological images.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
References
 J. Xu, X. Luo, G. Wang, H. Gilmore, and A. Madabhushi, “A deep convolutional neural Network for segmenting and classifying epithelial and stromal regions in histopathological images,” Neurocomputing, vol. 191, pp. 214–223, 2016. View at: Publisher Site  Google Scholar
 A. CruzRoa, A. Basavanhally, F. González et al., “Automatic detection of invasive ductal carcinoma in whole slide images with convolutional neural networks,” in Proceedings of the SPIE medical imaging, International Society for Optics and Photonics, vol. 9041, pp. 904103–904103, February 2014. View at: Publisher Site  Google Scholar
 J. Xu, L. Xiang, Q. Liu et al., “Stacked sparse autoencoder (SSAE) for nuclei detection on breast cancer histopathology images,” IEEE Transactions on Medical Imaging, vol. 35, no. 1, pp. 119–130, 2016. View at: Publisher Site  Google Scholar
 A. Esteva, B. Kuprel, R. A. Novoa et al., “Dermatologistlevel classification of skin cancer with deep neural networks,” Nature, vol. 542, no. 7639, pp. 115–118, 2017. View at: Publisher Site  Google Scholar
 K. Sirinukunwattana, S. E. A. Raza, Y.W. Tsang, D. R. J. Snead, I. A. Cree, and N. M. Rajpoot, “Locality sensitive deep learning for detection and classification of nuclei in routine colon cancer histology images,” IEEE Transactions on Medical Imaging, vol. 35, no. 5, pp. 1196–1206, 2016. View at: Publisher Site  Google Scholar
 T. H. Vu, H. S. Mousavi, V. Monga, G. Rao, and U. K. A. Rao, “Histopathological image classification using discriminative featureoriented dictionary learning,” IEEE Transactions on Medical Imaging, vol. 35, no. 3, pp. 738–751, 2016. View at: Publisher Site  Google Scholar
 G. Litjens, C. I. Sánchez, and N. Timofeeva, “Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis,” Scientific Reports, vol. 6, Article ID 26286, 2016. View at: Publisher Site  Google Scholar
 V. Murthy, L. Hou, D. Samaras, T. M. Kurc, and J. H. Saltz, “Centerfocusing multitask cnn with injected features for classification of glioma nuclear images,” in Proceedings of the 2017 IEEE Winter Conference on Applications of Computer Vision (WACV '17), pp. 834–841, Santa Rosa, CA, USA, March 2017. View at: Publisher Site  Google Scholar
 X. Shu, G.J. Qi, J. Tang, and J. Wang, “WeaklyShared deep transfer networks for heterogeneousdomain knowledge propagation,” in Proceedings of the 23rd ACM International Conference on Multimedia (MM '15), pp. 35–44, October 2015. View at: Publisher Site  Google Scholar
 T.S. Chua, J. Tang, R. Hong, H. Li, Z. Luo, and Y. Zheng, “NUSWIDE: A realworld web image database from national university of singapore,” in Proceedings of ACM International Conference on Image and Video Retrieval (CIVR '09), pp. 368–375, July 2009. View at: Publisher Site  Google Scholar
 X. Y. Du, J. Tang, Z. Li, and Z. Qin, “Wheel: Accelerating cnns with distributed gpus via hybrid parallelism and alternate strategy,” in Proceedings of the the 2017 ACM, pp. 393–401, Mountain View, California, USA, October 2017. View at: Publisher Site  Google Scholar
 Y. Sun, X. Wang, and X. Tang, “Deep learning face representation from predicting 10,000 classes,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '14), pp. 1891–1898, Columbus, OH, USA, June 2014. View at: Publisher Site  Google Scholar
 A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proceedings of the 26th Annual Conference on Neural Information Processing Systems (NIPS '12), pp. 1097–1105, Lake Tahoe, NV, USA, December 2012. View at: Google Scholar
 G. E. Hinton, N. Srivastava, A. Krizhevsky, I. Sutskever, and R. R. Salakhutdinov, “Improving neural networks by preventing coadaptation of feature detectors,” https://arxiv.org/abs/1207.0580. View at: Google Scholar
 S. Stewart, M. A. Ivy, and E. V. Anslyn, “The use of principal component analysis and discriminant analysis in differential sensing routines,” Chemical Society Reviews, vol. 43, no. 1, pp. 70–84, 2014. View at: Publisher Site  Google Scholar
 D. Giri, U. Rajendra Acharya, R. J. Martis et al., “Automated diagnosis of Coronary Artery Disease affected patients using LDA, PCA, ICA and Discrete Wavelet Transform,” KnowledgeBased Systems, vol. 37, pp. 274–282, 2013. View at: Publisher Site  Google Scholar
 U. Demšar, P. Harris, C. Brunsdon, A. S. Fotheringham, and S. McLoone, “Principal component analysis on spatial data: an overview,” Annals of the Association of American Geographers, vol. 103, no. 1, pp. 106–128, 2013. View at: Publisher Site  Google Scholar
 A. Vedaldi and K. Lenc, “MatConvNet: convolutional neural networks for matlab,” in Proceedings of the 23rd ACM International Conference on Multimedia, pp. 689–692, Brisbane, Australia, October 2015. View at: Publisher Site  Google Scholar
 M. Buda, A. Maki, and M. A. Mazurowski, “A systematic study of the class imbalance problem in convolutional neural networks”. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2017 ZhiFei Lai and HuiFang Deng. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.