Table of Contents Author Guidelines Submit a Manuscript
Computational Intelligence and Neuroscience
Volume 2017 (2017), Article ID 3792805, 12 pages
https://doi.org/10.1155/2017/3792805
Research Article

Image Classification Using Biomimetic Pattern Recognition with Convolutional Neural Networks Features

1College of IOT Engineering, Hohai University, Changzhou 213022, China
2Key Laboratory of Sensor Networks and Environmental Sensing, Hohai University, Changzhou 213022, China

Correspondence should be addressed to Qingwu Li

Received 1 August 2016; Revised 28 November 2016; Accepted 15 January 2017; Published 16 February 2017

Academic Editor: Leonardo Franco

Copyright © 2017 Liangji Zhou et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

As a typical deep-learning model, Convolutional Neural Networks (CNNs) can be exploited to automatically extract features from images using the hierarchical structure inspired by mammalian visual system. For image classification tasks, traditional CNN models employ the softmax function for classification. However, owing to the limited capacity of the softmax function, there are some shortcomings of traditional CNN models in image classification. To deal with this problem, a new method combining Biomimetic Pattern Recognition (BPR) with CNNs is proposed for image classification. BPR performs class recognition by a union of geometrical cover sets in a high-dimensional feature space and therefore can overcome some disadvantages of traditional pattern recognition. The proposed method is evaluated on three famous image classification benchmarks, that is, MNIST, AR, and CIFAR-10. The classification accuracies of the proposed method for the three datasets are 99.01%, 98.40%, and 87.11%, respectively, which are much higher in comparison with the other four methods in most cases.

1. Introduction

Image classification and recognition is a sophisticated task for machine, and it has been a hot issue in the field of Artificial Intelligence (AI) all the time. Feature extraction from an image is a significant step in automatic image classification. To effectively represent the image, many approaches have been proposed, and these approaches can be roughly categorized as hand-crafted features and machine learned features. The most representative hand-crafted features are scale-invariant feature transform (SIFT) [1] and Histogram of Oriented Gradient (HOG) [2]. These features are especially useful for the image classification on small-scale datasets. However, it is a too difficult problem to find proper features from images in the case of large-scale dataset. Moreover, the hand-crafted features are usually low-level features without enough mid-level and high-level information, which hinders the performance [3].

Over the last few years, Deep Neural Networks (DNNs) have achieved state-of-the-art performances in a wide range of areas [47]. Inspired by the mammalian visual system, Deep Convolutional Neural Networks (DCNNs) have become the most suitable architectures for many computer vision tasks [8]. CNNs, as generic feature extractors, have been continuously improving the image classification accuracy, avoiding the traditional hand-crafted feature extraction techniques in image classification problems. The features learned from CNNs are not designed by human engineers, but from data using a general-purpose learning procedure [9]. Because both hand-crafted features and machine learned features have their advantages, the reasonable combination of these two methods is becoming a hotspot recently [1012].

A typical architecture of CNNs usually contains many layers to automatically extract useful image features and exploit the softmax function (also known as multinomial logistic regression) for classification [13, 14]. However, the softmax classifier often shows a low prediction performance [15]. Moreover, a higher precision gained by CNNs also means a deeper structure, more learning parameters, and larger amount of training data, leading to a cost of increased training complexity. In addition, because overly increasing depth can harm the accuracy, even if the width/filter sizes are unchanged, a deeper structure does not always guarantee a better result, which has been validated by many experiments [16].

To tackle the problem mentioned above, some viable research has already been proposed. If a well-performed classifier was added behind the CNN, the classification accuracy will be improved in some degree, and this is exactly the starting point of CNN-SVM. CNN-SVM is a combination of CNN and SVM [17], which take CNN as a trainable feature extractor and SVM as a classifier. Firstly, CNN is utilized to learn feature vectors from the image data. Then the learned vector representations are fed to a SVM classifier as features for image classification. It should be noted that, in the whole process, CNN and SVM are trained separately to get a better result [18]. The results provided by a combination of CNNs and SVM show higher accuracy rate compared with alone use of CNNs or SVM. The running time of the combination method is significantly lower than that of SVM. Inspired by its success, this kind of combination is also adopted by other studies [19, 20].

Biomimetic Pattern Recognition (BPR) [21] is a new model of pattern recognition, which is based on “matter cognition” instead of “matter classification.” This new model is much closer to the recognition function of human beings, who cognize matters class by class, than traditional statistical pattern recognition using “optimal separating” as its main principle. In the BPR, “cognizing” one class of matters is essential to analyzing and “cognizing” the shape of infinite points set made up of all samples of the same class. In a mathematical work [22] written by pre-Soviet academician, Aleksandrov pointed out “The concept of topological space is very general, and the science about topological space—topology—is the most general mathematical branch about continuity.” The mathematical tool of the BPR is just the method to analyze the manifold in point set topology. Therefore, the BPR is also called the Topological Pattern Recognition (TPR).

In this paper, a new method that combines CNNs with BPR is proposed to reduce the complexity of training networks and to improve the performance of classification. Because CNNs represent an inspiration of the cognitive neuroscience while BPR implies the cognitive psychology, it is reasonable to combine them together in the framework of cognitive science. In our framework, CNNs are used to automatically learn feature vectors from raw images, and then the learned feature vectors are projected into high-dimensional space to be covered by BPR classifier. Such a combination is expected to combine the advantages of CNNs on feature representation and BPR on classification. Meanwhile, an adaptive technique is adopted to tackle the problem of setting coverage radius in BPR classifier. Evaluations on MNIST, AR, and CIFAR-10 datasets show that the combined model excels the classic CNN, CNN combined with SVM, and Principal Component Analysis (PCA) combined with BPR models in classification accuracy.

The rest of the paper is organized as follows. In Section 2, we give some brief introduction of CNNs and BPR. In Section 3, the proposed CNN-BPR model is presented. In Section 4, three different benchmark datasets are used to validate the superiority of the proposed model. Finally, Section 5 gives the conclusion.

2. Related Works

2.1. Convolutional Neural Networks

The idea of CNNs was firstly proposed in [23] by Fukushima, developed in [24] by LeCun et al., and improved in [25, 26] by Simard, Cireşan, and others. GPU acceleration hardware has facilitated development of deep CNN (DCNN), which includes a deeper architecture with additional convolutional layers.

Typical CNNs are composed of convolutional layer, pooling layer, and fully connected layer. A CNN consists of one or more pairs of convolution and pooling layers and finally ends with fully connected neural networks. Convolutional layers alternate with max-pooling layers mimicking the nature of complex and simple cells in mammalian visual cortex [27]. A typical convolutional network architecture is shown in Figure 1.

Figure 1: A typical convolutional network architecture.

The 2D raw pixels of the image can be accepted as the input of CNNs directly. The image is then convolved with multiple learned kernels using shared weights. A convolutional layer is parametrized by the number of maps, the size of the maps, and kernel sizes. Each layer has M maps of equal size (). A kernel of size () is shifted over the valid region of the input image. Each map in layer l is connected to all maps in layer . Neurons of a given map share their weights but have different input fields.

Next, the pooling layer reduces the size of the image while trying to maintain the contained information. The purpose of the pooling layers is to achieve spatial invariance by reducing the resolution of the feature maps. The output of the pooling layer is given by the maximum, mean, or stochastic activation, corresponding to max-pooling, mean-pooling, or stochastic-pooling, over nonoverlapping rectangular regions of size (). Pooling creates position invariance over larger local regions and downsamples the input image by a factor of and along each direction. It turned out to be that max-pooling leads to faster convergence rate by selecting superior invariant features, which improves generalization performance [28].

Convolutional layer and pooling layer compose the feature extraction part. Afterwards, the extracted features are weighted and combined in one or more fully connected layers. This represents the classification part of the convolutional network. These layers are similar to the layers in a standard Multilayer Perceptron (MLP), where the outputs of all neurons in layer are connected to every neuron in layer l.

Finally, there exists one output neuron for each object category in the output layer. The output layer has one neuron per class in the classification task. A softmax activation function is used; thus each neuron’s output represents the posterior class probability.

For a classification problem, it is standard for a CNN to use softmax or 1-of- encoding at the top. Let pi specify a discrete probability distribution, where , and , h is the activation of the penultimate layer nodes, W is the weight vector between the penultimate layer and the softmax layer, the total input into a softmax layer, given by a, isthen we haveand the predicted class would be

CNNs are usually trained with a variant of the gradient-based backpropagation method [29]. All training patterns along with the expected outputs are fed into the network. Afterwards the network error (the difference between the actual and expected output) is backpropagated through the network and used to compute the gradient of the network error with respect to the weights. This gradient is then used to update the weight values according to a specific rule (e.g., stochastic, momentum, etc.) [30].

2.2. Biomimetic Pattern Recognition

The “biomimetic” emphasizes that the viewpoint of the function and mathematical model of pattern recognition is the concept of “cognition,” which is much closer to the function of human beings. An important and essential focus of attention in BPR is the principle of homology-continuity (PHC) [31].

Theorem 1. In the feature space , suppose that set is a point set including all samples in class . If is given, there must be set

“All useful information is included in the training set”—it is the basic of the traditional pattern recognition, but the theorem of PHC is beyond this hypothesis. The traditional pattern recognition is completely based on the separation of different samples in feature space because of the consideration that there is no a priori knowledge among the same sample points. However, “Universal Relation” is the objective law in nature, and it is followed by BPR, which makes full use of the law to improve the cognitive abilities of things. Figure 2 shows the differences of the BPR and the traditional pattern recognition.

Figure 2: The schematic diagram of the difference of BP, RBF, and BPR.

In Figure 2, the circles represent the samples to be recognized; the squares and triangles represent samples to be distinguished from circles. These small signs represent an idealized distribution of the sample points in feature space, and the polygonal line denotes the classification boundaries of traditional backpropagation (BP) networks, big circle denotes radial-basis function (RBF) networks (which is the same as the template matching), and the sausage-like shape represents “cognition” manner of BPR. The specific differences between BPR and the traditional pattern recognition are described in Table 1.

Table 1: Comparison of traditional pattern recognition and BPR.

3. The Proposed Model

In this section, we present a CNN-BPR combined model for image classification. The system framework is shown in Figure 3. Firstly, an automatic feature extraction is proposed by using CNN. Secondly, BPR is adopted as the classifier exploiting the features extracted from the previous module, CNN. To obtain the optimal coverage, an adaptive method is used to compute an appropriate coverage threshold for each class.

Figure 3: Flowchart of the proposed CNN-BPR recognition method.
3.1. CNN-Based Feature Extraction

A CNN architecture with alternating convolutional and max-pooling layers is used here. Each node in the output layer corresponds to one character class. After CNN training, only the parameters of the fully connected layer are left to extract the final feature vector which will be fed to the BPR classifier. The CNN-based feature extractor that is irrelevant to the number of the character classes can be very compact. To extract the CNN-based feature, network training is used first. The training of CNN is composed of two main procedures, namely, forward propagation and backpropagation [32, 33].

3.1.1. Forward Propagation of CNN

Assuming that layer is a convolution layer and layer is a subsampling layer or an input layer, the feature map j of layer l is calculated as follows:where represents a selection of input maps, “” indicates the convolution computation; the essence of which is to convolve the convolution kernel on all the associated feature maps of the layer ; then sum them, together with the bias as the input of the activation function and finally get the output of convolution layer .

A pooling layer produces downsampled versions of the input maps. If there are input maps, then there will be exactly output maps, although the output maps will be smaller. More formally,where denotes the -th subsampling layer, and is pooling function which can be max-pooling, mean-pooling, or stochastic-pooling. Each output map is given its own multiplicative bias and an additive bias .

The procedure of forward propagation is composed of convolution and pooling alternately, and for full connection layer, all previous output maps are convoluted through each convolution kernel in this layer.

3.1.2. Backpropagation of CNN

The training procedure of the CNN is the same as BP model. In the following, we define a squared-error function as the loss function . For a multiclass problem with training examples, can be given aswhere represents the -th training example, is its label, and is the output.

However, different from the single structure of BP, there are differences in the training procedure for each layer of CNN. Here, we briefly present how to train parameters and compute the gradients for different types of layers.

(1) Convolution Layers. The backpropagation “error” through the network can be considered as “sensitivity.” Assuming that each convolution layer is followed by a subsampling layer , the residuals in the BP algorithm are equal to the weighted sums of the weights and residuals of all the nodes connected to the layer and then multiplied by the derivative value of this point. The next layer of the convolution layer is the subsampling layer , using one-to-one nonoverlapping sampling, so the residual calculation is simpler. The residual of the feature map of the layer is calculated as follows:where “” denotes element-wise multiplication and denotes an upsampling operation the purpose of which is to extend the size of layer l + 1 to the size of layer l, and it is also defined as the Kronecker product:

Now given sensitivities map, the bias gradient is computed by simply summing over all the entries in .Finally, to compute the gradients of the kernel weights, we sum the gradients for a given weight over all the connections that mention this weight:where is the patch in that was multiplied element-wise by during convolution in order to compute the element at (u, v) in the output convolution map .

(2) Subsampling Layers. For subsampling layer, the parameters of and b are needed to learn. It is also worth noting that the proposed subsampling layer is following connected by convolution layer. The sensitivity of subsampling layer l is defined aswhere is the sensitivity of convolution layer . The additive bias is again the sum over the elements of the sensitivity map and can be rewritten (10) as

The multiplicative bias involves the original downsampled map computed at the current layer during the feedforward pass. To compute the parameter , we defineThen the gradient for is given by

3.2. Classification Based on BPR

The process of BPR for classification mainly consists of constructing complex geometry coverage in high-dimensional space and discrimination based on minimum distance. Some associated theories will be introduced in the following context briefly.

3.2.1. Theory of Complex Geometry Coverage in High-Dimensional Space

The BPR uses topological high-dimensional manifold theory as a mathematical tool and realizable method, which is also the reason why it is called TPR. High-dimensional manifold theory is used by BPR to study the topological properties of the similar samples in feature space. The method—“Complex Geometry Coverage (CGC) in high-dimensional space”—is used to study the samples’ distribution in feature space and to give reasonable cover, so the samples can be cognized [34].

To form the CGC in high-dimensional space, the Multiweighted Neuron is specifically used to cover the Simplex one by one. The definitions of Simplex and Multiweighted Neuron are as follows.

Definition 2. Suppose are some irrelevant points in -dimensional Euclidean space , which is to say the vectors , are linearly independent, and then point set is a -dimensional Simplex with as its vertexes.
Simply, respectively, line segment, plane triangle, and tetrahedron can be regarded as a 1-dimensional, 2-dimensional, and 3-dimensional Simplex in the Euclidean space.

Definition 3. Suppose is a polyhedron in feature space and the distance between x and V meetsIf meetsthen is called a coverage of polyhedron.
When V is a line segment, a plane triangle, or a tetrahedron, the Multiweighted Neuron U is a Hyper Sausage Neuron (HSN), a ψ3 neuron, or a three degrees of freedom’ (DOF) neuron, respectively. Figure 4 shows a schematic representation of these types of neuron model.

Figure 4: Some kinds of neural model of BPR.
3.2.2. Recognition Algorithm Based on a Triangle Coverage

Suppose is a training set including N classes, and is the kth class which contains N sample points; here are the steps to construct CGC [35]:

Step 1. Calculate the distance ρ between any of two points in , and find two points P11 and P12 from the training set , and let , where . Then find out a third point , where is the nearest point away from P11 and P12 but not in line of them. Connect these three points to constitute the first plane triangle T1, which can be covered with a neuron and the coverage space θ1 iswhere indicates the distance between X and T1, Th is the threshold, and the rest set .

Step 2. Judge whether or not points in S are in the coverage of θ1; if it is true, then exclude these points from S and let . Find out another point P21 from set to make the minimum sum of distance from , , and . Rename the two points of as P22 and P23, which are the nearest two to , and then construct the second plane triangle T2. Likewise, T2 is covered with a ψ3 neuron generating the coverage space θ2, and the rest set .

Step 3. Find out other point as Step 2 does, marked as Pi1, and the two nearest points are marked as Pi2 and Pi3. construct the plane triangle and further make the coverage θi, and the rest set .

Step 4. Judge whether set is empty; if it is true then end the construction; else repeat Step 3.
After the steps above, the eventual coverage of class k is the union of all coverage of neurons, which is

The basic recognition process is to judge which coverage the test sample would be covered with. Full coverage of training samples in different classes will inevitably result in overlapping space. A test sample might fall into none or overlapped coverage; then it belongs to the one closest to it. Therefore, the smallest distances need to be calculated between the test sample and each coverage. Let be the distance between sample and coverage space of class ; we havewhere is the number of neurons in class , is the total of classes, and is the distance between a sample to be recognized and the coverage of neuron in class . Then discrimination function is defined asThe pseudocode of CNN-BPR is given in Pseudocode 1.

Pseudocode 1: Pseudocode of the CNN-BPR algorithm.

4. Experiments and Discussions

To validate the proposed algorithm, three different datasets are used, namely, MNIST, AR, and CIFAR-10. Each dataset will be introduced briefly in the following paragraphs. For each group of experiments, the performance among CNN, CNN-SVM, PCA-BPR [36], HOG-SVM [37], and the proposed method is compared in the condition of different amount of training data. For CNN, we use the code downloaded from https://github.com/rasmusbergpalm/DeepLearnToolbox, and for the other three compared methods, they are all combined algorithms; we reimplement them according to the specific steps from their papers. The results demonstrate the validity of the proposed method.

4.1. Experiments on MNIST

The MNIST [38] dataset contains grayscale images of handwritten digits. Some images of the MNIST dataset are shown in Figure 5. It possesses ten different categories, namely, one for each digit from zero to nine. Each image has a fixed size of 28×28 pixels. The digits are centered inside the image and normalized in size. Totally, MNIST contains 70,000 images, including 60,000 training and 10,000 test images.

Figure 5: 100 randomly selected images from MNIST.

For MNIST, the input layer is followed by a convolutional layer C1 with 5 × 5 filters and 10 maps of size 24 × 24. The subsequent max-pooling layer reduces the previous layer size to 12 × 12 by 2 × 2 filters. C3 also employs 5 × 5 filters but has 12 maps with dimensions of 8 × 8 pixels. with 2 × 2 pooling windows yields 4 × 4 feature maps that are fully connected to 100 hidden neurons. These 100 hidden neurons are finally connected to the 10 output units. The structure of CNN can be briefly described as 28 × 28Input-10C5-MP2-12C5-MP2-10Output.

Because the CNN is employed as an automatic feature extractor and BPR as a classifier in our proposed method, after training for 20 epochs with a learning rate of 0.1, the fully connected layer with dimensions of 100 is projected into the feature space. Then, a series of ψ3 neurons are used to cover these feature points class by class.

CNN, CNN-SVM, PCA-BPR, and the proposed method are tested on the dataset. For CNN-SVM, we employ the 100 dimensional fully connected neurons above as the input of SVM, which is from LIBSVM with RBF kernel function. For PCA-BPR, same dimensional size of features are extracted from the top-100 principal components, and then ψ3 neurons are used to cover these feature points class by class. The experimental results are shown in Table 2. Moreover, in order to facilitate comparison with other methods, we set different numbers of training samples, which are 500, 1000, 6000, 10,000, and 60,000. Figure 6 shows the comparison result.

Table 2: Classification accuracy of different methods on MNIST.
Figure 6: Classification accuracies versus number of training samples for MNIST.
4.2. Experiments on AR

The AR database consists of over 3,200 frontal images of 70 men and 56 women, and there are 26 images of each individual [39]. The faces in AR contain variations such as illumination change, expressions, and facial disguises (i.e., sunglasses or scarf). We randomly selected 100 subjects (50 male and 50 female, 2,600 face images in total) in the experiments, and the images are cropped with dimension 165 × 120. For each individual, we select five images, totaled 500 face images for testing, and set different numbers of training samples by the rest images, which are 500, 800, 1,600, and 2,100. Some face images are shown in Figure 7.

Figure 7: Sample images in AR dataset.

The architecture of CNN is represented as 165 × 120Input-20C5-MP4-50C5-MP2-80C3-MP2-120FC-100Output, training 50 epochs with a learning rate of 0.01. Then the 120-dimensional feature points are covered by 3 neurons. The experimental results are shown in Table 3 and Figure 8.

Table 3: Classification accuracy of different methods on AR.
Figure 8: Classification accuracies versus number of training samples for AR.
4.3. Experiments on CIFAR-10

CIFAR-10 is a dataset of natural RGB images of 32 × 32 pixels [40]. It contains 10 classes with 50,000 training images and 10,000 test images. All of these images have different backgrounds with different light sources. Objects in the image are not restricted to the one at center, and these objects have different sizes that range in orders of magnitude. Some images of CIFAR-10 are shown in Figure 9.

Figure 9: Sample images in CIFAR-10 dataset.

Because of RGB input images, there would be three channels in each filter for the first convolutional layer, which means the size of the filter would be and three-dimensional convoluted with the input image, resulting in 12 maps of size 30 × 30 in layer . The following structure is -MP2-16C3-MP2-120FC-10Output, training 40 epochs with a learning rate of 0.1. Then the 120-dimensional feature points are covered by ψ3 neurons. The experimental results are shown in Table 4 and Figure 10.

Table 4: Classification accuracy of different methods on CIFAR-10.
Figure 10: Classification accuracies versus number of training samples for CIFAR-10.

From the above experiments, it can be seen that the CNN-BPR generally outperforms the other four methods. In the condition of the maximum training datasets, the CNN-SVM shows 0.17%, 2%, and 1.9% improvements compared to CNN, respectively, and CNN-BPR shows generally higher improvements of 0.12%, 3.4%, and 3.38% compared to CNN.

It can also be seen that HOG and BPR perform much better than the other methods in the case of small-sized homogeneous datasets, while with the increase of training samples, CNN-SVM surpasses the PCA-BPR, which means that CNN can better represent the feature than HOG and PCA do in the case of large-scale heterogeneous datasets.

5. Conclusion

In this paper, a CNN-BPR combined model for image classification is proposed. The proposed model treats CNN as a feature extractor, which can automatically learn the feature representation. BPR is adept in providing an accurate classifier. The results in terms of accuracy on the datasets of MNIST, AR, and CIFAR-10 show that the proposed method generally outperforms the other methods, which verify the effectiveness of the CNN-BPR combined image classification model.

Benefited from the unified framework of cognitive science, the combination of CNN and BPR represents a better performance than other methods. In the future, more choices of classification methods inspired by biology will be researched and compared in order to determine the best CNN-based framework for the image classification task.

Competing Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported by the Nation Nature Science Foundation of China (no. 41306089), the Science and Technology Program of Jiangsu Province (no. BY2014041), the Innovation Program for Graduate Students of Jiangsu Province (no. 2016B48514), and the Science and Technology Support Program of Changzhou (no. CE20150068).

References

  1. D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004. View at Publisher · View at Google Scholar · View at Scopus
  2. N. Dalal and B. Triggs, “Histograms of oriented gradients for human detection,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '05), vol. 2, San Diego, Calif, USA, June 2005. View at Publisher · View at Google Scholar · View at Scopus
  3. J. Yim, J. Ju, H. Jung, and J. Kim, “Image classification using convolutional neural networks with multi-stage feature,” Advances in Intelligent Systems and Computing, vol. 345, pp. 587–594, 2015. View at Publisher · View at Google Scholar · View at Scopus
  4. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “ImageNet classification with deep convolutional neural networks,” in Proceedings of the 26th Annual Conference on Neural Information Processing Systems (NIPS '12), pp. 1097–1105, Lake Tahoe, Nev, USA, December 2012. View at Scopus
  5. G. Hinton, L. Deng, D. Yu et al., “Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups,” IEEE Signal Processing Magazine, vol. 29, no. 6, pp. 82–97, 2012. View at Publisher · View at Google Scholar · View at Scopus
  6. M. K. K. Leung, H. Y. Xiong, L. J. Lee, and B. J. Frey, “Deep learning of the tissue-regulated splicing code,” Bioinformatics, vol. 30, no. 12, pp. I121–I129, 2014. View at Publisher · View at Google Scholar · View at Scopus
  7. A. Bordes, S. Chopra, and J. Weston, “Question answering with subgraph embeddings,” in Proceedings of the Conference on Empirical Methods in Natural Language Processing (EMNLP '14), pp. 615–620, Doha, Qatar, October 2014. View at Scopus
  8. W. Hu, Y. Huang, L. Wei, F. Zhang, and H. Li, “Deep convolutional neural networks for hyperspectral image classification,” Journal of Sensors, vol. 2015, Article ID 258619, 12 pages, 2015. View at Publisher · View at Google Scholar · View at Scopus
  9. Y. Lecun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015. View at Publisher · View at Google Scholar · View at Scopus
  10. L. Jin, S. Gao, Z. Li, and J. Tang, “Hand-crafted features or machine learnt features? together they improve RGB-D object recognition,” in Proceedings of the 16th IEEE International Symposium on Multimedia (ISM '14), Taichung, Taiwan, December 2014. View at Publisher · View at Google Scholar · View at Scopus
  11. G. Antipov, S.-A. Berrani, N. Ruchaud, and J.-L. Dugelay, “Learned vs. hand-crafted features for pedestrian gender recognition,” in Proceedings of the 23rd ACM International Conference on Multimedia (MM '15), pp. 1263–1266, Brisbane, Australia, October 2015. View at Publisher · View at Google Scholar · View at Scopus
  12. W. Li, S. Manivannan, S. Akbar, J. Zhang, E. Trucco, and S. J. McKenna, “Gland segmentation in colon histology images using hand-crafted features and convolutional neural networks,” in Proceedings of the IEEE 13th International Symposium on Biomedical Imaging: From Nano to Macro (ISBI '16), pp. 1405–1408, Prague, Czech Republic, April 2016. View at Publisher · View at Google Scholar · View at Scopus
  13. C. Szegedy, W. Liu, Y. Jia et al., “Going deeper with convolutions,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '15), pp. 1–9, IEEE, Boston, Mass, USA, June 2015. View at Publisher · View at Google Scholar · View at Scopus
  14. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” 2014, https://arxiv.org/abs/1409.1556.
  15. Y. Tang, “Deep learning using linear support vector machines,” https://arxiv.org/abs/1306.0239.
  16. K. He and J. Sun, “Convolutional neural networks at constrained time cost,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR '15), June 2015. View at Publisher · View at Google Scholar · View at Scopus
  17. I. Wichakam and P. Vateekul, “Combining deep convolutional networks and SVMs for mass detection on digital mammograms,” in Proceedings of the 8th International Conference on Knowledge and Smart Technology (KST '16), vol. 2016, pp. 239–244, Chiangmai, Thailand, February 2016. View at Publisher · View at Google Scholar · View at Scopus
  18. F. J. Huang and Y. LeCun, “Large-scale learning with SVM and convolutional nets for generic object categorization,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR '06), pp. 284–291, New York, NY, USA, June 2006. View at Publisher · View at Google Scholar · View at Scopus
  19. X.-X. Niu and C. Y. Suen, “A novel hybrid CNN-SVM classifier for recognizing handwritten digits,” Pattern Recognition, vol. 45, no. 4, pp. 1318–1325, 2012. View at Publisher · View at Google Scholar · View at Scopus
  20. S. Bianco, M. Buzzelli, D. Mazzini, and R. Schettini, “Logo recognition using CNN features,” in Image Analysis and Processing—ICIAP 2015, vol. 9280 of Lecture Notes in Computer Science, pp. 438–448, Springer, Cham, Switzerland, 2015. View at Publisher · View at Google Scholar · View at MathSciNet · View at Scopus
  21. S. Wang and J. Lai, “A more complex neuron in biomimetic pattern recognition,” in Proceedings of the International Conference on Neural Networks and Brain Proceedings (ICNNB '05), vol. 3, Beijing, China, October 2005. View at Scopus
  22. A. D. Aleksandrov, Mathematics, Its Essence, Methods and Role, USSR Academy of Sciences, Moscow, Russia, 1956.
  23. K. Fukushima, “Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position,” Biological Cybernetics, vol. 36, no. 4, pp. 193–202, 1980. View at Publisher · View at Google Scholar · View at Scopus
  24. Y. LeCun, L. Bottou, Y. Bengio, and P. Haffner, “Gradient-based learning applied to document recognition,” Proceedings of the IEEE, vol. 86, no. 11, pp. 2278–2324, 1998. View at Publisher · View at Google Scholar · View at Scopus
  25. P. Y. Simard, D. Steinkraus, and J. C. Platt, “Best practices for convolutional neural networks applied to visual document analysis,” in Proceedings of the 7th International Conference on Document Analysis and Recognition (ICDAR '03), pp. 958–963, August 2003. View at Publisher · View at Google Scholar · View at Scopus
  26. D. C. Cireşan, U. Meier, J. Masci, L. M. Gambardella, and J. Schmidhuber, “Flexible, high performance convolutional neural networks for image classification,” in Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI '11), vol. 22, pp. 1237–1242, July 2011.
  27. D. H. Hubel and T. N. Wiesel, “Receptive fields and functional architecture of monkey striate cortex,” The Journal of Physiology, vol. 195, no. 1, pp. 215–243, 1968. View at Publisher · View at Google Scholar · View at Scopus
  28. D. Scherer, A. Müller, and S. Behnke, “Evaluation of pooling operations in convolutional architectures for object recognition,” in Proceedings of the International Conference on Artificial Neural Networks (ICANN '10), vol. 6354, no. 3, pp. 92–101, Thessaloniki, Greece, September 2011.
  29. Y. Lecun, L. Bottou, G. B. Orr, and K.-R. Müller, “Efficient backprop,” Neural Networks: Tricks of the Trade, vol. 7700, pp. 9–48, 2012. View at Google Scholar
  30. D. Strigl, K. Kofler, and S. Podlipnig, “Performance and scalability of GPU-based convolutional neural networks,” in Proceedings of the 18th Euromicro Conference on Parallel, Distributed and Network-Based Processing (PDP '10), pp. 317–324, Pisa, Italy, February 2010. View at Publisher · View at Google Scholar · View at Scopus
  31. Y. Ren, Y. Wu, and Y. Ge, “A co-training algorithm for EEG classification with biomimetic pattern recognition and sparse representation,” Neurocomputing, vol. 137, pp. 212–222, 2014. View at Publisher · View at Google Scholar · View at Scopus
  32. J.-J. Seo, H.-I. Kim, and Y. M. Ro, “Pose-robust and discriminative feature representation by multi-task deep learning for multi-view face recognition,” in Proceedings of the 17th IEEE International Symposium on Multimedia (ISM '15), pp. 166–171, Miami, Fla, USA, December 2015. View at Publisher · View at Google Scholar · View at Scopus
  33. G. Jingyu, J. Yang, J. Zhang, and M. Li, “Natural scene recognition based on Convolutional Neural Networks and Deep Boltzmannn Machines,” in Proceedings of the IEEE International Conference on Mechatronics and Automation (ICMA '15), vol. 2015, pp. 2369–2374, 2015. View at Publisher · View at Google Scholar
  34. Y. Zhai, J. Li, J. Gan, and Y. Xu, “A novel SAR image recognition algorithm with rejection mode via biomimetic pattern recognition,” Journal of Information and Computational Science, vol. 10, no. 11, pp. 3363–3374, 2013. View at Publisher · View at Google Scholar · View at Scopus
  35. W. Cao, H. Feng, L. Hu, and T. He, “Space target recognition based on biomimetic pattern recognition,” in Proceedings of the 1st International Workshop on Database Technology and Applications (DBTA '09), April 2009. View at Publisher · View at Google Scholar · View at Scopus
  36. J.-Y. Zeng, J.-Y. Gan, and Y.-K. Zhai, “A novel partially occluded face recognition method based on biomimetic pattern recognition,” in Proceedings of the International Conference on Wavelet Analysis and Pattern Recognition (ICWAPR '12), pp. 175–179, Xian, China, July 2012. View at Publisher · View at Google Scholar · View at Scopus
  37. C. Yao, F. Wu, H.-J. Chen, X.-L. Hao, and Y. Shen, “Traffic sign recognition using HOG-SVM and grid search,” in Proceedings of the 12th IEEE International Conference on Signal Processing (ICSP '14), pp. 962–965, IEEE, Hangzhou, China, October 2014. View at Publisher · View at Google Scholar · View at Scopus
  38. The MNIST database, http://yann.lecun.com/exdb/mnist/index.html.
  39. “The AR face database,” http://www2.ece.ohio-state.edu/~aleix/ARdatabase.html.
  40. “The CIFAR-10 dataset,” http://www.cs.utoronto.ca/~kriz/cifar.html.