Decentralization and Security Issues in Blockchain Enabled Internet of Things
View this Special IssueResearch Article  Open Access
Mengxi Xu, Yingshu Lu, Xiaobin Wu, "Annular Spatial Pyramid Mapping and Feature FusionBased Image Coding Representation and Classification", Wireless Communications and Mobile Computing, vol. 2020, Article ID 8838454, 9 pages, 2020. https://doi.org/10.1155/2020/8838454
Annular Spatial Pyramid Mapping and Feature FusionBased Image Coding Representation and Classification
Abstract
Conventional image classification models commonly adopt a single feature vector to represent informative contents. However, a single image feature system can hardly extract the entirety of the information contained in images, and traditional encoding methods have a large loss of feature information. Aiming to solve this problem, this paper proposes a feature fusionbased image classification model. This model combines the principal component analysis (PCA) algorithm, processed scale invariant feature transform (PSIFT) and color naming (CN) features to generate mutually independent image representation factors. At the encoding stage of the scaleinvariant feature transform (SIFT) feature, the bagofvisualword model (BOVW) is used for feature reconstruction. Simultaneously, in order to introduce the spatial information to our extracted features, the rotation invariant spatial pyramid mapping method is introduced for the PSIFT and CN feature division and representation. At the stage of feature fusion, we adopt a support vector machine with two kernels (SVM2K) algorithm, which divides the training process into two stages and finally learns the knowledge from the corresponding kernel matrix for the classification performance improvement. The experiments show that the proposed method can effectively improve the accuracy of image description and the precision of image classification.
1. Introduction
Image classification is a major topic in the field of image processing and pattern recognition [1, 2]. The conventional image classification methods [3, 4] focus on some specific targets which extract effective features to represent the informative contents of images. However, this kind of method has obvious drawbacks. For example, some specific image features cannot be generalized to strange objects. In addition, image information is likely lost during the coding phase of the method.
Recently, image color features have been widely considered. Traditional color features include the color histogram, color moments, color sets, the color coherence vector, and the color correlogram. These color features are combined with contourbased features (i.e., Hu invariant moments and histogram of gradient) in the field of image classification and have achieved excellent performance results [5, 6]. A major research focus on human linguistics in the color feature representation is color names (CN). In the computer vision, the color properties contain human linguistic labels for pixels in an image. Berlin and Kay [7] presented that languages include eleven universal basic color terms. According to their analysis, Khan et al. [8] proposed to exploit the conventional statistical model to learn the knowledge of the naming color by the human’s brain, and Khan et al. [6] adopted this color feature combined with the histogram of gradient for image features fusion, which were successfully applied in target recognition. Therefore, this paper utilizes CN and dimensionreduced scale invariant feature transformation as the complementary fused features. In order to further investigate and optimize image information, a modified spatial pyramid matching method (SPM) is proposed to add the features’ spatial location information. In this method, an image would be split into patches whose features are extracted and encoded so that a uniform vector with a spatial location information can be formed. However, the simple division in SPM cannot maintain the vector representation after the rotation of the image.
Aiming to solve the problems above, we propose an image classification model based on an annular spatial pyramid matching and multifeature fusion. Considering the success of the SIFT feature, we adopt the dimensionreduced SIFT feature to speed up the encoding process. A large part of image information is contained in color features, while SIFT never takes this point into account. Thus, if color information can be added into our feature fusion model, the proposed image classification system can be improved significantly. Additionally, an annular spatial pyramid matching is applied to map the extracted features for the rotation invariance spatial vector representation. Due to various PSIFT features extracted from different images, the sparse coding algorithm is adopted for the generation of the uniform vector. In the phase of feature fusion and the learning parameters of the classification algorithm, support vector machine with two kernels algorithm (SVM2K) [9, 10] is utilized in our proposed model for training the labeled data set and predicting the unlabeled images. The SVM2K algorithm is the primary issue when combining the kernel canonical correlation analysis algorithm (KCCA) [11, 12] and SVM classification algorithm in which KCCA is adopted for preprocessing the features (PSIFT and CN), and two independent SVM models of two features are trained together. The novelty of our method lies in our novel feature fusionbased image classification method that obtains satisfying results in contrast to existing methods, while PSIFT and CN features can be used as complementary descriptions of the image.
2. Features Extraction and Spatial Mapping
2.1. Color Naming Feature
Recently, the research of image features has focused on the target shapebased local feature instead of the informationrich color features. Compared with the traditional color descriptors including HUE, OPP, and color moments, the termbased histogram statistical feature, color names, is used as the complementary to PSIFT. Through mapping RGB color channels, the extraction algorithm of CN labels in each pixel of an image is using one of 11 color names: black, blue, brown, grey, green, orange, pink, purple, red, white and yellow. The algorithm also employs the histogram and normalization to get the vector representation. CN labels are different forms of the same color to a specific color name so that the CN feature is equipped with an illumination invariance. The CN algorithm is to predict the color description by humans for a specific color in essence. The experiments in paper [5] shows that, compared with HUE and OPP color feature, the CN descriptor is more selective and has been successfully applied in the field of image recognition. Hence, this paper utilizes the CN feature to mitigate the vacancy of color in our image classification model. The comparison of images with and without color names is shown in Figure 1. For a given pixel X, the CN description of this point should be defined as the probability of the pixel belonging to one of the 11 color names.where represents the probability of the pixel assigned to the ith label, and the probability mapping matrix is determined from a large dataset. This paper utilizes a mapping matrix, which statistically computes the image dataset of 11 color names collected by Google. The max probability of color names is in equation (1). Following the above flows, all pixels in an image would be assigned a specific color name; after that, a histogram vector can be formed.
2.2. DimensionReduced SIFT Feature
SIFT originally proposed by Lowe [13] is a local gradient histogram used to locate the target shape. The algorithm identifies even the extreme points in a multiscale image space and has been widely applied in the field of image classification and pattern recognition. In recent years, SIFT has had various modifications, including SURF, PCASIFT, and HSVSIFT. However, SIFT is a vector and can be utilized to detect image features even when the image properties, including scale and image noise, are changed. However, the SIFT descriptor has a negative impact on the computation and performance of a reconstruction method. In addition, SIFT is typically taken for its accuracy in reconstruction, and the task of image classification in this paper does not need the same level of accuracy. Thus, a dimensionreduced SIFT feature generated from SIFT using a principal component analysis algorithm (PCA) [14] is proposed to extract the main information of the image feature. Due to the various number of SIFT in different image patches, the extracted SIFT cannot be used as the final representation. Thus, we have included the sparse representation in our model to encode the SIFT features of an image.
2.3. Rotational Annular Spatial Pyramid Matching
SIFT features contain corresponding location information. Lazebnik et. al [2] introduced the spatial pyramid matching model (SPM) in which images would be divided into multiscale patches so that SIFT can be a more valuable mapped for encoding. In this SPM model, SIFT features are first extracted from all images and clustered to get a visual dictionary which includes K visual words. Each image is then divided into three layers (, , ), and 1/4, 1/4, and 1/2 are independently given to the corresponding layer. Finally, the SIFT features in their patches are encoded, and the series is connected as a vector. Unfortunately, SPM has no capability of dealing with violent rotation images, so it is unable to maintain the spatial location information. While targets in the image recognition and classification are normally accompanied by large position changes, the vectors of the model would become the original SIFT vectors or CN features if SPM is integrated into our model for mapping.
To make full use of the features’ location information, we suggest an annular division based on a rotational spatial pyramid matching model (RSPM) to map SIFT and CN features. In this model, as shown in SPM, each image will be annularly divided into three layers (, , and ), which are attached to weights 1/8, 1/4, and 1/2, respectively (Figure 2). Because of the relatively small image patches in RSPM, the location information of features is more effective. For the larger patches of images, this effect is suppressed, so less weight is given to the larger layers. Finally, all features in annular circle patches will be encoded, and the series is connected as a vector.
In this way, the location information in image patches would be attached to the specific location of the vector representation. No matter how the targets rotate, the vectors of the circle patches never change so that the mapped vector by RSPM owns the rotational spatial locative information.
3. The Theory of the Sparse Representation
In the traditional bagofvisualwords model (BOVW), the kmeans algorithm [15] is used to cluster the terms for a visual dictionary. The vector quantification method (VQ) is then adopted for encoding features by computing Euclidean distances between features and visual words to get a histogram representation. However, VQ fails to consider that the Euclidean distance is unsuitable for histogrambased features (i.e., SIFT and HOG). For the reduction of the feature information in the encoding process, instead of VQ, ScSPM [2] is proposed to utilize the sparse representation algorithm and generate more sparse and selective encoding vectors. The sparse representation [16] is a kind of soft encoding algorithm which can be seen as the extension of the kmeans.
3.1. Sparse Representation
The goal of the sparse representation (SR) is to learn an ultracomplete dictionary and use rare atoms to reconstruct original signals for the successfully extraction of the embedded image information. The sparse representation has a wide range of research and practical applications, especially for the collection, compression, and representation of high dimension vectors. For instance, in face recognition, image features are first extracted from a training set and then acquired into an ultracomplete dictionary, generating sparse representations of that training set. The feature extraction and sparse representation processes for the test image set are also then performed, and the test and training samples are compared. The nature of the sparse representation is to solve the convex problem, but some specific problems, such as the low convex degree of sparse matrices, cannot be efficiently solved by a traditional convex optimization algorithm. The sparse dictionary is developed by a KSVD algorithm [17] which iteratively and simultaneously regenerates atoms and sparse coefficients. However, the time required to run an optimization algorithm is excessive, which has a significant negative effect on the classification performance. Therefore, this paper used the Kmeans++ algorithm [18] to get a stable ultracomplete dictionary, and orthogonal matching pursuit (OMP) was performed to get reconstruction coefficients. The core problem of the sparse representation is to solve equation (2).where represents the feature to be reconstructed, is the fixed dictionary, and is the sparse coefficients. The former term is the reconstruction error, and the latter term is to take control of the sparsity.
3.2. Orthogonal Matching Pursuit
An orthogonal matching pursuit algorithm (OMP) [19, 20] is a modified iterative version of a matching pursuit algorithm (MP). Due to the nonorthogonality of the selected atoms by MP, the sparse coefficients are normally local peaks. In the process of the atom selection, OMP follows the laws of MP, but converts the selected atoms into orthogonal, which decreases the iterative number of optimal convergences. At the same time, the OMP algorithm has set a maximum iterative number, so that when the number has achieved the set value, the OMP algorithm is forcibly stopped.
Ideally, each image would get a coefficient matrix when all SIFT are to be encoded. In addition, a pooling algorithm should be used to pool the matrices for uniform vectors. Experiments [21, 22] show that, when compared with other pooling methods, the maximum pooling is more effective for the generation of the sparse vector representation. Thus, we have adopted a maximum pooling algorithm in this paper. If we set that a given image has M features (K dimensions) after the mapping process of the RSPM model, the coefficient matrix is . The method of the maximum pooling can be defined in equation (3).where is the final vector representation of this image.
4. Support Vectors Machine with Double Fusion Kernels Algorithm
In the field of image recognition, the selection of a specific image feature is determined by the characteristics of the targets to be predicted. There is no effective standard to assess the feature selection. In order to mitigate any influence of feature selection on the performance of image recognition and classification, the learning of the corresponding knowledge from the training image set through machine learning algorithms has been widely researched. This feature fusion learning algorithm mainly focuses on two requirements. The first is that the classifier directly learns the series of connected vectors to achieve the performance of feature fusion. The second requirement is that each image feature is considered an individual unit to train their own models and give each model different weight.
Following the above method, researchers [23, 24] combined the preserved invariant feature with SIFT to design an image classification model that based on feature fusion and got a more brilliant performance than a single feature based model. Therefore, we proposed the support vector machine with a two fusion kernels algorithm (SVM2K) to complete the task of the image feature fusion and classification. SVM2K combines the preprocess of the kernel canonical correlation analysis (KCCA) and the parameters of an SVM algorithm to make two features that independently and complementarily describe the images. In the SVM2K algorithm, a similarity constraint between two hyper plane mappings for the organic combination of preprocess and parameter learning is introduced.
If there are two completely different features (set as A and B) for the same data set after their individual kernel mapping, the mapped features can be set as and . Then, a specific image can be described in equation (4).
The similarity constraint is defined as follows.where is the weights and thresholds of the SVM model. This constraint is introduced into the SVM function for further optimization.
The decisive function of SVM2K can be expressed as follows.
In this paper, our classification model adopts the SVM2K algorithm to perform feature fusion and classification learning on PSIFT and CN for better performance than the single featurebased SVM model. Because the SVM2K algorithm is a binary classifier, however, we follow the implementation of LibSVM [25] to utilize the “one VS one” method to extend SVM2K to multiclass classification tasks. The “one VS one” method is to train the binary SVM2K model to vote for the final predicted results of the test image. Figure 3 shows the flow chart of our feature fusionbased image classification model.
5. Experiment Results and Analysis
The experimental datasets used in this paper are Caltech256 [26] and PASCAL VOC 2011 [27]. Caltech256 is a traditional dataset in the field of computer vision. It includes 256 category image sets and each set has different number of image patches (from 31 to 800). PASCAL VOC 2011 is a benchmark test set for the detection of visual object classification which provides standard images for testing algorithms and learning performance. We randomly selected 9 categories from each dataset and divided them into training and test datasets, which are shown as Figures 4 and 5. The training dataset contained 50 images, the rest of the images in each category were set as the test dataset.
The original SIFT features are extracted by the vl_feat function library [28]. All experiments in this paper were implemented on the MATLAB 2013b platform, and Average Precision (AP) [29, 30] is introduced to assess the performance of the image classification models.
5.1. The Performance Analysis of PSIFT
PSIFT was generated by implementing a PCA algorithm on SIFT for dimension reduction. In order to assess the influence of dimensionreduced percentages on the speed of feature reconstruction and the performance of image classification, we collected the mean time (MT) spent by sparse coding and the AP performance of the proposed model. Caltech256 is utilized as the baseline dataset in this section.
Table 1 shows that when the dimensionreduction percentage increases, and the mean time spent by feature reconstruction shortens, but the corresponding AP performance is extremely slow. When the percentage is larger than 0.7, the mean time is relatively short, and the performance of classification model tends to be stable. Thus, the experiments all adopt 0.7 as the dimensionreduced percentage.

5.2. The Performance Analysis of RSPM
For the sake of elevating the rotational performance of RSPM, this section utilizes an aircraft image and implements 6 kinds of rotation transformations on it as shown in Figure 6. The SPM and RSPM models, respectively, are used to map the extracted features for further sparse representation. The length of dictionary of the sparse representation is set to be 300. After each transformed image in Figure 6 gets its vector representation, their difference degree (Diff) values are defined in equation (8), and the results are shown in Table 2.

where V is the vector of the original image, and represents the vector of the transformed vector.
As shown in Table 2, with the rotation transformation of the image, the different degree values of SPM become much larger. However, the proposed RSPM model does not have changes in degree and always keeps a relatively low level which supports the idea that RSPM has a strong adaptability for the rotation transformation of images.
In order to further compare RSPM with SPM in terms of performance, we designed experiments to utilize SPM and RSPM models to map SIFT features and use SVM and AdaBoost algorithms to recognize the images selected from Caltech256. The kernel of SVM is set as a histogram intersection kernel. The combination of SPM and SVM is set as SSVM, while the combination of SPM and AdaBoost is set as SABT. RSPM and SVM are set as RSVM, and RSPM and AdaBoost are set as RABT. The length of reference is set as 400, and other parameters are selected as the above experimental setups.
It is clearly evident from Figures 7 and 8 that the average accuracy of RSPM is higher than the 23% SPM in terms of the classification performance. The performance of RSVM in the four models is also optimal. At the same time, the image recognition accuracy of each category using RSPM is higher than SPM, which confirms that the proposed RSPM model can better obtain the spatial characteristics of images and makes the image vector representation be more selective and robust.
5.3. The Performance of SVM2K Algorithm
The experiment in this section also adopts BOWSVM [31, 32] and KSPM [33, 34] which are implemented on the Caltech256 and PASCAL VOC 2011 datasets for the comparison with the proposed model. The feature fusion model in this thesis is set as SIFTCN; the model using PSIFT and SVM is set as SSVM; and the term using CN and SVM is set as CSVM. In the SVM2K algorithm of the proposed method, HIK is utilized as the kernel of PSIFT, and the lineal kernel works as the kernel of CN. The length of the dictionary in the sparse representation varies from 200 to 500 to obtain the optimal value. The performance of the mentioned models is shown in Figures 9 and 10.
As shown in Figures 9 and 10, with the same length of the dictionary, the proposed model using PSIFT and CN for feature fusion achieved the best performance of all classification models. Compared with other models, the AP results of the fusion model increased by 510%. It is clear that the image classification model proposed in this paper has a better recognition effect on the image types. When the sparse dictionary length reaches 400, the performance of SIFTCN is optimal. Therefore, the length of the dictionary is fixed at the same time for BOWSVM, KSPM, and SIFTCN image classification models implemented on the rest of the experiments.
To better illustrate the performance of the image classification model, we follow recent literature [35, 36] that uses the kappa coefficient (equation (9)) as a measure of standards. The essence of the kappa coefficient is to measure the agreement between the interpretations of different observers. When the Kappa coefficient is 1, a negative correlation is expressed, while when the Kappa coefficient is 1, it indicates that the classification result is in complete agreement.
Table 3 gives the Kappa performance of each image description model with a 400 length of the dictionary. Taken the Caltech256 dataset for example, the Kappa performance result of the SIFTCN model is 0.68, which is nearest to 1 among models listed in Table 3. It can be concluded that the SIFTCN model is better than the other classification models. It shows that the results of the feature fusion model have a high reliability, in other words, the corresponding classification performance is the most ideal. And in the interior of the SIFTCN model, Table 3 shows that the feature fusion method is superior to the single feature performance which further verifies the excellent performance of the proposed model.

Through a series of experiments, we can see that the PSIFT feature modified by this paper has the ability to increase the speed of feature coding through the dimension reduction processing. At the same time, compared with the traditional SPM model, the proposed RSPM model maps PSIFT and CN features, which have the ideal rotation invariant phase position information. Finally, experimental comparisons show that the use of a SVM2K feature fusion classification algorithm is significantly better than using a single feature classification model.
6. Conclusion
In this paper, we propose a method of the rotation invariant spatial pyramid mapping and a feature fusionbased image classification model. In feature extraction and representation, this paper proposes a novel PSIFT feature, which is the dimensionreduced SIFT feature vector that extracts the principal information of SIFT features to speed up the sparse representation. At the same time, CN and PSIFT features are fused to describe the image. In order to explore the location information of the feature, an annular divisionbased spatial pyramid model is proposed, which makes the space vector representation of the features invariant in rotation. At the stage of feature fusion, the SVM2K algorithm is used to train two independent SVM models, and the final result is obtained by a weighted voting method. In the experimental evaluations, the Caltech256 and VOC PASCAL image databases are selected. The experimental results show that, compared with other single feature image classification model, our feature fusionbased image description model can extract the characteristic information of the image, so that PSIFT and CN characteristic can be used as complementary descriptions of the image and finally achieve the satisfying performance improvement. However, we acknowledge that beside the features that we investigated in this paper, there are many other advanced features which are required to be studied in our future work, while the robust should be concerned which means different kinds of data could be tested.
Data Availability
The image data used to support the findings of this study are included within the article.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
Acknowledgments
This work is supported partly by the UniversityLevel Research Fund Project of Nanjing Institute of Technology (No. ZKJ201907).
References
 S. Lazebnik, C. Schmid, and J. Ponce, “Beyond bags of features: Spatial pyramid matching for recognizing natural scene categories,” in 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 2169–2178, New York, NY, USA, 2006. View at: Publisher Site  Google Scholar
 J. Yang, K. Yu, Y. Gong, and T. Huang, “Linear spatial pyramid matching using sparse coding for image classification,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition, pp. 1794–1801, Miami, FL, USA, June 2009. View at: Publisher Site  Google Scholar
 J. Wang, J. Yang, K. Yu, F. Lv, T. Huang, and Y. Gong, “Localityconstrained linear coding for image classification,” in 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. 3360–3367, San Francisco, CA, USA, June 2010. View at: Publisher Site  Google Scholar
 H. Wang, Z. Chen, X. Wang, and Y. Ma, “Random finite sets based UPFCPHD multiobject tracking,” Journal on Communications, vol. 33, no. 12, pp. 147–153, 2012. View at: Google Scholar
 F. S. Khan, J. van de Weijer, and M. Vanrell, “Modulating shape features by color attention for object recognition,” International Journal of Computer Vision, vol. 98, no. 1, pp. 49–64, 2012. View at: Publisher Site  Google Scholar
 F. S. Khan, J. Weijer, A. D. Bagdanov, and M. Vanrell, “Portmanteau vocabularies for multicue image representation,” in Advances in Neural Information Processing Systems, pp. 1323–1331, Granada, Spain, 2011. View at: Google Scholar
 Y. Wang, J. Liu, J. Wang, Y. Li, and H. Lu, “Color names learning using convolutional neural networks,” in 2015 IEEE International Conference on Image Processing (ICIP), pp. 217–221, Quebec City, QC, September 2015. View at: Publisher Site  Google Scholar
 F. S. Khan, R. M. Anwer, J. Weijer, A. D. Bagdanov, M. Vanrell, and A. M. Lopez, “Color attributes for object detection,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, pp. 3306–3313, Providence, RI, USA, June 2012. View at: Publisher Site  Google Scholar
 J. D. Farquhar, D. R. Hardoon, H. Meng, J. ShaweTaylor, and S. Szedmak, “Two view learning: SVM2K, theory and practice,” in Advances in Neural Information Processing Systems, pp. 355–362, Vancouver, BC, Canada, 2005. View at: Google Scholar
 Y. Lu, Research on targets detection method in high resolution remote sensing image, Hohai University, 2016.
 W. Zheng, X. Zhou, C. Zou, and L. Zhao, “Facial expression recognition using kernel canonical correlation analysis (KCCA),” Neural Networks, vol. 17, no. 1, pp. 233–238, 2006. View at: Publisher Site  Google Scholar
 X. Wang, X. Yan, G. Lv, and T. Fan, “Balloonborne spectrum–polarization imaging for river surface velocimetry under extreme conditions,” Infrared Physics & Technology, vol. 58, pp. 5–11, 2013. View at: Publisher Site  Google Scholar
 D. G. Lowe, “Distinctive image features from scaleinvariant keypoints,” International Journal of Computer Vision, vol. 60, no. 2, pp. 91–110, 2004. View at: Publisher Site  Google Scholar
 H. Abdi and L. J. Williams, “Principal component analysis,” Wiley Interdisciplinary Reviews: Computational Statistics, vol. 2, no. 4, pp. 433–459, 2010. View at: Publisher Site  Google Scholar
 S. E. Hickman, N. D. Kingery, T. K. Ohsumi et al., “The microglial sensome revealed by direct RNA sequencing,” Nature Neuroscience, vol. 16, no. 12, pp. 1896–1905, 2013. View at: Publisher Site  Google Scholar
 H. Lee, A. Battle, R. Raina, and A. Y. Ng, “Efficient sparse coding algorithms,” in Advances in Neural Information Processing Systems, pp. 801–808, Vancouver, BC, Canada, 2006. View at: Google Scholar
 Z. Jiang, Z. Lin, and L. S. Davis, “Label consistent KSVD: learning a discriminative dictionary for recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 35, no. 11, pp. 2651–2664, 2013. View at: Publisher Site  Google Scholar
 D. Arthur and S. Vassilvitskii, “Kmeans++: the advantages of careful seeding,” in Proceedings of the eighteenth annual ACMSIAM symposium on Discrete algorithms. Society for Industrial and Applied Mathematics, pp. 1027–1035, Philadelphia, PA, USA, 2007. View at: Google Scholar
 M. Tan, I. W. Tsang, and L. Wang, “Matching pursuit lasso part I: sparse recovery over big dictionary,” IEEE Transactions on Signal Processing, vol. 63, no. 3, pp. 727–741, 2013. View at: Publisher Site  Google Scholar
 W. Guo, N. Xiong, A. V. Vasilakos, G. Chen, and C. Yu, “Distributed k–connected fault–tolerant topology control algorithms with PSO in future autonomic sensor systems,” International Journal of Sensor Networks, vol. 12, no. 1, pp. 53–62, 2012. View at: Publisher Site  Google Scholar
 J. X. Wu and J. M. Rehg, “Beyond the euclidean distance: creating effective visual codebooks using the histogram intersection kernel,” in 2009 IEEE 12th International Conference on Computer Vision, pp. 630–637, Kyoto, Japan, September 2009. View at: Publisher Site  Google Scholar
 H. Liang, J. Zou, Z. Li, M. J. Khan, and Y. Lu, “Dynamic evaluation of drilling leakage risk based on fuzzy theory and PSOSVR algorithm,” Future Generation Computer Systems, vol. 95, pp. 454–466, 2019. View at: Publisher Site  Google Scholar
 Q. Wang, X. Zhang, M. Li, X. Dong, Q. Zhou, and Y. Yin, “Adaboost and multiorientation 2D Gaborbased noisy iris recognition,” Pattern Recognition Letters, vol. 33, no. 8, pp. 978–983, 2012. View at: Publisher Site  Google Scholar
 J. Li, N. Xiong, J. H. Park, C. Liu, S. MA, and S. E. Cho, “Intelligent model design of cluster supply chain with horizontal cooperation,” Journal of Intelligent Manufacturing, vol. 23, no. 4, pp. 917–931, 2012. View at: Publisher Site  Google Scholar
 C. C. Chang and C. J. Lin, LIBSVM: a library for support vector machines, 2001, Software, http://www.csie.ntu.edu.tw/cjlin/libsvm.
 G. Griffin, A. Holub, and P. Perona, Caltech256 Object Category Dataset, California Institute of Technology, 2007.
 K. Chatfield, V. Lempitsky, A. Vedaldi, and A. Zisserman, “The devil is in the details: an evaluation of recent feature encoding methods,” BMVC, vol. 2, no. 4, p. 8, 2011. View at: Publisher Site  Google Scholar
 J. Li, X. Li, B. Yang, and X. Sun, “Segmentationbased image copymove forgery detection scheme,” IEEE Transactions on Information Forensics and Security, vol. 10, no. 3, pp. 507–518, 2015. View at: Publisher Site  Google Scholar
 Y. Yue, T. W. Finley, F. Radlinski, and T. Joachims, “A support vector method for optimizing average precision,” in Proceedings of the 30th annual international ACM SIGIR conference on Research and development in information retrieval  SIGIR '07, pp. 271–278, New York, NY, USA, July 2007. View at: Publisher Site  Google Scholar
 Y. Liu, M. Ma, X. Liu, N. Xiong, A. Liu, and Y. Zhu, “Design and analysis of probing route to defense sinkhole attacks for Internet of Things security,” IEEE Transactions on Network Science and Engineering, vol. 7, no. 1, pp. 356–372, 2020. View at: Publisher Site  Google Scholar
 A. P. LópezMonroy, M. MontesyGómez, H. J. Escalante, A. CruzRoa, and F. A. González, “Improving the BoVW via discriminative visual ngrams and MKL strategies,” Neurocomputing, vol. 175, pp. 768–781, 2016. View at: Publisher Site  Google Scholar
 M. Xu, Q. Sun, Y. Lu, and C. Shen, “Nearestneighbors based weighted method for the BOVW applied to image classification,” Journal of Electrical Engineering & Technology, vol. 10, no. 4, pp. 1877–1885, 2015. View at: Publisher Site  Google Scholar
 K. He, X. Zhang, S. Ren, and J. Sun, “Spatial pyramid pooling in deep convolutional networks for visual recognition,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 37, no. 9, pp. 1904–1916, 2015. View at: Publisher Site  Google Scholar
 H. Liang, J. Zou, K. Zuo, and M. J. Khan, “An improved genetic algorithm optimization fuzzy controller applied to the wellhead back pressure control system,” Mechanical Systems and Signal Processing, vol. 142, article 106708, 2020. View at: Publisher Site  Google Scholar
 Y. Chen, Z. Lin, X. Zhao, G. Wang, and Y. Gu, “Deep learningbased classification of hyperspectral data,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 7, no. 6, pp. 2094–2107, 2014. View at: Publisher Site  Google Scholar
 F. Long, N. Xiong, A. V. Vasilakos, L. T. Yang, and F. Sun, “A sustainable heuristic QoS routing algorithm for pervasive multilayered satellite wireless networks,” Wireless Networks, vol. 16, no. 6, pp. 1657–1673, 2010. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2020 Mengxi Xu et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.