Computational Intelligence and Neuroscience

Computational Intelligence and Neuroscience / 2019 / Article

Research Article | Open Access

Volume 2019 |Article ID 2060796 | https://doi.org/10.1155/2019/2060796

Jianfang Cao, Chenyan Wu, Lichao Chen, Hongyan Cui, Guoqing Feng, "An Improved Convolutional Neural Network Algorithm and Its Application in Multilabel Image Labeling", Computational Intelligence and Neuroscience, vol. 2019, Article ID 2060796, 12 pages, 2019. https://doi.org/10.1155/2019/2060796

An Improved Convolutional Neural Network Algorithm and Its Application in Multilabel Image Labeling

Academic Editor: Leonardo Franco
Received01 Mar 2019
Revised27 May 2019
Accepted23 Jun 2019
Published04 Jul 2019

Abstract

In today’s society, image resources are everywhere, and the number of available images can be overwhelming. Determining how to rapidly and effectively query, retrieve, and organize image information has become a popular research topic, and automatic image annotation is the key to text-based image retrieval. If the semantic images with annotations are not balanced among the training samples, the low-frequency labeling accuracy can be poor. In this study, a dual-channel convolution neural network (DCCNN) was designed to improve the accuracy of automatic labeling. The model integrates two convolutional neural network (CNN) channels with different structures. One channel is used for training based on the low-frequency samples and increases the proportion of low-frequency samples in the model, and the other is used for training based on all training sets. In the labeling process, the outputs of the two channels are fused to obtain a labeling decision. We verified the proposed model on the Caltech-256, Pascal VOC 2007, and Pascal VOC 2012 standard datasets. On the Pascal VOC 2012 dataset, the proposed DCCNN model achieves an overall labeling accuracy of up to 93.4% after 100 training iterations: 8.9% higher than the CNN and 15% higher than the traditional method. A similar accuracy can be achieved by the CNN only after 2,500 training iterations. On the 50,000-image dataset from Caltech-256 and Pascal VOC 2012, the performance of the DCCNN is relatively stable; it achieves an average labeling accuracy above 93%. In contrast, the CNN reaches an accuracy of only 91% even after extended training. Furthermore, the proposed DCCNN achieves a labeling accuracy for low-frequency words approximately 10% higher than that of the CNN, which further verifies the reliability of the proposed model in this study.

1. Introduction

With the rapid development and increasing popularity of multimedia devices and network technologies, increasing amounts of information are being presented in image form. The large number of rich image resources has attracted users, who can find the information that they need in the images. According to statistics from Flickr, a website for social image sharing on the Internet, image storage is growing at an annual rate of 100 million units, while Facebook image storage is growing at a rate of 15 billion units per year [1]. However, this massive amount of image information can easily overwhelm users. Determining how to rapidly and effectively query, retrieve, and organize image information has become a major problem that must be solved [2]. As a result, the field of image retrieval technology has emerged and received considerable attention. Specifically, image annotation can provide more search information than traditional methods and result in rapid retrieval of corresponding images. However, because images often contain complex and diverse semantic information, they are typically labeled with more than one label; thus, it is necessary to consider the case of multilabel annotation.

Generally, the methods for automatically labeling multilabel images can be divided into three main categories: generative models, discriminant models, and nearest neighbor models. Generative models can generate training data randomly, particularly when certain implicit parameters are given. These models construct the joint distribution probability of the visual features and the text semantic tags first and then calculate the posterior probability of each semantic feature of the known image with a Bayesian probabilistic model, which they use to complete the semantic annotation of the image [3]. Duygulu et al. [4] proposed a generative model called the translation model, which transforms the image semantic annotation process into a translation process by transforming visual image keywords into semantic keywords. Jeon et al. [5] proposed the cross-media relevance model (CMRM), which models images to perform image annotation by constructing the joint probability of the visual and semantic information. Although the above model considers the semantics of objects and regions, the discrete processing of visual features can result in the feature loss. In addition, the labeling effect is largely influenced by the clustering granularity, but the ideal granularity parameters are difficult to determine in advance. To solve this problem, Feng et al. [6] proposed the multiple Bernoulli relevance model (MBRM), and Alkaoud et al. [7] proposed the fuzzy cross-media relevance model (FCRM). These models use a nonparametric Gaussian kernel to perform a continuous estimation of the feature generation probability. Compared with the discrete model, these models significantly improve the labeling accuracy. Although the annotation process of the abovementioned production annotation model is relatively simple, the gap between the underlying features of the image and the high-level semantics and the nonindependence among the semantics can lead to inaccurate joint probabilities [8].

A discriminative model defines image annotation as a traditional supervised classification problem. This approach performs image annotation mainly by determining the correlations between visual features and predefined labels [9]. The authors of [10] used the K-nearest neighbor (KNN) method to select the nearest K images by calculating the distance between graphs and then labeling the unlabeled image using a label propagation algorithm. Li et al. [11] used a K-means algorithm to construct a classifier by combining a semantic vocabulary with annotated words using semantic constraints and used the classifier for subsequent image annotation. Qiu et al. [12] used a support vector machine (SVM) to semantically denote some areas and then label unlabeled areas based on the relationships among the areas. Regardless of whether a method is based on one-to-one classification or one-to-many classification, it is subject to the constraints of the number of classifiers and the training effect of the classifier, especially in the case of unbalanced training samples. If the classifier training effect is poor, the overall labeling accuracy rate will be affected. As the size of the label set increases, the required classifier size also increases, which increases the complexity of the labeling model; thus, some methods may not be applicable in big data environments [13].

The nearest neighbor model has become popular as the requirements of data training have expanded. The authors of [14] introduced the transmission mechanism of nearest neighbor labeling. In this approach, image annotation is treated as a retrieval problem. The nearest neighbor depends on the averages of several distances calculated from visual features, also known as the joint equal contribution (JEC). For a given image, a label is passed through a neighbor. Visual characteristics such as color and texture are used for comparison and testing, and feature selection regularization is performed based on label similarity. However, this approach does not increase the sparsity or improve the accuracy of labels in all cases. The TagProp model [15] is a new type of nearest neighbor model. It creates combined weights based on the existence or nonexistence of neighbor labels and achieves good results. The traditional methods described above have advanced the field of image annotation, but they require manual feature selection, which can result in information loss, poor annotation accuracy, and a low recall rate [16].

In recent years, as deep learning has received increasing attention, some scholars have begun to apply deep learning to computer vision tasks. In 2012, Hinton et al. used a multilayer convolutional neural network to classify images using the widely used large-scale ImageNet database [17] for image recognition and achieved exceptional recognition results [18]. Since then, a large number of studies have developed improved network structures and increased CNN performance. For example, Google’s GoogLeNet network [19] won the championship in the 2014 large-scale image recognition competition. The Visual Computing Group of Microsoft Research Asia developed a computer vision system based on a deep convolutional neural network that—for the first time—surpassed humans in its ability to identify and classify objects in the ImageNet 1000 challenge [20]. Although deep learning models have performed well on image recognition and classification tasks, most studies have focused on the network itself or on improvements in single-label learning. Specifically, the task of image annotation for multilabel learning has been given little attention, particularly for unbalanced datasets. Currently, the approaches to solve the issue of dataset imbalance primarily focus on the composition of the datasets themselves. Briefly, a balance is achieved in the entire dataset by reducing the number of the types of images that are overrepresented in the dataset (downsampling) or increasing the number of the types of images that are underrepresented (upsampling). Despite their easy operation, these approaches require the image quantities in the dataset to reach a certain level and therefore do not possess wide applicability. Images types that rarely occur in daily life are difficult to even obtain, let alone to obtain in large numbers, and simple translation transformation sometimes fails to satisfy the requirement.

By combining deep learning and multilabel image annotation and targeting the problems of insufficient training and the poor annotation effects of underrepresented images due to data imbalance, in this study, we design a dual-channel convolutional neural network (DCCNN) and propose a new multilabel image annotation method. To increase the annotation accuracy, particularly that of low-frequency words, the proposed model is designed with two input channels and one output channel constructed by two six-layer convolutional neural networks that use different parameters. One of the two input channels is trained on the entire dataset, while the other is specially trained on the low-frequency portions of the dataset. These two channels are independent; the low-frequency datasets undergo training two times, thus increasing the training weights of low-frequency words in the integral dataset. During testing, the outputs from both input channels are combined to form a joint decision, thus achieving an optimal annotation effect.

The main contributions of this study are summarized as follows:(1)The combination of deep learning and multilabel image annotation solves the issues of the complex annotation process, poor annotation efficiency, the deficiencies in determining characteristics, and the “semantic gap” that affect traditional annotation methods.(2)We design a DCCNN model that fuses two different convolutional neural networks. Based on an understanding of the convolutional neural networks themselves as well as the experimental results, the parameters are adjusted, and a fusion ratio is set between the two subnetworks that result in a satisfactory performance. The designed model is aimed at solving the poor annotation effect problem that occurs on underrepresented image types in datasets due to insufficient training. Compared with other methods to address such problems, the method proposed in this study is both convenient and fast, and its application is not restricted by datasets.(3)Based on an understanding of the general process of multilabel image annotation, this study proposes a multilabel image annotation algorithm that employs the DCCNN. This algorithm contains training and annotation phases, and the inputs and outputs differ according to the different phases. During the training phase, the two branch models are trained independently. Then, in the testing phase, these branch models are fused so that they exert joint contributions to decision-making with regard to the final annotation results.

2. Methods

2.1. Convolutional Neural Network

The first convolutional neural network (CNN) was proposed by Hubel and Wiesel [21] in the 1960s through studies of neurons in monkey cortexes related to local sensitivity and direction selection. CNNs use weighted sharing, downsampling, and local connection techniques that greatly reduce the number of required parameters and the complexity of the neural network. CNNs have been compared to traditional methods of image feature extraction such as the Histogram of Oriented Gradient (HOG) and Scale-Invariant Feature Transform (SIFT) methods; however, CNNs can typically extract more abstract and comprehensive features. In addition, CNNs avoid the need for complex image preprocessing because they can use the original images directly as input.

CNNs are mainly composed of a convolutional layer, a pooled layer, and a fully connected layer. The convolutional layer is a key part of the CNN. The function of this layer is to extract features from input images or feature maps. Each convolutional layer can have multiple convolution kernels, which are used to obtain multiple feature maps. The convolution layer is calculated as follows [22]:where is the characteristic map of the output of the previous layer, is the output of the th channel of the th convolution layer, and is called the activation function. Here, is a subset of the input feature maps used to calculate , is a convolution kernel, and is the corresponding offset.

A pooling layer is generally sandwiched between two convolutional layers. The main function of this layer is to reduce the dimensions of the feature map and maintain the scale invariance of the features to some extent. There are two main pooling methods: mean pooling and max pooling. A pooling effect diagram is shown in Figure 1.

The pooling process is similar to the convolution process in that it involves a sliding window similar to a filter, but the calculation is simpler. Mean pooling uses the average value in an image area as the pooled value of the area. This approach preserves the background of the image well. Max pooling takes the maximum value of the image area as the pooled value of the area and preserves the texture of the image well. The function of the fully connected layer is to integrate the multiple image maps obtained after the image is passed through multiple convolution layers and pooling layers to obtain the high-layer semantic features of the image for subsequent image classification.

2.2. Dual-Channel Convolutional Neural Network (DCCNN)

In the image annotation problem, one image often corresponds to a plurality of annotated words, and different annotated words correspond to different scenes. Some scenes correspond to many images, that is, the corresponding frequency of the annotated words is large, such as sun, white clouds, mountains, and rivers. Additionally, some scenes correspond to few images, and their corresponding word frequency is small, such as crocodile and lizard. Unbalanced input data can result in the insufficient training of low-frequency annotated words, resulting in a poor recognition rate. To increase the recognition accuracy and the overall recognition efficiency for low-frequency annotated words, this paper designs a DCCNN model (Figure 2).

In Figure 2, the DCCNN model consists of the convolutional neural networks CNN0 and CNN1. Each of these networks possesses three convolution layers and three fully connected layers.

CNN0 is trained on all training sets and the parameters of each layer are as follows:Layer 1 consists of 20 10 × 10 convolution kernels that perform convolution operations on the input images. The step length is initially set to 4. Then, 3 × 3 max pooling windows with a step length of 2 are used for downsampling.Layer 2 consists of 40 5 × 5 convolution kernels that perform convolution operations on the feature maps. The step length is initially set to 2. Then, 3 × 3 max pooling windows with a step length of 2 are used for downsampling.Layer 3 consists of 60 3 × 3 convolution kernels that perform convolution operations on the input feature maps. The step length is set to 1. The remaining three layers are fully connected layers. A dropout layer is applied for full connection to avoid overfitting. The keep pro (proportion) parameter is set to 0.5 (i.e., 50% of the neurons at each of the fully connected layers participate in the operation). The number of output nodes is 20. Considering that the ReLU activation function possesses powerful expression ability and is free from the vanishing gradient problem, enabling the convergence rate of the model to be maintained stably, we used the ReLU function in this study for all activations. The learning rate was adjusted to 0.001 after the experiment.

CNN1 is trained on the low-frequency training sets, and the parameters for each layer are as follows:Layer 1 consists of 20 12 × 12 convolution kernels that perform convolution operations on the input images. The step length is initially set to 2. Then, 5 × 5 max pooling windows with a step length of 4 are used for downsampling.Layer 2 consists of 40 5 × 5 convolution kernels that perform convolution operations on the input feature maps. The step length is initially set to 1. Then, 4 × 4 max pooling windows with a step length of 2 are used for downsampling.Layer 3 consists of 60 4 × 4 convolution kernels that perform convolution operations on the input feature maps. The step length is set to 1. Then, 4 × 4 max pooling windows with a step length of 2 are used for downsampling.The final three layers are all fully connected layers whose parameters are the same as those of CNN0.

The model increases the training weight of the low-frequency samples using special training channels. The training samples are first processed separately to achieve sample equalization. Then, during the labeling process, the final labeling result is jointly determined by the two channels. Because the low-frequency channel is trained only with low-frequency samples, the parameters of this channel are more suitable for identifying low-frequency samples, which reduces the labeling impact of training with insufficient numbers of low-frequency samples.

2.3. Multilabel Image Annotation

In this paper, the labeling algorithm is divided into two phases: the training phase and the labeling phase, as shown in Figure 3.

The algorithm corresponding to the training phase is as follows.Step 1. Launch the algorithm corresponding to the training phase. Sum the number of samples corresponding to each tagged word and determine the low-frequency annotation word set.Step 2. Through the program, all the low-frequency samples in the training sample are extracted to form a low-frequency training set.Step 3. Construct a CNN model with two channels: CNN0 and CNN1. CNN0 corresponds to the channel with a small convolutional kernel and a large step, and CNN1 corresponds to the channel with a large convolutional kernel and a small step. The first layer of the CNN1 channel is fully connected to the second layer of the fully connected layer for feature fusion.Step 4. Input all the training sets into the CNN0 channel and input only the low-frequency training samples into the CNN1 channel. Conduct model training until the model becomes stable.

The labeling phase algorithm is as follows.Step 1. Input the test image into both channels (CNN0 and CNN1) of the trained two-channel CNN for feature extractionStep 2. Fuse the output vectors of the two channels in a 2 : 1 manner (the specific ratio is experimentally determined)Step 3. Combine the decision results of the two channels to perform image annotation

2.4. Experimental Data

To validate the proposed dual-channel CNN (DCCNN) algorithm, we performed experiments using the following freely available datasets: Caltech-256 [23], Pascal VOC 2007 [24], and Pascal VOC 2012 [25]. The Caltech-256 dataset contains 256 categories, each with at least 80 images and 30,608 overall images. The Pascal VOC 2007 dataset contains 20 categories and 9,963 images, with approximately 450 images per category. Based on the Pascal VOC 2007 dataset, Pascal VOC 2012 includes more images in each category, extending the dataset to 22,531 images over 20 categories, with each category containing approximately 1,000 images. Figure 4 shows some example images from the three datasets.

2.5. Experimental Design

In this paper, we conducted simulation experiments based on the framework of the deep learning algorithm TensorFlow. Most of the images investigated in the multilabel image annotation task correspond to more than one label; thus, the evaluation criteria for single-label image classification are not fully applicable to multilabel image tasks. In this paper, we use mean average precision (MAP) as a new metric for multilabel images. The MAP score is obtained by calculating the average precision (AP). For a given task or category, the corresponding precision-recall curve can be calculated. Then, a set of thresholds is established [0, 0.1, 0.2, …, 1]. If the recall rate is greater than each threshold, the corresponding maximum precision value is obtained. AP is the average of the 11 precision scores, and its formula is as follows:where is the maximum precision value corresponding to each threshold:and is the accuracy rate corresponding to each . Finally, the MAP is calculated bywhere is the number of categories.

We adopted MAP as a metric because it encompasses the accuracy, recall rate, and single-point value limitations of the F1 value and reflects the overall global performance.

Therefore, to verify the effectiveness of the proposed dual-channel CNN algorithm, we combine the characteristics of the three datasets (the Pascal VOC dataset image categories are small, and each category contains more images; the Caltech-256 dataset image categories are large, and each category contains fewer images) and compare the results with those reported in the literature in terms of labeling accuracy, automatic labeling effect, iteration number, and MAP score.

3. Results and Discussion

3.1. Labeling Accuracy Comparison
3.1.1. Comparison on the Pascal VOC 2012 Datasets

The Pascal VOC 2012 datasets feature fewer categories and more images per category than does the Caltech-256 dataset. Therefore, we conducted an experimental comparison of the accuracy of the proposed DCCNN model with various methods from the literature [2628] and the common CNN based on each category of the Pascal VOC 2012 dataset. The results are shown in Table 1.


Image categoryLabeling accuracy
Zuo et al. [26]Zhang et al. [27]Islam et al. [28]CNNDCCNN
2007201220072012200720122007201220072012

Plane0.8020.7770.9880.9730.9470.9240.9920.9831.00.999
Bike0.5010.4250.8120.7480.4980.4510.9050.8770.9960.973
Bird0.5610.4540.8730.8080.9620.9461.00.9771.00.984
Boat0.6190.5330.8990.8530.6710.6520.9350.9200.9900.972
Bottle0.280.240.6910.6080.7910.7580.8950.8790.9240.919
Bus0.7840.7220.9310.8990.9660.9510.9760.9710.9860.980
Car0.5840.5060.8970.8680.9050.8910.9530.9490.9890.987
Cat0.6070.5420.9410.8930.9410.9230.9620.9550.9840.970
Chair0.5090.4530.6130.5540.4220.390.8260.7940.9070.893
Cow0.3090.260.8480.7780.8660.8571.00.9991.01.0
Dining table0.3980.3660.8290.7510.7490.7040.8430.8250.8990.885
Dog0.5070.4260.8850.830.8950.8860.9180.9050.9920.971
Horse0.4410.3890.9160.8750.9120.8940.9290.9270.9800.978
Motorbike0.570.5070.8250.7920.7980.7610.8510.8490.9310.931
Person0.7690.7030.8990.8470.8310.7940.8990.8970.9620.957
Potted plant0.3050.2340.6360.5780.6810.6580.8330.8290.8840.881
Sheep0.4060.3620.8240.7920.8950.8620.9600.9600.9960.993
Sofa0.3850.3140.4490.3950.3890.3390.7250.7190.8190.816
Train0.6990.6160.9350.9060.8640.8181.00.9891.01.0
TV monitor0.5230.430.7930.7740.7850.7290.8190.8160.8570.856
MAP value0.5280.4630.8240.7760.7880.7590.9110.9010.9550.947

Table 1 shows that the DCCNN method proposed in this paper yields a significantly higher labeling accuracy for each category than do the other four methods. Compared with the three methods from the literature and the CNN, the proposed method increased the MAP values on the Pascal VOC 2007 dataset by 42.7%, 13.1%, 16.7%, and 4.4%, and the MAP values on the Pascal VOC 2012 dataset increased by 48.4%, 17.1%, 18.8%, and 4.6%, respectively.

3.1.2. Comparisons on the Mixed Datasets from Caltech-256 and Pascal VOC 2012

We combined the Caltech-256 and Pascal VOC 2012 datasets to form a larger dataset containing 50,000+ images and 276 categories and then compared the average labeling accuracy of the various models, as shown in Figure 5.

Additionally, to verify the influence of the number of image categories on the labeling accuracy of the algorithm proposed in this paper, we increased the number of image categories to evaluate the stability of the accuracy of the algorithm on the large dataset composed of Caltech-256 and Pascal VOC 2012 images. The results are shown in Figure 6.

The results in Figures 5 and 6 show that when the number of images reaches approximately 50,000, the DCCNN algorithm proposed in this paper achieves a higher labeling accuracy (93%) than do the other algorithms. Moreover, the results in Figure 6 show that the accuracy of the DCCNN method does not fluctuate as the number of image categories increases, further demonstrating that the DCCNN method can effectively avoid the information limitations associated with manual feature selection.

3.2. Comprehensive Comparison with a CNN Method
3.2.1. Comparison of the Pascal VOC 2012 Datasets

To verify the improvement of the proposed method over the CNN algorithm, we first performed an experimental comparison of the accuracy of the two methods based on the same number of iterations using the Pascal VOC 2012 dataset. Table 2 shows the experimental results when both models were trained 100 times.


Image categoryCNNDCCNN

Plane0.9830.999
Bike0.8770.973
Bird0.9180.984
Boat0.9200.972
Bottle0.7220.892
Bus0.9200.980
Car0.8190.939
Cat0.9160.970
Chair0.6680.804
Cow0.9991.0
Dining table0.5700.757
Dog0.8940.971
Horse0.9270.978
Motorbike0.8490.931
Person0.8710.957
Potted plant0.7290.881
Sheep0.9600.993
Sofa0.6180.827
Train1.00.999
TV monitor0.7460.866
MAP value0.8450.934

Table 2 shows that the DCCNN proposed in this paper achieves good results within 100 training steps. The ordinary CNN achieves nearly the same effect only after being trained approximately 2,500 times. This finding indicates that the DCCNN proposed in this study is more accurate and efficient than the traditional CNN. The time and number of iterations required to reach the same accuracy percentage are shown in Table 3.


MethodsCNNDCCNN

Time (s)277.41141.446
Number of iterations2500100

3.2.2. Comparison of the Mixed Caltech-256 and Pascal VOC 2012 Datasets

For the large dataset consisting of Caltech-256 and Pascal VOC 2012 images, the MAP values of the CNN and the DCCNN are compared based on the number of training iterations required, as shown in Figure 7.

Figure 7 shows that the MAP value of the CNN is approximately 82% initially but finally stabilizes at 91.9%. In contrast, the MAP value of the DCCNN eventually reaches 93.8%, 1.9% higher than that of the CNN.

In addition, we report the AP values of the two models for some vocabulary words after achieving stable training (Table 4 and Figure 8).


VocabularyCNNDCCNNDifference in AP

Plane0.9970.9980.001
Bird0.9770.9720.005
Boat0.9750.9900.015
Bus0.9840.9790.005
Dog0.9700.9710.001
Horse0.9800.9790.001
Train1.00.9990.001

As shown in Table 4 and Figure 8, the annotation results of the CNN and DCCNN are similar because the effect of each model is nearly saturated after sufficient training. However, as shown in Table 3, the training rate of the CNN is much slower than that of the DCCNN. In addition, Figure 8 shows that the AP values of the DCCNN for some words are higher than those of the CNN, indicating a better annotation effect.

3.3. Comparisons of Low-Frequency Word Efficiency

To better verify the efficacy of the method proposed in this study, we compared the annotation effects of different methods on low-frequency words. As shown in Table 1, because the methods of [2628] require manual feature selection, the labeling accuracy is very low for certain categories. According to [29], under the standard practice for low-frequency vocabulary, the highest and lowest frequency words are eliminated, and the frequencies of the remaining words are averaged. A labeled word corresponding to a below-average value is a low-frequency word. The results in Table 5 show that categories such as bike, bottle, chair, dining table, potted plant, sofa, and TV monitor are low-frequency words when labeling is based on manual feature selection.


Sample imageAutomatic labeling results
Islam et al. [28]CNNDCCNN

PeoplePeople, sofaPeople, sofa, chair

SofaSofa, potted plantSofa, potted plant, chair, TV monitor

Chair, sofaChair, sofaChair, sofa, dining table

People, carPeople, carPeople, car, TV monitor

Chair, dining tablePotted plant, chair, dining tablePotted plant, chair, dining table, sofa

Dining table, chairDining table, chair, potted plantDining table, chair, potted plant, sofa

MotorbikePeople, motorbikePeople, motorbike, chair

People, busPeople, bus, carPeople, bus, car, bike

Chair, TV monitorSofa, chair, TV monitorSofa, chair, TV monitor, dining table

Chair, peopleDog, chair, people, sofaDog, chair, people, sofa, potted plant

3.3.1. Comparison of Labeling Effects

In this section, we compare the DCCNN method with the method proposed in [27] and with CNN-based automatic labeling (Table 5).

The labeling results in Table 5 indicate that the DCCNN method provides more complete image descriptions and more comprehensive annotations than do the other two methods. Additionally, its recognition rate for low-frequency words such as chair, sofa, TV monitor, and dining table is higher.

3.3.2. Comparison of Annotation Accuracy

For low-frequency words, as shown in Figure 9, the labeling accuracies when using the CNN and the DCCNN method proposed in this paper are significantly higher, and the DCCNN method yields the highest accuracy. This result indicates that compared with traditional methods (the methods in [2628]), the extracted image features are more abstract, more comprehensive, and more suitable for the high-level semantics associated with the human understanding of images. Therefore, its labeling accuracy is significantly higher. The DCCNN model proposed in this paper is an integrated model of two different CNNs. One CNN has a small convolutional kernel and a large step, and the other has a large convolutional kernel and a small step. During training, to increase the training weights of low-frequency vocabulary words, images corresponding to the low-frequency vocabulary are input into the CNN with a large convolutional kernel and a small step. In contrast, the other CNN is trained on all the training images. This approach results in the DCCNN having a higher accuracy than the classical CNN method alone.

Moreover, we also evaluated the annotation accuracy of each method based on some low-frequency words in the large dataset composed of Caltech-256 and Pascal VOC 2012 images after the two models became relatively stable. The experimental pairs are shown in Table 6 and Figure 10.


Low-frequency wordsLabeling accuracy
CNNDCCNNDifference

Bike0.9350.9640.029
Dining table0.7560.8040.048
Potted plant0.8850.8980.013
Sofa0.5620.6690.107
TV monitor0.8490.8820.033

Figure 10 is a visual diagram of the information in Table 6. Most of the average accuracy improvements are similar. The annotation accuracies of “bike,” “dining table,” “potted plant,” “sofa,” and “TV monitor” increased by 2.9%, 4.8%, 1.3%, 10.7%, and 3.3%, respectively, when using the DCCNN rather than the CNN. Thus, the DCCNN designed in this paper improves the proportions of low-frequency annotated words in the model, further validating the effectiveness of the model.

4. Conclusions

In real life, with the continuous development of technology, access to information has grown explosively, and the amounts of image data have increased dramatically. Determining how to efficiently and rapidly solve large-scale image processing problems has become a popular and challenging research topic. Automatic image annotation is the key to text-based image retrieval. While the traditional methods made significant progress in the field of image annotation, their reliance on manual feature selection caused some information to be missed, resulting in poor annotation accuracy and low recall rates. Although deep learning models have been successfully applied for image recognition and classification, most studies have focused on the specific network used or on improving single-label learning. Notably, the application and improvement of multilabel image annotation in deep learning have been given little attention.

Therefore, based on the characteristics of multilabel learning and considering the uneven distribution of labeled words, we propose a DCCNN to improve the training weights of low-frequency words and the overall labeling efficiency. We validated the model with classic, commonly used multilabel image datasets: the Caltech-256, Pascal VOC 2007, and Pascal VOC 2012 datasets. In this study, we compared the DCCNN with existing methods from the literature and a traditional CNN. The methods based on CNNs are more effective for image annotation than are the traditional methods based on the manual feature selection. We also conducted a comprehensive comparison between the DCCNN and CNN. The results verify that the DCCNN improves both the accuracy of low-frequency vocabulary labeling and the overall labeling efficiency.

The next steps in this research are threefold: (1) training samples will be grouped according to the word frequency, and a multichannel CNN model will be established to further reduce the influence of word frequency on the model; (2) the labeling results will be further improved by considering the symbiotic relationships among words and the distances between mapped words; and (3) tests will be performed using larger datasets. Finally, based on the results, we will make improvements that further enhance the stability of the solution. The use of larger image datasets has certain benefits for network training and avoids overfitting.

Data Availability

The data used to support the findings of this study are included within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

Acknowledgments

This work was supported by the Natural Science Foundation of Shanxi Province (grant no. 201701D21059), an Art and Science Planning Project of Shanxi Province (grant no. 2017F06), an Education Science Planning Project of the 13th Five-year Plan of the Key Discipline Project of Shanxi Province (grant no. GH-17059), and the Platform and Personnel Specialty of Xinzhou (20180601).

References

  1. Z. Y. Chu, Q. Fang, J. T. Sang, and C. S. Xu, “Image annotation based on region context perception,” Journal of Computer Science, vol. 37, pp. 1390–1397, 2014. View at: Google Scholar
  2. D. Zhang, M. M. Islam, and G. Lu, “A review on automatic image annotation techniques,” Pattern Recognition, vol. 45, no. 1, pp. 346–362, 2012. View at: Publisher Site | Google Scholar
  3. S. H. Amiri and M. Jamzad, “Automatic image annotation using semi-supervised generative modeling,” Pattern Recognition, vol. 48, no. 1, pp. 174–188, 2015. View at: Publisher Site | Google Scholar
  4. P. Duygulu, K. Barnard, J. F. G. de Freitas, and D. A. Forsyth, “Object recognition as machine translation: learning a lexicon for a fixed image vocabulary,” in Proceedings of the Computer Vision (ECCV 2002), pp. 97–112, Springer Berlin Heidelberg, Copenhagen, Denmark, May 2002. View at: Google Scholar
  5. J. Jeon, V. Lavrenko, and R. Manmatha, “Automatic image annotation and retrieval using cross-media relevance models,” in Proceedings of the 26th Annual International ACM SIGIR Conference on Research and Development in Informaion Retrieval, pp. 119–126, ACM, Ann Arbor, MI, USA, July 2003. View at: Google Scholar
  6. S. L. Feng, R. Manmatha, and V. Lavrenko, “Multiple Bernoulli relevance models for image and video annotation,” in Proceedings of the 2004 IEEE Computer Society Conference on Computer Vision & Pattern Recognition (CVPR 2004), vol. 2, pp. 1002–1009, Washington, DC, USA, June 2004. View at: Google Scholar
  7. M. Alkaoud, I. Ashshohail, and M. M. B. Ismail, “Automatic image annotation using fuzzy cross-media relevance models,” Journal of Image and Graphics, vol. 2, pp. 59–63, 2014. View at: Publisher Site | Google Scholar
  8. S. Zhu, S. Shen, and X. Li, “Multimodal deep network learning-based image annotation,” Electronics Letters, vol. 51, no. 12, pp. 905-906, 2015. View at: Publisher Site | Google Scholar
  9. Z. Zeng, H. Shi, Y. Wu, and Z. Quan, “Annotation-retrieval reinforcement by visual cognition modeling on manifold,” Neurocomputing, vol. 215, pp. 150–159, 2016. View at: Publisher Site | Google Scholar
  10. X. Ke, S. Li, and G. Chen, “Real web community based automatic image annotation,” Computers & Electrical Engineering, vol. 39, no. 3, pp. 945–956, 2013. View at: Publisher Site | Google Scholar
  11. Z.-X. Li, Z.-P. Shi, Z.-Q. Li, and Z.-Z. Shi, “Automatic image annotation by fusing semantic topics,” Journal of Software, vol. 22, no. 4, pp. 801–812, 2011. View at: Publisher Site | Google Scholar
  12. Z. Y. Qiu, Q. Fang, J. T. Sang, and C.-S. Xu, “Regional context-aware image annotation,” Chinese Journal of Computers, vol. 37, no. 6, pp. 1390–1397, 2014. View at: Google Scholar
  13. A. Bahrololoum and H. Nezamabadi-Pour, “A multi-expert based framework for automatic image annotation,” Pattern Recognition, vol. 61, pp. 169–184, 2017. View at: Publisher Site | Google Scholar
  14. A. Makadia, V. Pavlovic, and S. Kuma, “A new baseline for image annotation,” in Proceedings of the Computer Vision (ECCV 2008), pp. 316–329, Springer Berlin Heidelberg, Marseille, France, October 2008. View at: Google Scholar
  15. M. Guillaumin, T. Mensink, J. Verbeek, and C. Schmid, “Tagprop: discriminative metric learning in nearest neighbor models for image auto-annotation,” in Proceedings of the IEEE 12th International Conference on Computer Vision, pp. 309–316, Kyoto, Japan, September 2009. View at: Google Scholar
  16. Q. Cheng, Q. Zhang, P. Fu, C. Tu, and S. Li, “A survey and analysis on automatic image annotation,” Pattern Recognition, vol. 79, pp. 242–259, 2018. View at: Publisher Site | Google Scholar
  17. J. Deng, W. Dong, R. Socher, L. J. Li, K. Li, and F. F. Li, “Imagenet: a large-scale hierarchical image database,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2009), pp. 248–250, Miami, FL, USA, June 2009. View at: Google Scholar
  18. A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” in Proceedings of the Advances in Neural Information Processing Systems, pp. 1097–1105, Lake Tahoe, NV, USA, December 2012. View at: Google Scholar
  19. C. Szegedy, W. Liu, Y. Jia et al., in Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 1–9, Boston, MA, USA, June 2015.
  20. K. He, X. Zhang, S. Ren, and J. Sun, “Delving deep into rectifiers: surpassing human-level performance on Imagenet classification,” in Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), pp. 1026–1034, Santiago, Chile, December 2015. View at: Google Scholar
  21. D. H. Hubel and T. N. Wiesel, “Receptive fields and functional architecture of monkey striate cortex,” Journal of Physiology, vol. 195, no. 1, pp. 215–243, 1968. View at: Publisher Site | Google Scholar
  22. J. Bouvrie, “Notes on convolutional neural networks,” 2006, http://cogprints.org/5869/1/cnn_tutorial.pdf. View at: Google Scholar
  23. G. Griffin, A. Holub, and P. Perona, Caltech-256 Object Category Dataset, Caltech, Pasadena, CA, USA, 2007, Technical Report.
  24. M. Everingham, A. Zisserman, C. Williams, J. Wincc, and G. Luc, The PASCAL Visual Object Classes Challenge 2007 (VOC 2007) Results, PASCAL VOC Challenge, Kaggle, San Francisco, CA, USA, 2007, Technical Report.
  25. M. Everingham, L. Gool, and C. Williams, The PASCAL Visual Object Classes Challenge 2012 (VOC 2012) Results, PASCAL VOC Challenge, Kaggle, San Francisco, CA, USA, 2012, Technical Report, https://www.pascal-network.org/challenges/VOC/voc2012/workshop/index.html.
  26. Z. Zuo, G. Wang, B. Shuai, L. Zhao, and Q. Yang, “Exemplar based deep discriminative and shareable feature learning for scene image classification,” Pattern Recognition, vol. 48, no. 10, pp. 3004–3015, 2015. View at: Publisher Site | Google Scholar
  27. C. Zhang, J. Cheng, and Q. Tian, “Image-level classification by hierarchical structure learning with visual and semantic similarities,” Information Sciences, vol. 422, pp. 271–281, 2018. View at: Publisher Site | Google Scholar
  28. M. A. Islam, M. Rochan, N. D. B. Bruce, and Y. Wang, “Gated feedback refinement network for dense image labeling,” in Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 4877–4885, IEEE, Honolulu, HI, USA, July 2017. View at: Google Scholar
  29. L. X. Jaing and J. Hou, “Image annotation using the ensemble learning,” Acta Autamatic Sinica, vol. 38, no. 8, pp. 1257–1262, 2012. View at: Publisher Site | Google Scholar

Copyright © 2019 Jianfang Cao et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


More related articles

 PDF Download Citation Citation
 Download other formatsMore
 Order printed copiesOrder
Views5574
Downloads1125
Citations

Related articles