Cognitive Modeling of Multimodal Data Intensive Systems for Applications in Nature and Society (COMDICS)
View this Special IssueResearch Article  Open Access
Yan Guo, Jin Zhang, Chengxin Yin, Xiaonan Hu, Yu Zou, Zhipeng Xue, Wei Wang, "Plant Disease Identification Based on Deep Learning Algorithm in Smart Farming", Discrete Dynamics in Nature and Society, vol. 2020, Article ID 2479172, 11 pages, 2020. https://doi.org/10.1155/2020/2479172
Plant Disease Identification Based on Deep Learning Algorithm in Smart Farming
Abstract
The identification of plant disease is the premise of the prevention of plant disease efficiently and precisely in the complex environment. With the rapid development of the smart farming, the identification of plant disease becomes digitalized and datadriven, enabling advanced decision support, smart analyses, and planning. This paper proposes a mathematical model of plant disease detection and recognition based on deep learning, which improves accuracy, generality, and training efficiency. Firstly, the region proposal network (RPN) is utilized to recognize and localize the leaves in complex surroundings. Then, images segmented based on the results of RPN algorithm contain the feature of symptoms through Chan–Vese (CV) algorithm. Finally, the segmented leaves are input into the transfer learning model and trained by the dataset of diseased leaves under simple background. Furthermore, the model is examined with black rot, bacterial plaque, and rust diseases. The results show that the accuracy of the method is 83.57%, which is better than the traditional method, thus reducing the influence of disease on agricultural production and being favorable to sustainable development of agriculture. Therefore, the deep learning algorithm proposed in the paper is of great significance in intelligent agriculture, ecological protection, and agricultural production.
1. Introduction
Plant disease can directly lead to stunted growth causing bad effects on yields [1–3]. An economic loss of up to $20 billion per year is estimated all over the world [4–6]. Diverse conditions are the most difficult challenge for researchers due to the geographic differences that may hinder the accurate identification [7, 8]. In addition, traditional methods mainly rely on specialists, experience, and manuals [9], but the majority of them are expensive, timeconsuming, and laborintensive with difficulty detecting precisely [10]. Therefore, a rapid and accurate approach to identify plant diseases seems so urgent for the benefit of business and ecology to agriculture.
Internet technologies, in particular the availability of multimodality data from various sensors including the Internet of things and sensor networks, have developed rapidly [11]. Herein, a novel plant leaf identification model based on deep learning algorithm is designed to solve the above issues. The function contains leaf retrieval, image segmentation, and identification with the utilization of integrated deep learning algorithm throughout the whole process. The first task is leaf retrieval, but many factors pose the challenge of identification accuracy such as soil and illumination in the complex environment [12]. Hence, the model is investigated RPN algorithm for manipulating retrieval and represents the good adaption in practice. Image segmentation is the second step that is considered to be the most crucial because diagnostic precision plays an important role in detection results. The Chan–Vese algorithm based on region shows promising results for segmenting images free of noise and weak edge. The last step is to identify the disease of leaves based on the migration learning algorithm. Based on the pretrained model, the migration learning model uses the dataset of disease leaves in a simple background to train the model. The rest of the paper is constructed as follows: Section 2 previews other scholars’ researches thoroughly. The detailed information about the model is shown in Section 3. Section 4 demonstrates the procedure of experiment and study. Conclusions and discussions are in Section 5.
2. Literature Review
At present, the research of plant disease recognition in the complex environment mainly focuses on three aspects: disease leaves image segmentation, feature extraction, and disease identification.
2.1. Image Segmentation
In the complex environment, the most crucial task is how to segment the images while localizing and detecting diseased plant leaves, since the major aim of image segmentation is to set the symptom information apart from the background. There are many researchers making a deep investigation on it. In 2017, Ali et al. applied the Delta E color difference algorithm to separate the diseaseinfected area [13]. In general, four major methods are used to perform the image segmentation which are discussed the detail in the following paragraph [14].
Some researchers integrate the region of interest (ROI) and other methods to segment images. For example, Kao et al. claimed that the convolutional autoencoder served as the background filter to determine the ROI in an image [15]. The second method only concerns region segmentation. In 2013, Pujari et al. claimed that images were divided into various regions which had a special meaning and extracted the images’ feature [16]. Akram and other colleagues provided an image processing model with realtime synchronous processing. By dividing the image into three color spaces, it can carry out contrast stretching, feature vector, and salient region recognition [17]. In addition, other researchers chose deep learning techniques to segment and detect images. Marko et al. recommended a depthbased target detection algorithm and used the twostage algorithm to optimize plant disease images detection [18]. At last, the thresholding is common in segmentation. In 2018, Li et al. applied multilevel thresholding techniques based on gray histogram for image segmentation [19]. Mohamed and Diego presented a new multiobjective metaheuristic on the basis of a multiverse optimization algorithm to segment grayscale images via multilevel thresholding [20].
However, there is a fact that cannot be ignored. Because of the complexity of color information in the complicated environment, the machine vision algorithm based on color, ROI, and threshold performs poorly in practice.
2.2. Feature Extraction
The feature extraction of plant disease faces many problems in identifying plant disease. The distinct image features include textures, shape, color, and motionrelated attributes, which are the essential conditions for disease feature extraction [21, 22]. Raza and his colleagues described a method that uses color and texture features to extract disease spots [23]. Hu et al. proposed the Dempster–Shafer (DS) evidence theory and multifeature fusion for extracting features as well as the results were processed by introducing variance to improve decision rules of DS evidence theory [24]. In addition, Turkoglu depicted improved versions of the Local Binary Patterns (LBP) methodology, which uses the original LBP local quadratic value to transform the image into grayscale and processes the R and G channels of the image by considering overall and region [25]. Li et al. researched an IoT feature extraction for the intelligent city based on the deep migration learning model [26]. There was an application in music, which can extract meaningful audio features in order to enable the visualizations to be responsive to the music [27]. And in recent studies, lots of novel approaches have been put forward to implementing feature extraction. For example, concerning the challenging task that the extraction of relevant and distinct features from electroencephalogram (EEG), Meziani et al. proposed two new spectral estimators that were robust against nonGaussian, nonlinear, and nonstationary signals [28]. What is more, as Liu et al. reported, the highdimensional time–frequency spectrum features were extracted by using the residual neural network and the improved signaltoclutter ratio (SCR) [29]. Xu et al. introduced a feature extraction method based on the Hilbert marginal spectrum to perform the wear of milling tools [30].
2.3. Disease Identification
As for the precise identification, so many techniques are developed and researched to get accurate results. The identification model focused on using class labels for training images and built a finegrained image classification system [31]. Zhang et al. reported a recognition method for plant disease leaf images based on a hybrid clustering [32]. In 2017, Patil et al. described a contentbased image retrieval (CBIR) system to extract texture features and means value to compute color features, and support vector machine (SVM) classifier was used for classification [33]. Through above researches, the major goal was to design the classification schemes and image analysis for feature extraction and identification. Recently, other new approaches have been introduced to identify the disease more accurately and precisely. A novel system based on the selection of pictures and short text descriptions helped nonexperts in identifying plant diseases that can be used remotely from a desktop as well as in a smart phone or personal digital assistants [34]. Pertot et al. presented a scheme that used mobile phones for realtime onfield imaging of diseased plants and used mobile devices for leaf image segmentation and spotting of disease patch with improved kmeans clustering [35]. Yang et al. presented a microscopy image detection methodology based on the synergistic judgment of texture and shape features and the decision treeconfusion matrix [36].
Additionally, the convolutional neural network is numerously utilized in identifying diseases. Chad et al. established a system capable of automatically identifying plant disease in fieldacquired images of maize plants [37]. Ni et al. used the deep convolution neural network to train 1632 images of corn kernels and designed an automatic corn detector [38]. Lu et al. proposed a rice diseases identification method based on deep convolutional neural networks (CNNs) techniques [39]. Zhang et al. designed an agricultural machinery image recognition network using the deep learning algorithm [40]. Zhang et al. improved deep convolution neural network to improve the accuracy of maize leaf disease identification [41]. Images were input into two deep learningbased architectures, namely, AlexNet and VGG16 net, to perform detection [42]. Coulibaly et al. suggested a method using transfer learning for feature extraction to build an identification system [43]. However, due to the requirement for high hardware resources and traditional neural network models of high quality and quantity of data sets in the training process, the training wastes much time that is not conducive to the promotion and use of the model. In this paper, we recommend a transfer learning model for identification combined with the pretrained model, using the dataset of disease leaves to train the model.
From the above research findings, some achievements have been achieved in three aspects: leaf image segmentation, leaf lesion feature extraction, and leaf disease recognition. However, there are still many problems to be solved to realize plant disease identification in the complex environment.
3. Modeling
3.1. The Solution Framework
The full plant disease identification model framework based on deep learning is shown in Figure 1, including three steps, the localization of plant leaves, the segmentation of images, the extraction of plant disease, and the identification of disease. The model used in this paper mainly consists of the following three steps. The first step is to locate the diseased leaves. The RPN algorithm is used to train the leaf dataset in the complex environment, and the frame regression neural network and classification neural network are used to locate and retrieve the diseased leaves in the complex environment. The second step is the segmentation of diseased leaves. The Chan–Vese algorithm is used to segment the image of diseased leaves. Based on the set zero level set and the minimum energy function as the goal, the leaf contour is obtained by iterative calculation, so as to realize the image segmentation of diseased leaves in the complex environment. The third step is the identification of leaf disease species. The pretrained transfer learning model is trained to realize plant disease recognition in the simple background.
3.2. The Leaf Localization
Aiming at the localization of diseaseplant leaves, the paper manipulates the leaf dataset under complex background to train the RPN algorithm and integrates boundary regression neural network and classification neural network to perform localization and retrieval.
As for the classification neural network, the core task is to distinguish whether the image in the boundary box is an object or a background. During the process of training, making Intersection over Union (IoU) as a criterion of classification, the boundary box with IoU greater than 0.5 is annotated as an object and the boundary box with IoU less than 0.1 is labeled as a background. IoU is applied in calculating the relevance between predicting boundary box and artificial marked boundary box. The formula of IoU is shown as follows:where represents the overlap area of predicting boundary box and artificially marked boundary box, and represents the total area of it. Due to the fact that the classification neural network is only used for binary classification problem, sigmoid function is employed as loss function.
With regard to the adjustment parameters of boundary regression neural network and output boundary box, one boundary box can be represented by fourdimensional variable . represents the given boundary box, represents the target boundary box, and represents the predicting boundary box. In order to find a mapping relationship of boundary regression neural network, and are defined.
The movement of boundary consists of pan and zoom.
The parameter of pan is , given that and . The formula is shown as
The parameter of zoom is , given that and . The formula is shown as
According to the above formula, the real learning objectives of boundary regression neural network are represented with , and the real transform parameters between predicting boundary box and artificially marked boundary box are shown as .
The objective function of boundary regression neural network is , where represents the learning parameter of boundary regression neural network. The loss function is shown as follows:
3.3. The Leaf Image Segmentation
As for the segmentation of images in complex environment, based on the results of previous step, the model performs the segmentation by Chan–Vese algorithm. Laying the foundation of the set zero level set, aiming at minimizing the energy function and obtaining blade profiles by iterative computing, the model may perform the segmentation of diseased plant leaves images. The Chan–Vese algorithm uses the level set to constructing an energy function to constrain the whole region rather than to control surface evolution by the explicit control speed . The energy function is defined as the minimum sum of variances between the gray values of the image inside and outside the contour, and the contour length is increased to make it converge. Given a closed curve in the image, the energy function is expressed as follows:
represents inside the contour, represents outside the contour, represents the gray values of the image, represents the average gray values in contour, and represents the average gay values out contour. Then, the given formulas are as follows:
When and , the computing ends.
Level set method is used to solve (11) and zero level set is used to express contour lines. Heaviside’s function and Dirac’s function are introduced:
The level set equation of energy function is as follows:
By minimizing (16) with variational method and combining Euler–Lagrange equation, the following partial differential equations are obtained, where and .
3.4. The Diseased Leaf Identification
In identifying disease types, the paper utilizes the disease leaf dataset training model under simple background to train the pretrained transfer learning model. This method finishes training in a short period of time and performs the disease identification in the simple environment, reducing the requirement of deep learning algorithms for the hardware equipment. Due to the fact that shallow network has similar characteristics for various learning objects, the shallow neural network for source task can be transferred to the neural network for a target task by using the transfer learning algorithm. Transfer learning has better performance in convergences and ultimate results than new learning in practice.
Setting the domain as , it includes two contents, where represents feature space and contains all possible characteristic values. represents a specific feature sampling instance in the feature space:
Setting the task as , it includes two parts, where represents label space, that is, all vector space consisting of all tags. As prediction function, is obtained by learning from the features and labels of input data:
4. Experimental Study
4.1. The Acquisition of Data
First of all, this study needs to obtain the leaf dataset in the complex environment. The paper employs Crawler technology and obtains 1000 leaf photos from the Plant Photo Bank of China (PPBC), including the leaves of various plants at each growth stage. The shapes of these leaves are different, and the health of the leaves is also different. Aiming at watermark less shelter, obvious leaves, and easy labeling, 189 images are screened out as leaf photos in the complex environment. This dataset is used to train the RPN algorithm to detect and locate the leaf in the complex environment.
Then, using LabelImg, images are quickly annotated and XML files are generated in PASCAL VOC format. It can directly input the target detection neural network as training data. Finally, this study needs to obtain the dataset of diseased leaves in the simple environment. This paper downloads four kinds of images of black rot, bacterial plaque, rust, and healthy leaves from PlantVillage Agricultural QuestionandAnswer Forum as training data of transfer learning model, including 537 black rot, 1032 bacterial plaque disease, 293 rust, and 2852 healthy leaves. This dataset is used to train the transfer learning model.
4.2. The Parameter Setup
4.2.1. The Parameter Setup of Leaf Localization
The parameter setup of classification neural network and boundary regression neural network is shown in Tables 1 and 2. The anchors in the table represent the number of candidate boxes generated.


4.2.2. The Parameter Setup of Leaf Segmentation
The main parameter setup of Chan–Vese algorithm is initial zero level set and iteration number setting. In this paper, we set the initial zero level set as a circle with the center of the picture and onethird of the diagonal length of the picture as the radius and set up the Chan–Vese algorithm to calculate 500 iterations.
The image obtained by RPN algorithm is input into the Chan–Vese algorithm and the image in the zero level set is preserved. The image outside the zero level set is set to black to get the image segmentation result.
4.2.3. The Parameter Setup of Leaf Retrieval
Resnet101 is selected as the pretraining model, and the network is trained by using the dataset of disease leaves under a simple background in this paper. Its network parameters are shown in Table 3.

In this paper, all its parameters are modified and initialized in the last output layer of Resnet101, and the classification number is changed from 1000 to 4, which corresponds to the identification results of four kinds of leaf diseases.
5. Results
5.1. The Result of Leaf Localization
The test image is input into VGG16 model and RPN algorithm, and the results are shown in Figure 2.
(a)
(b)
(c)
(d)
With regard to the above images, there is inaccuracy of the frame selection range in Figure 2(b) and the blades in Figure 2(d) are missing. But RPN algorithm can basically frame the main blade structure, which has better performance than the original model.
5.2. The Result of Leaf Segmentation
The iterative calculation process of Chan–Vese algorithm is shown in Figures 3–6. The results of Chan–Vese algorithm compared with watershed algorithm are shown in Figure 7.
(a)
(b)
(c)
(d)
(a)
(b)
(c)
(d)
(a)
(b)
(c)
(d)
(a)
(b)
(c)
(d)
(a)
(b)
(c)
(d)
According to the above results, it can be found that after 500 iterations, Chan–Vese algorithm can get better leaf image segmentation results. Although Chan–Vese algorithm cannot effectively extract the edge contour of the blade compared with the watershed algorithm, it retains the complete structure of the central blade including leaf venation, spot color, and spot shape. The complete central structure of the blade obtained by Chan–Vese algorithm can be used for disease identification of the next step.
5.3. The Result of Disease Identification
The test image is input into VGG16 model and RPN algorithm, and the results are shown in Figure 8.
(a)
(b)
(c)
(d)
According to the above pictures, although it can be found that the frame selection range in Figure 8(b) is not accurate enough and there are omissions in Figure 8(d), RPN algorithm can basically frame the main blade structure and can be used for the next operation.
5.4. Comparative Test
In this paper, the parameters of the transfer learning model include gradient descent optimization parameters and training parameters. The specific parameters are set as shown in Table 4.

After 4000 iterations of training, the loss value and training set accuracy of the transfer learning model and the traditional model are shown in Figure 9.
(a)
(b)
In Figure 9(a), ResNet101 represents traditional model. According to Figure 9(a), after the same 4000 iterations training, the transfer learning model has faster convergence speed and lower model loss value than the traditional model. According to Figure 9(b), it can be found that in the process of model training, transfer learning has higher accuracy, lower variance, and better recognition effect than new learning.
Therefore, compared with the new learning, this paper uses the transfer learning to converge faster and achieve better model identification effect. It can meet the requirements of smart agriculture for low hardware resources, fast training time, and high training efficiency.
Then, the image is input into transfer learning model based on the segmentation of Chan–Vese algorithm. As a contrast, the image that has not been processed in this paper is input into the traditional ResNet101 model for identification, and the results are shown in Table 5.

According to the comparison results in the above table, the average correct rate of the proposed method is 83.75%, which is significantly better than that of the traditional ResNet101 model (42.5%). Comparing the performances of this method in four samples, we can find that rust and healthy leaves can get better results than black rot and bacterial plaque.
6. Discussion and Conclusion
This paper shows that the plant disease recognition model based on deep learning has the characteristics of unsupervised, high accuracy, good universality, and high training efficiency. However, there are many challenges in accuracy practicability of plant disease detection in the complex environment. In order to solve these problems and optimize the identification method, this paper proposes a recognition model integrating RPN algorithm, CV algorithm, and TL algorithm, which can effectively solve the problem of plant disease identification in the complex environment. The model not only adapts to complex environments, but also increases the accuracy of identification. Compared with the traditional model, the model proposed in this paper not only guarantees the robustness of the convolutional neural network, but also reduces the number and quality requirements of the convolutional neural network on the data set and obtains better results. Therefore, the model could help agricultural production personnel to prevent and cure the plant disease quickly. The model which overcomes the problem of environment complexity can get an accurate identification result in practical application. Furthermore, this study enriches the existing theory and helps to improve the accuracy. At the same time, it is of great significance for the study of plant disease identification in the field of environmental complexity and helps researchers pay attention to the important role of environmental complexity in plant disease identification. Therefore, the model applies information technology to agricultural production and is favorable to sustainable development of smart agriculture.
Although the plant disease identification model based on deep learning proposed in this paper can overcome the complexity of the environment and improve the accuracy of identification, there are still some problems to be pointed out. For example, the Chan–Vese algorithm needs repetitive iterative calculation and runs for a long time, which is not conducive to the fast identification results of this method. In future research, we will use the neural network to generate zero initial set corresponding to different leaves, which will increase the end of calculation limit for the iterative process of Chan–Vese algorithm, speed up the training speed, and end the iteration ahead of time.
Data Availability
Data were made available with the help of the Key Laboratory of Agricultural Information Engineering of Sichuan Province. The data used to support the findings of this study were provided by the laboratory under license. Access to these data will be considered by the corresponding author upon request, with permission of the laboratory.
Conflicts of Interest
The authors declare that there are no conflicts of interest regarding the publication of this paper.
Acknowledgments
This study was supported by the Key Laboratory of Agricultural Information Engineering of Sichuan Province and Social Science Foundation of Sichuan Province in 2019 (19GL030).
References
 Y. Ampatzidis, L. De Bellis, and A. Luvisi, “iPathology: robotic applications and management of plants and plant diseases,” Sustainability, vol. 9, no. 6, p. 1010, 2017. View at: Publisher Site  Google Scholar
 A. Breukers, D. L. Kettenis, M. Mourits, W. V. D. Werf, and A. O. Lansink, “Individualbased models in the analysis of disease transmission in plant production chains: an application to potato brown rot,” Academy of Sciences, vol. 90, no. 1–3, pp. 112–131, 2006. View at: Publisher Site  Google Scholar
 S. Ghosal, D. Blystone, A. K. Singh, B. Ganapathysubramanian, A. Singh, and S. Sarkar, “An explainable deep machine vision framework for plant stress phenotyping,” Proceedings of the National Academy of Sciences, vol. 115, no. 18, pp. 4613–4618, 2018. View at: Publisher Site  Google Scholar
 E.C. Oerke, “Crop losses to pests,” The Journal of Agricultural Science, vol. 144, no. 1, pp. 31–43, 2006. View at: Publisher Site  Google Scholar
 X. E. Pantazi, D. Moshou, and A. A. Tamouridou, “Automated leaf disease detection in different crop species through image features analysis and One Class Classifiers,” Computers and Electronics in Agriculture, vol. 156, pp. 96–104, 2019. View at: Publisher Site  Google Scholar
 J. G. A. Barbedo, “Factors influencing the use of deep learning for plant disease recognition,” Biosystems Engineering, vol. 172, pp. 84–91, 2018. View at: Publisher Site  Google Scholar
 G. Geetharamani and J. Arun Pandian, “Identification of plant leaf diseases using a ninelayer deep convolutional neural network,” Computers & Electrical Engineering, vol. 76, pp. 323–338, 2019. View at: Publisher Site  Google Scholar
 P. F. Konstantinos, “Deep learning models for plant disease detection and diagnosis,” Computers & Electrical Engineering, vol. 145, pp. 311–318, 2018. View at: Google Scholar
 V. Singh and A. K. Misra, “Detection of plant leaf diseases using image segmentation and soft computing techniques,” Information Processing in Agriculture, vol. 4, no. 1, pp. 41–49, 2017. View at: Publisher Site  Google Scholar
 S. P. Mohanty, D. P. Hughes, and S. Marcel, “Using deep learning for imagebased plant disease detection,” Frontiers in Plant Science, vol. 7, p. 1419, 2016. View at: Publisher Site  Google Scholar
 Y. Guo, X. Hu, Y. Zou et al., “Maximizing Etailers’ sales volume through the shippingfee discount and product recommendation system,” Discrete Dynamics in Nature and Society, vol. 2020, pp. 1–14, 2020. View at: Google Scholar
 R. Amanda, B. Kelsee, M. C. Peter, A. Babuali, L. James, and D. P. Hughes, “Deep learning for imagebased cassava disease detection,” Frontiers in Plant Science, vol. 8, p. 1852, 2017. View at: Google Scholar
 H. Ali, M. I. Lali, M. Z. Nawaz, M. Sharif, and B. A. Saleem, “Symptom based automated detection of citrus diseases using color histogram and textural descriptors,” Computers and Electronics in Agriculture, vol. 138, pp. 92–104, 2017. View at: Publisher Site  Google Scholar
 H. M. Alexander, K. E. Mauck, A. E. Whitfield, K. A. Garrett, and C. M. Malmstrom, “Plantvirus interactions and the agroecological interface,” European Journal of Plant Pathology, vol. 138, no. 3, pp. 529–547, 2014. View at: Publisher Site  Google Scholar
 I.H. Kao, Y.W. Hsu, Y.Z. Yang, Y.L. Chen, Y.H. Lai, and J.W. Perng, “Determination of Lycopersicon maturity using convolutional autoencoders,” Scientia Horticulturae, vol. 256, p. 108538, 2019. View at: Publisher Site  Google Scholar
 D. Pujari, R. Yakkundimath, and A. S. Byadgi, “Grading and classification of anthracnose fungal disease of fruits based on statistical texture features,” International Journal of Advanced Science and Technology, vol. 52, pp. 121–132, 2013. View at: Google Scholar
 T. Akram, S. R. Naqvi, S. A. Haider, and M. Kamran, “Towards realtime crops surveillance for disease classification: exploiting parallelism in computer vision,” Computers & Electrical Engineering, vol. 59, pp. 15–26, 2017. View at: Publisher Site  Google Scholar
 A. Marko, K. Mirjana, S. Srdjan, A. Andras, and S. Darko, “Solving current limitations of deep learning based approaches for plant disease detection,” SymmetryBaseline, vol. 11, no. 7, p. 939, 2019. View at: Google Scholar
 J. Li, W. Tang, J. Wang, X. Wang, and X. Zhang, “Multilevel thresholding selection based on variational mode decomposition for image segmentation,” Signal Processing, vol. 147, pp. 80–91, 2018. View at: Publisher Site  Google Scholar
 M. A. Elaziz, D. Oliva, A. A. Ewees, and S. Xiong, “Multilevel thresholdingbased grey scale image segmentation using multiobjective multiverse optimizer,” Expert Systems with Applications, vol. 125, pp. 112–129, 2019. View at: Publisher Site  Google Scholar
 A. Cruz, Y. Ampatzidis, R. Pierro et al., “Detection of grapevine yellows symptoms in Vitis vinifera L. with artificial intelligence,” Computers and Electronics in Agriculture, vol. 157, pp. 63–76, 2019. View at: Publisher Site  Google Scholar
 Z. Iqbal, M. A. Khan, M. Sharif, J. H. Shah, M. H. ur Rehman, and K. Javed, “An automated detection and classification of citrus plant diseases using image processing techniques: a review,” Computers and Electronics in Agriculture, vol. 153, pp. 12–32, 2018. View at: Publisher Site  Google Scholar
 M. Raza, M. Sharif, M. Yasmin, M. A. Khan, T. Saba, and S. L. Fernandes, “Appearance based pedestrians' gender recognition by employing stacked auto encoders in deep learning,” Future Generation Computer Systems, vol. 88, pp. 28–39, 2018. View at: Publisher Site  Google Scholar
 M. Hu, X. Bu, X. Sun, Z. Yu, and Y. Zheng, “Rape plant disease recognition method of multifeature fusion based on DS evidence theory,” Mathematical and Computational Applications, vol. 22, no. 1, p. 18, 2017. View at: Publisher Site  Google Scholar
 M. Turkoglu and D. Hanbay, “Leafbased plant species recognition based on improved local binary pattern and extreme learning machine,” Physica A: Statistical Mechanics and Its Applications, vol. 527, p. 121297, 2019. View at: Publisher Site  Google Scholar
 D. Li, L. Deng, M. Lee, and H. Wang, “IoT data feature extraction and intrusion detection system for smart cities based on deep migration learning,” International Journal of Information Management, vol. 49, pp. 533–545, 2019. View at: Publisher Site  Google Scholar
 Dhiraj, R. Biswas, and N. Ghattamaraju, “An effective analysis of deep learning based approaches for audio based feature extraction and its visualization,” Multimedia Tools and Applications, vol. 78, no. 17, pp. 23949–23972, 2019. View at: Publisher Site  Google Scholar
 A. Meziani, K. Djouani, T. Medkour, and A. Chibani, “A Lasso quantile periodogram based feature extraction for EEGbased motor imagery,” Journal of Neuroscience Methods, vol. 328, p. 108434, 2019. View at: Publisher Site  Google Scholar
 Y. Xu, H. Ding, Y. Xue, and J. Guan, “Highdimensional feature extraction of sea clutter and target signal for intelligent maritime monitoring network,” Comput. Commun., vol. 147, pp. 76–84, 2019. View at: Google Scholar
 C. Xu, Y. Chai, H. Li, Z. Shi, L. Zhang, and Z. Liang, “A feature extraction method for the wear of milling tools based on the Hilbert marginal spectrum,” Machining Science and Technology, vol. 23, pp. 847–868, 2019. View at: Google Scholar
 Y. Zhang, X.S. Wei, J. Wu et al., “Weakly supervised finegrained categorization with partbased image representation,” IEEE Transactions on Image Processing, vol. 25, no. 4, pp. 1713–1725, 2016. View at: Publisher Site  Google Scholar
 S. Zhang, Z. You, and X. Wu, “Plant disease leaf image segmentation based on superpixel clustering and EM algorithm,” Neural Computing and Applications, vol. 31, no. S2, pp. 1225–1232, 2019. View at: Publisher Site  Google Scholar
 J. K. Patil and R. Kumar, “Analysis of content based image retrieval for plant leaf diseases using color, shape and texture features,” Engineering in Agriculture, Environment and Food, vol. 10, pp. 69–78, 2016. View at: Google Scholar
 K. P. Ferentinos, “Deep learning models for plant disease detection and diagnosis,” Computers and Electronics in Agriculture, vol. 145, pp. 311–318, 2018. View at: Publisher Site  Google Scholar
 I. Pertot, T. Kuflik, I. Gordon, S. Freeman, and Y. Elad, “Identificator: a webbased tool for visual plant disease identification, a proof of concept with a case study on strawberry,” Computers and Electronics in Agriculture, vol. 84, pp. 144–154, 2012. View at: Publisher Site  Google Scholar
 N. Yang, Y. Qian, H. S. EL‐Mesery, R. Zhang, A. Wang, and J. Tang, “Rapid detection of rice disease using microscopy image identification based on the synergistic judgment of texture and shape features and decision treeconfusion matrix method,” Journal of the Science of Food and Agriculture, vol. 99, no. 14, pp. 6589–6600, 2019. View at: Publisher Site  Google Scholar
 D. Chad, W. H. Tyr, S. Chen et al., “Automated identification of northern leaf blightinfected maize plants from field imagery using deep learning,” Phytopathology, vol. 107, pp. 1426–1432, 2017. View at: Google Scholar
 C. Ni, D. Wang, R. Vinson, M. Holmes, and Y. Tao, “Automatic inspection machine for maize kernels based on deep convolutional neural networks,” Biosystems Engineering, vol. 178, pp. 131–144, 2019. View at: Publisher Site  Google Scholar
 Y. Lu, S. Yi, N. Zeng, Y. Liu, and Y. Zhang, “Identification of rice diseases using deep convolutional neural networks,” Neurocomputing, vol. 267, pp. 378–384, 2017. View at: Publisher Site  Google Scholar
 Z. Zhang, H. Liu, Z. Meng, and J. Chen, “Deep learningbased automatic recognition network of agricultural machinery images,” Computers and Electronics in Agriculture, vol. 166, p. 104978, 2019. View at: Publisher Site  Google Scholar
 X. Zhang, Y. Qiao, F. Meng, C. Fan, and M. Zhang, “Identification of maize leaf diseases using improved deep convolutional neural networks,” IEEE Access, vol. 6, pp. 30370–30377, 2018. View at: Publisher Site  Google Scholar
 R. A. Krishnaswamy, P. Raja, and R. Aniirudh, “Tomato crop disease classification using pretrained deep learning algorithm,” Procedia Computer Science, vol. 133, pp. 1040–1047, 2018. View at: Google Scholar
 S. Coulibaly, B. KamsuFoguem, D. Kamissoko, and D. Traore, “Deep neural networks with transfer learning in millet crop images,” Computers in Industry, vol. 108, pp. 115–120, 2019. View at: Publisher Site  Google Scholar
Copyright
Copyright © 2020 Yan Guo et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.