Table of Contents Author Guidelines Submit a Manuscript
The Scientific World Journal
Volume 2014, Article ID 214674, 13 pages
Research Article

Integrating SOMs and a Bayesian Classifier for Segmenting Diseased Plants in Uncontrolled Environments

1ITESM, Autopista del Sol, 62790 Xochitepec, MOR, Mexico
2Centro de Investigación en Biotecnología/Laboratorio de Investigaciones Ambientales, Universidad Autónoma del Estado de Morelos, Avenida Universidad 1001, 62209 Cuernavaca, MOR, Mexico

Received 18 July 2014; Revised 5 October 2014; Accepted 6 October 2014; Published 4 November 2014

Academic Editor: Gonzalo Pajares

Copyright © 2014 Deny Lizbeth Hernández-Rabadán et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


This work presents a methodology that integrates a nonsupervised learning approach (self-organizing map (SOM)) and a supervised one (a Bayesian classifier) for segmenting diseased plants that grow in uncontrolled environments such as greenhouses, wherein the lack of control of illumination and presence of background bring about serious drawbacks. During the training phase two SOMs are used: one that creates color groups of images, which are classified into two groups using -means and labeled as vegetation and nonvegetation by using rules, and a second SOM that corrects classification errors made by the first SOM. Two color histograms are generated from the two color classes and used to estimate the conditional probabilities of the Bayesian classifier. During the testing phase an input image is segmented by the Bayesian classifier and then it is converted into a binary image, wherein contours are extracted and analyzed to recover diseased areas that were incorrectly classified as nonvegetation. The experimental results using the proposed methodology showed better performance than two of the most used color index methods.

1. Introduction

Modern agriculture is applying new scientific and technological processes to automate and provide more accurate and appropriate solutions to real problems in agricultural systems [1]. Machine vision represents one of these technological achievements, which is being integrated into the agricultural field because it is considered as accurate and nondestructive and yields consistent results [2]. Machine vision is a feasible sensing technique for plant specific direct applications (PSDA) not only due to its superior spatial resolutions [3], but also for providing numerical attributes of the objects or the scene being imaged [2]. This technology has been developed for many agricultural applications, such as detection of diseases and pests in plants [410], weed detection [3, 1113], and plant species identification [1416], among others. In particular, the development of vegetation segmentation algorithms from images is a fundamental and complex process in agricultural applications, because it is highly dependent on environmental conditions, which can be controlled or uncontrolled.

Previous studies under controlled environments have proposed a variety of approaches for segmenting diseased areas of leaves before their recognition. Most of them consider a uniform color background and controlled illumination, and, in addition, a single leaf of the plant is commonly analyzed, without considering the overlapping of leaves and the effects caused by the illumination such as shadows, brightness, and highlights, among others [14, 1719]. The segmentation process of these captured images under such conditions represents nonrelevant complexity and yields efficient results. Many algorithms have been developed for segmenting vegetation, without considering areas of diseased vegetation, from images under uncontrolled environments such as agricultural fields or greenhouses. Such algorithms use commonly color vegetation indices to segment a plant from the background of field images [20], such as the color index of vegetation extraction (CIVE) [21] and the excess green minus excess red (ExG-ExR) [22]; besides color indices, thresholds techniques have also been employed applying, for example, a fixed threshold [23] or the Otsu method [24]; in [25] a more sophisticated algorithm using a mean-shift procedure and a back-propagation neural network (BPNN) was presented and yielded better results than the two methods based on color indices known as CIVE and ExG; Jeon et al. [3] presented an algorithm for segmenting vegetation from a normalized excessive-green image using an adaptive threshold which reported efficient results even under different illuminations.

Owing to the problems mentioned above, not much research has been done on segmenting diseased vegetation in uncontrolled environments. In order to cope with the complexities of the environment and improve vegetation segmentation rates, some methodologies have been proposed. One of them is a methodology combining an unsupervised and a supervised learning method that was proposed by Meunkaewjinda et al. [4] for disease detection in grape leaves, where the grape leaf color was extracted from the background by applying a self-organizing feature map (SOFM) and a back-propagation neural network (BPNN). Even though this work yielded very promising performance, some ambiguous color pixels from the background or the vegetation of the image were incorrectly classified. This problem was due to the fact that the developed algorithms only used the color as input feature, thus reducing the effectiveness of the segmentation algorithms for images captured in natural scenes where objects and the background often exhibit common intensities.

Our research has been developed for tomato greenhouse environments whose illumination and background parameters are considered to be under uncontrolled conditions. Therefore, the task of segmenting vegetation, plants with diseased leaves in our case, is a serious complex task to be carried out.

In this work, the images of tomato plants were captured by a low cost digital camera which produced poor quality images. In the methodology proposed in this paper (hereafter called SEVUE, which stands for segmenting vegetation in uncontrolled environments) the following main processes were developed: an image enhancement process to improve both the quality of the images and the highlight features of interest; a color clustering process using two SOMs, wherein the second SOM refines the results of the first SOM; a classification process using a Bayesian classifier; and the extraction and analysis of contours from a binary image to recover areas that were incorrectly classified as nonvegetation and eliminate noise.

The SEVUE methodology was implemented and compared to the color index methods more commonly used for segmenting vegetation under uncontrolled environments (CIVE and ExG-ExR) using the Wilcoxon signed rank sum test to assess the algorithm performance. The remainder of this paper is organized as follows. In Section 2 the SEVUE methodology is described. The procedure used to assess the algorithm performance is presented in Section 3. A discussion of the results is made in Section 4, and finally the extracted conclusions and future works are exposed in Section 5.

2. SEVUE Methodology

A set of 138 images of tomato plants were captured by a Kodak EasyShare 913 camera with features corresponding to automatic mode. The image resolution of 320 × 240 pixels was selected to minimize image processing efforts. 30 out of 138 captured images were selected by experts who considered only those captured images of plants showing leaves with early visual symptoms of powdery mildew. Some considerations were made during the image capture sessions, such as the sessions were regularly carried out around midday without considering the climatic conditions (sunny, partly, cloudy, and overcast); the images were captured manually by a person standing in front of and pointing the camera at the vegetation, at a height of approximately 110 cm from the ground and holding the camera at an angle of approximately ; no polarizing filter was attached to the camera.

The SEVUE methodology was tested over the set of 30 images. Due to small statistics of individuals, we applied the -fold cross-validation test aiming at gaining a reasonable estimation of SEVUE methodology accuracy. Figure 1 illustrates a block diagram that shows the main processes of SEVUE methodology. The next sections will describe in detail the processes of SEVUE by following Figure 1.

Figure 1: Blocks diagram of the SEVUE methodology. Two phases are performed: training and testing phases. The main processes of SEVUE are the following: an image enhancement to adjust the color and brightness levels of the input images; a color clustering process using SOM1 and SOM2 (SOM2 refines the results of SOM1); a classification process by a Bayesian classifier which segments vegetation from input images; and finally a process for recovering areas incorrectly classified as nonvegetation, which at the output will yield the segmented image.

2.1. Image Enhancement

As a result of the different illumination conditions, the characteristics of light reflected by the objects in the images brought about problems related to different colors and brightness levels. To cope with such problems, it was necessary to adjust the levels of these parameters to homogenize the input images [26], thus achieving a better segmentation of vegetation. For this purpose, the actions described below were applied for each input image.A reference image was chosen showing less visual effects due to variable illumination, such as color inconstancy, highlights, shadows, and uneven lighting. The reference image was selected through a visual inspection.The reference image and the input image were converted from RGB to YIQ color model.The color and brightness of the input image were adjusted using the equations described as follows: where is the intensity value for each channel from the input image; is the mean value for each channel from the input image; is the mean value for each channel from the reference image with a matrix of size ;    is the new image with an adjustment for both the color and brightness to each channel .The image is reconverted from YIQ to RGB color model.

The resulting image after the enhancement process is an RGB image (see Figure 2). RGB color model is not a good choice for color image processing because it is highly correlated [27], for this reason in the training phase, a color transformation from the RGB color model into HSV and CIE color models are applied to the enhanced image. HSV and CIE are nonlinear color models and are less sensitive to variations of illumination intensity [4]; consequently a more detachable feature space is produced.

Figure 2: Image enhancement. (a) Original image and (b) image with color and brightness correction.
2.2. SOM Training

A color clustering process is carried out with the set of training images by applying two SOMs. SOM is regarded as a specific region-based image segmentation technique by [28]. SOM combined with -means has achieved better segmentation results in natural color image segmentation compared to JSEG, which is one of the most used techniques for segmentation based on regions of images taken under natural (uncontrolled) conditions [29]. The SOM training generates a map with output neurons, where each neuron represents a color group. The SOM training algorithm applied in SEVUE methodology was proposed by Wu et al. [30], wherein schemes of growing, pruning, and merging of neurons are implemented to find an appropriate number of output neurons automatically. Some considerations were made during the SOM training.(i) The first SOM considered as input values (features) to the map the and components from HSV and CIE color model for each pixel of the image. The components and were chosen because they can distinguish between vegetation and background color as is demonstrated in [26].(ii) The SOM training algorithm is initialized with a number of neurons , which is in accordance with two clusters. The weight vectors for each neuron are initialized with the two first input vectors .(iii) The training algorithm considered a maximum number of neurons to be created, which attempts to represent the color groups on the image effectively. A trial-and-error process was made and best results were observed when the maximum number was defined with a value of 9.(iv) The winner competition neuron was found applying an heuristic, which is described as follows: where is the output value of the th neuron with a weight vector ; is a threshold with a value of chosen by a trial-and-error process; is an input vector; is the similarity measure between the input vector and the weight vector; in this work the Euclidean distance was considered as measurement. A new neuron is created when (growing scheme). The input vector is assigned as the weight vector of the new neuron.(v) The thresholds to decide when to create (growing scheme), eliminate (pruning scheme), and merge neurons were determined by a trial-and-error process. Best results were obtained with a value of 10 for growing and 5 for merging. The elimination of neurons, pruning scheme, was done by deleting that neurons rarely winner (low density) with a frequency less than a predefined percentage, which is decremented in one for each SOM iteration. The initial percentage was defined with a value of 15%.

After the training of the SOM, the set of training images were classified into the 9 color groups (see Figure 3) by assigning each pixel of the images to the color group that corresponds with the minimum Euclidean distance. This process is a presegmentation of the training images based on regions because it consists in grouping pixels of similar intensity levels. This process creates 9 new images for each training image. As observed in Figure 3, the images created for the color groups with a high value in its color component (see Table 1) grouped the green color (vegetation) and those with low values grouped the background color (nonvegetation). The labeling process as vegetation or nonvegetation of the set of new images generated by each color group was possible by using -means. -means clustered into two clusters the color groups by applying the Euclidean square distance as measurement. Then, these clusters were labeled as vegetation or nonvegetation in accordance with their value in the color component by applying a predefined rule (3) as follows: where is the cluster and (the two result clusters by -means); Max is the maximum value of the channel for the two clusters. Table 1 shows the results for the 10 iterations () of cross-validation where it is observed that mean values for the vegetation cluster correspond to for , respectively.

Table 1: Weights of the color groups.
Figure 3: Images generated for the different color groups from an input image. The presented color groups correspond to 2nd iteration of cross-validation.

After classification, it was observed that some color groups labeled as vegetation incorrectly classified background color within its group, or conversely, in groups labeled as nonvegetation some disease or vegetation areas were incorrectly classified. Generally, these classifications errors are presented because of noise information with similar tonalities. For example, in Figure 3 the image corresponding to the color group 3 grouped regions of moldy ground due to similarity with green vegetation. To address these misclassifications, we train a second SOM with the images generated for that color group with the objective of separating those areas and integrating them into their corresponding color group, vegetation or nonvegetation. Through visual inspections to the set of images, it was observed that the color groups with misclassifications presented values lower than the mean value for vegetation . Considering those observations, it was possible to implement a procedure to select automatically the color group of nonvegetation or vegetation to be trained again.Search in the vegetation color groups , where is the color group, , and is the fold in the cross-validation .If some color group is selected, thenthe color group with the maximum Euclidean distance to is chosen to be trained,search in the rest of vegetation color groups where and relabel them as nonvegetation.elsesearch in the nonvegetation color groups if some color group is selected, thenthe color groups with the minimum Euclidean distance to is chosen to be trained.

This procedure is repeated for each one of the 10-fold. The color groups selected by the procedure are highlighted in Table 1. For example, for the color group was selected to be trained again and the color group was relabeled as nonvegetation. All images generated for the selected color group assume the role of input to the second SOM training. The training considerations for the second map are the same as presented at the beginning of this section, except that, in this case, we considered the channel from the HSV color model and from the CIE color model as the input values to the map because they showed better discrimination between vegetation and background with similar tonalities to vegetation (moldy ground). The and color components were chosen based on the observations of the resulting images due to the fact that it was not possible to quantify the segmentation performance between color components because the manual segmentation of the background was not considered. As final step of the color clustering process, all the color groups images labeled as vegetation are got together to create a new set of images (see Figure 4(a)). The same process is applied to the color group images labeled as nonvegetation (see Figure 4(b)). Therefore the new set of images consists of 30 vegetation images and 30 nonvegetation images which will be used as knowledge by the supervised classifier (Bayesian classifier) for vegetation segmentation purposes. The next section describes details about the Bayesian classifier.

Figure 4: The resultant images by joining the different color groups. (a) The resultant image for the color groups labeled as vegetation and (b) the resultant image for the color groups labeled as nonvegetation.
2.3. Image Segmentation by the Bayesian Classifier

The vegetation segmentation process is carried out by a Bayesian classifier, performing a pixel level classification of the input images for labeling each pixel as vegetation or nonvegetation. Bayesian reasoning is based on the assumption that optimal decisions can be made by relating probability distributions with observed data. When these probabilities are not known a priori they are often estimated by a training process, where examples are incorporated into an algorithm and a supervisor determines the different classes. The set of vegetation and background images created by the SOM clustering process are considered as examples for training the classifier; in our case the classes were determined automatically by the SEVUE methodology. The conditional probabilities are estimated by creating two color histogram models (vegetation, nonvegetation) from the examples images in the training process. The histogram represents the relative frequency of each combination in the image. For classification purposes, the histogram counts are converted into discrete probability distributions as follows: where represents the count in the histogram bin associated with the color combination and is the total count obtained by summing the counts in all of the bins. We formed the histograms with a bins number of . Given vegetation and nonvegetation histograms, we can compute the probability that a given color value (a combination) belongs to the vegetation and nonvegetation classes using the Bayes theorem, described as where the conditional probabilities and and a priori probabilities and are directly computed from the vegetation and nonvegetation histograms, respectively, where is the pixel count contained in bin of the vegetation histogram, is the equivalent count from the nonvegetation histogram, and and are the total counts contained in the vegetation and nonvegetation histograms, respectively.

A pixel is classified as vegetation if where is a threshold. Alternatively, a faster way is to apply the following rule

2.4. Recovering Areas Incorrectly Classified as Nonvegetation

In applications of vegetation segmentation in natural and complex environments, the color has yielded efficient results as shown by methods such as CIVE [21] and excess-green [22], among others. However, some ambiguities related to color pixels may arise, due to similarities between the disease color and the background color [4], when the segmentation of diseased vegetation is being carried out. In this work, a color ambiguity problem appears in the images because of color similarities between the background and the visual symptoms of powdery mildew disease observed in the leaves. After applying the Bayesian classification, it was observed that some lesion areas of the images were labeled as nonvegetation (see Figure 5), which is due to the fact that the similarity of color tones causes the SOM to integrate both the background and the disease areas into the same color group.

Figure 5: (a) Original image and (b) same image after applying Bayesian classifier.

This work proposes the extraction and analysis of contours from the binary image generated after the Bayesian classification for recovering areas incorrectly classified as nonvegetation. The algorithm proposed in [31] was used for the contour extraction. The basic concept of this algorithm consists of tracking the edges by considering a topological analysis. Therefore, the relationship between inner and outer edges is extracted.

The most important contours for the purpose of this work were those that represent the inner edges. Once contours were extracted, the next step is to select and analyze the contours whose measurement in pixels is less than a predefined threshold. The process is described in Figure 6.

Figure 6: A process that extracts and selects contours which encloses disease areas. (a) Image segmented with the Bayesian classifier, (b) binary image, (c) contour extraction, and (d) recovering of contours corresponding to vegetation/disease areas and noise elimination.

Powdery mildew is manifested in its initial stage with small whitish spots on leaves. For this reason, it was considered that small contours that correspond to areas surrounded by pixels labeled as vegetation probably are contours of disease areas. After testing multiple values of thresholds, it was considered that contours with a pixel count less than 300 pixels and larger than 20 pixels represent diseased areas. It was inferred that contours larger than 300 pixels correspond to vegetation regions or background; therefore they are not analyzed.

A contour with a pixel count less than 20 pixels is considered as noise. The area that these small contours enclose is labeled as vegetation or nonvegetation by analyzing its neighboring pixels from the binary image, counting those pixels labeled with 1 (vegetation neighbor pixel) and with 0 (nonvegetation neighbor pixel). If the count for neighbor pixels labeled as vegetation is bigger than the count for nonvegetation pixels, then the contour area is labeled as vegetation, otherwise as nonvegetation.

In the case of contours larger than 20 pixels, the analysis of the area that encloses the contour was made by considering its standard deviation for both and channels, which is obtained by where is the area (number of pixels) that encloses the contour, represents the or value of the pixel , and is the mean value of or in the area enclosed by the contour. The rule for labeling the contour area as vegetation or nonvegetation is described as follows: where is the analyzed contour and is a threshold with a value of 2.4. The value was chosen after carrying out experiments with multiple values and the best results were observed with this threshold.

3. Assessing the Performance of SEVUE Methodology

The performance assessment of the vegetation segmentation methodology proposed in this work is quite difficult because the uncontrolled illumination and background conditions bring about highly complex images to be analyzed. Some techniques that have been used for this purpose consist of comparing an image that has been segmented manually with the same image but segmented using a proposed methodology [6, 22, 25]. The procedure used in this work to assess the methodology performance is based on the approach mentioned before which was proposed by Camargo and Smith [6]. They compared the two images from two perspectives, the first one by analyzing all the pixels from the image   (11) and the second one by analyzing the pixels that are considered as disease   (12). Consider the following:

The manual segmentation was done by overlapping a grid on the image; then each box was evaluated and labeled according to a color schema; white was used to represent a disease region and black(0) was used to represent a nondiseased region. The same process was done for labeling vegetation regions.

The images used by Camargo and Smith [6] correspond to a single leaf. They focused on the disease areas without considering complex background. In our work we did not analyze all the pixels of the image, but only those areas of the image that correspond to leaves with lesion regions which are localized in the foreground. Therefore,   (13) was modified to analyze only the pixels labeled manually as vegetation:

The manual segmentation carried out in this work consisted of marking separate contours on the image in areas considered either as vegetation or as diseased; the pixels into the area enclosed by the contour were labeled with the value of 1. The set of pixels manually labeled either as vegetation or as disease are represented by or , respectively, and those automatically labeled using the SEVUE methodology are represented by or . Then the comparison was made according to the following:

The percentage of missegmentation of disease and vegetation areas is represented by the following equation:

4. Results and Discussion

The -fold cross-validation technique was applied in this work with the goal of quantitatively assessing the performance of the proposed methodology (SEVUE). The value of was considered in -fold that correspond to iterations. A set of images was used. Most of the images presented leaves with areas that corresponded to a visual symptom of disease. The images were manually segmented by three users, who separately marked on the images the regions that they considered as vegetation or as disease. The users segmented only those regions in foreground due to the fact that these provide relevant information from a practical point of view, as opposed to regions at the rear of the scene because these were blurred.

The most relevant results to be assessed are related to aspects such as the segmentation rate of diseased and healthy vegetation in tomato images and the recuperation of these disease or vegetation areas incorrectly classified as nonvegetation after the application of the Bayesian classification.

4.1. Segmentation of Healthy and Diseased Vegetation

An initial assessment was made to determine the accuracy of the SEVUE methodology for segmenting greenness from color images. The methodology results were assessed up to the Bayesian classification process shown in Figure 1, without considering the process of extraction and analysis of contours for the recuperation of incorrectly classified areas.

The vegetation segmentation results obtained by the SEVUE methodology were assessed by comparing them with the manual segmentation (as described in Section 3) done by each user. Additionally, two methods were implemented and assessed for vegetation segmentation: the two most used color index methods, CIVE and ExG-ExR. The results of the mentioned methods were compared with those yielded by the SEVUE methodology.

The ExG-ExR is calculated using (16), from [22], which are described as where , , and are the normalized components and they are determined as follows:

, and are the values of the color channels of RGB image, which have also been previously normalized as a range from 0 to 1; they are defined as follows:

and , , are the maximum total value for each primary color. The color index CIVE is calculated using the following equation from [21]:

After applying the color indices, the output images are gray level images. An Otsu threshold method was adopted to convert each CIVE image to a binary image and a zero threshold in the case of result image from ExG-ExR.

Images segmented by the three methods are shown in Figure 7. Compared with CIVE and ExG-ExR, the SEVUE methodology results showed that segmentation rate is superior by incorporating some disease areas that are not segmented by these methods. Also ripe fruits are removed and less background information is included.

Figure 7: Example results of segmented images by the three methods being compared.

Table 2 shows the average for matching and missegmentation between the manual segmentation performed by each one of the three users and the automatic segmentation performed by the three methods. The column Match (matching) of Table 2 shows the averages of the matching results between the vegetation areas (including disease areas) manually segmented by each user with the vegetation areas automatically segmented by SEVUE (the proposed methodology), CIVE, and ExG-ExR; meanwhile, the column Miss. (missegmentation) shows the missegmentation results between the disease areas manually segmented and those automatically segmented for each method (only areas of disease). The percentages shown in the table are the average results of the cross-validation after 10 iterations for each user and for each method.

Table 2: Matching and missegmentation results.

It is observed from Table 2 that the average difference for matching between the manual and automatic segmentation was 5.2% for SEVUE, 24.2% for CIVE, and 11.6% for ExG-ExR. The smallest average difference for matching was obtained by SEVUE and also the smallest average missegmentation with 28.69%, considerably better than the other methods.

4.2. Recovery of Areas Incorrectly Classified as Nonvegetation

The second aspect to be assessed is related to the obtained results after executing the process aimed at recovering vegetation and disease areas that were not properly segmented. This process consists in the extraction and analysis of contours, as described in Section 2.4. Figure 8 shows some examples of resulting images after applying the process for recovering disease and vegetation areas incorrectly classified as nonvegetation to the same images presented in the previous section (Figure 7). Derived from a visual inspection of the image results some conclusions were extracted, which are described below.(i)The SEVUE methodology recovers a larger area of diseased regions.(ii)Small isolated areas classified as vegetation, and considered as noise, were accurately removed by SEVUE.(iii)We have to highlight that in spite of the fact that the changing illumination conditions and the background areas with similar color as vegetation could have affected the vegetation segmentation results in the image, these results were considerably robust as shown before.(iv)A limitation of the methodology is related to the nonelimination of green fruits from the images and also to small areas belonging to background labeled by the algorithm as vegetation.

Figure 8: Example of resulting images after applying the process for recovering vegetation and disease areas incorrectly classified as nonvegetation.

The results presented in Table 3, which have been obtained after the execution of all of the process of the SEVUE methodology, show the matching and missegmentation percentage between the manual and automatic segmentation. These results correspond to each set of images that were segmented by the three users ().

Table 3: Matching and missegmentation results.

The percentages are represented by the average results of each set of images belonging to the iterations of the cross-validation. It was observed that the SEVUE methodology obtained better results, with a missegmentation average of 8.62% (from Table 3), compared to CIVE with an average of 61.9% (from Table 2) and ExG-ExR with 52.7% (from Table 2). It is observed, from Table 3, that the average difference for missegmentation, between the manual and automatic segmentation, was improved by SEVUE with an average of 20% after applying the process for recovering areas incorrectly classified as nonvegetation.

The Wilcoxon signed rank test assessed the reliability of the SEVUE results taking into consideration the percentages of matching and missegmentation for the whole set of training images (30 images). A statistical significance level of 5% was considered. The analysis shows that the values for the missegmentation of disease areas are less than 0.05 and hence, for all cases, null hypothesis is rejected and alternative hypothesis is accepted to confirm that there is a significant difference between the proposed methodology SEVUE and the methods proposed by CIVE and ExG-ExR.

5. Conclusions

This work dealt with a methodology for segmenting healthy and diseased vegetation from images of tomato plants under uncontrolled outdoor illumination conditions. These conditions make the problem a complex task to be confronted. However, in spite of using vision equipment with technical limitations, we proposed in this work a methodology that copes with this problem yielding better results compared with other previously reported methods. The methodology includes an enhancement process of the captured image quality, coping in this way with the unpredictable effects of illumination. In addition, an unsupervised learning method (SOM) whose function consisted in separating the color images previously enhanced in vegetation and nonvegetation color groups. The methodology also includes a supervised method (Bayesian classifier) whose a priori knowledge was the vegetation and nonvegetation colors extracted from images by the SOM, and its function consisted in the vegetation segmentation from images. The proposed procedure for recovering areas incorrectly classified as nonvegetation, after the application of the Bayesian classification and the inclusion of a second SOM training into the methodology, demonstrated its capacity to achieve a better segmentation rate than the performance yielded by the color index methods CIVE and ExG-ExR. A comparison process that confronted the performance of the proposed methodology with images manually segmented by users yielded good results. The average difference of the matching between the manual and automatic vegetation segmentation (including disease areas) was 2.2% for SEVUE, 11.6% for ExG-ExR, and 24.2% for CIVE. The average of missegmentation of disease areas was 8.62% for SEVUE, 52.7% for ExG-ExR, and 61.9% for CIVE. A nonparametric statistical hypothesis test using the Wilcoxon signed rank sum test demonstrated that missegmentation results of the proposed methodology SEVUE were significantly better compared with CIVE and ExG-ExR. Summarizing, it is concluded that the developed methodology is relatively robust vis-à-vis the outdoor illuminations and stable for applications under uncontrolled environments. We aimed at presenting the methodology SEVUE in practical terms. Therefore, we consider that this methodology can be replicated by the interested users. Some important future works to be performed are to use the segmented images in applications of disease recognition; to assess the performance of the proposed methodology in the segmentation of diseases areas associated with other vegetables.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.


The authors would like to thank Ing. Edson Efren Rios Cuevas for his valuable support to this research.


  1. B. Ji, W. Zhu, B. Liu, C. Ma, and X. Li, “Review of recent machine-vision technologies in agriculture,” in Proceedings of the 2nd International Symposium on Knowledge Acquisition and Modeling (KAM '09), pp. 330–334, Wuhan, China, December 2009. View at Publisher · View at Google Scholar · View at Scopus
  2. Y.-R. Chen, K. Chao, and M. S. Kim, “Machine vision technology for agricultural applications,” Computers and Electronics in Agriculture, vol. 36, no. 2-3, pp. 173–191, 2002. View at Publisher · View at Google Scholar · View at Scopus
  3. H. Y. Jeon, L. F. Tian, and H. Zhu, “Robust crop and weed segmentation under uncontrolled outdoor illumination,” Sensors, vol. 11, no. 6, pp. 6270–6283, 2011. View at Publisher · View at Google Scholar · View at Scopus
  4. A. Meunkaewjinda, P. Kumsawat, K. Attakitmongcol, and A. Srikaew, “Grape leaf disease detection from color imagery using hybrid intelligent system,” in Proceedings of the 5th International Conference on Electrical Engineering/Electronics, Computer, Telecommunications and Information Technology (ECTI-CON '08), pp. 513–516, IEEE, May 2008. View at Publisher · View at Google Scholar · View at Scopus
  5. A. Camargo and J. S. Smith, “Image pattern classification for the identification of disease causing agents in plants,” Computers and Electronics in Agriculture, vol. 66, no. 2, pp. 121–125, 2009. View at Publisher · View at Google Scholar · View at Scopus
  6. A. Camargo and J. S. Smith, “An image-processing based algorithm to automatically identify plant disease visual symptoms,” Biosystems Engineering, vol. 102, no. 1, pp. 9–21, 2009. View at Publisher · View at Google Scholar · View at Scopus
  7. S. Sankaran, A. Mishra, R. Ehsani, and C. Davis, “A review of advanced techniques for detecting plant diseases,” Computers and Electronics in Agriculture, vol. 72, no. 1, pp. 1–13, 2010. View at Publisher · View at Google Scholar · View at Scopus
  8. D. Al Bashish, M. Braik, and S. Bani-Ahmad, “Detection and classification of leaf diseases using K-means-based segmentation and neural-networks-based classification,” Information Technology Journal, vol. 10, no. 2, pp. 267–275, 2011. View at Publisher · View at Google Scholar · View at Scopus
  9. G. Li, Z. Ma, and H. Wang, “Image recognition of grape down y mildew and grape powdery mildew based on support vector machine,” in Computer and Computing Technologies in Agriculture V, pp. 151–162, Springer, Berlin, Germany, 2012. View at Publisher · View at Google Scholar
  10. Y. Tian, L. Wang, and Q. Zhou, “Grading method of crop disease based on image processing,” in Computer and Computing Technologies in Agriculture V, vol. 369 of IFIP Advances in Information and Communication Technology, pp. 427–433, Springer, Berlin, Germany, 2012. View at Publisher · View at Google Scholar
  11. D. C. Slaughter, D. K. Giles, and D. Downey, “Autonomous robotic weed control systems: a review,” Computers and Electronics in Agriculture, vol. 61, no. 1, pp. 63–78, 2008. View at Publisher · View at Google Scholar · View at Scopus
  12. A. Tellaeche, G. Pajares, X. P. Burgos-Artizzu, and A. Ribeiro, “A computer vision approach for weeds identification through Support Vector Machines,” Applied Soft Computing Journal, vol. 11, no. 1, pp. 908–915, 2011. View at Publisher · View at Google Scholar · View at Scopus
  13. D. A. Kumar and P. Prema, “A review on crop and weed segmentation based on digital images,” in Multimedia Processing, Communication and Computing Applications, vol. 213 of Lecture Notes in Electrical Engineering, pp. 279–291, Springer, New Delhi, India, 2013. View at Publisher · View at Google Scholar
  14. J.-X. Du, X.-F. Wang, and G.-J. Zhang, “Leaf shape based plant species recognition,” Applied Mathematics and Computation, vol. 185, no. 2, pp. 883–893, 2007. View at Publisher · View at Google Scholar · View at Scopus
  15. T. Beghin, J. S. Cope, P. Remagnino, and S. Barman, “Shape and texture based plant leaf classification,” in Advanced Concepts for Intelligent Vision Systems, vol. 6475 of Lecture Notes in Computer Science, pp. 345–353, Springer, Berlin, Germany, 2010. View at Publisher · View at Google Scholar
  16. F. Zhang and X. Zhang, “Classification and quality evaluation of tobacco leaves based on image processing and fuzzy comprehensive evaluation,” Sensors, vol. 11, no. 3, pp. 2369–2384, 2011. View at Publisher · View at Google Scholar · View at Scopus
  17. C. L. McCarthy, N. H. Hancock, and S. R. Raine, “Applied machine vision of plants: a review with implications for field deployment in automated farming operations,” Intelligent Service Robotics, vol. 3, no. 4, pp. 209–217, 2010. View at Publisher · View at Google Scholar · View at Scopus
  18. M. El-Helly, A. A. Rafea, and S. El-Gammal, “An integrated image processing system for leaf disease detection and diagnosis,” in Proceedings of the 1st Indian International Conference on Artificial Intelligence (IICAI '03), pp. 1182–1195, Hyderabad, India, December 2003.
  19. L. M. Contreras-Medina, R. A. Osornio-Rios, I. Torres-Pacheco, R. D. J. Romero-Troncoso, R. G. Guevara-González, and J. R. Millan-Almaraz, “Smart sensor for real-time quantification of common symptoms present in unhealthy plants,” Sensors, vol. 12, no. 1, pp. 784–805, 2012. View at Publisher · View at Google Scholar · View at Scopus
  20. X. P. Burgos-Artizzu, A. Ribeiro, M. Guijarro, and G. Pajares, “Real-time image processing for crop/weed discrimination in maize fields,” Computers and Electronics in Agriculture, vol. 75, no. 2, pp. 337–346, 2011. View at Publisher · View at Google Scholar · View at Scopus
  21. T. Kataoka, T. Kaneko, H. Okamoto, and S. Hata, “Crop growth estimation system using machine vision,” in Proceedings of the IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM '03), vol. 2, pp. b1079–b1083, July 2003. View at Publisher · View at Google Scholar
  22. G. E. Meyer and J. C. Neto, “Verification of color vegetation indices for automated crop imaging applications,” Computers and Electronics in Agriculture, vol. 63, no. 2, pp. 282–293, 2008. View at Publisher · View at Google Scholar · View at Scopus
  23. J. Hemming and T. Rath, “Computer-vision-based weed identification under field conditions using controlled lighting,” Journal of Agricultural Engineering Research, vol. 78, no. 3, pp. 233–243, 2001. View at Publisher · View at Google Scholar · View at Scopus
  24. R. Ji, Z. Fu, and L. Qi, “Real-time plant image segmentation algorithm under natural outdoor light conditions,” New Zealand Journal of Agricultural Research, vol. 50, no. 5, pp. 847–854, 2007. View at Publisher · View at Google Scholar · View at Scopus
  25. L. Zheng, J. Zhang, and Q. Wang, “Mean-shift-based color segmentation of images containing green vegetation,” Computers and Electronics in Agriculture, vol. 65, no. 1, pp. 93–98, 2009. View at Publisher · View at Google Scholar · View at Scopus
  26. A. Meunkaewjinda, Hybrid artificial intelligence approach for grape leaf disease diagnosis system from color imagery, Suranaree University of Technology, Nakhon-Ratchasima, Thailand, 2007.
  27. M. Emambakhsh, H. Ebrahimnezhad, and M. Sedaaghi, “Integrated region-based segmentation using color components and texture features with prior shape knowledge,” International Journal of Applied Mathematics and Computer Science, vol. 20, no. 4, pp. 711–726, 2010. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at Scopus
  28. Y. Jiang, K.-J. Chen, and Z.-H. Zhou, “Som based image segmentation,” in Rough Sets, Fuzzy Sets, Data Mining, and Granular Computing, pp. 640–643, Springer, 2003. View at Google Scholar
  29. D. Chi, “Self-organizing map-based color image segmentation with k-means clustering and saliency map,” ISRN Signal Processing, vol. 2011, Article ID 393891, 18 pages, 2011. View at Publisher · View at Google Scholar
  30. Y. Wu, Q. Liu, and T. S. Huang, “An adaptive self-orga nizing color segmentation algo rithm with application to robust real-time human hand localization,” in Proceedings of Asian Conference on Computer Vision, pp. 1106–1111, 2000.
  31. S. Suzuki and K. Be, “Topological structural analysis of digitized binary images by border following,” Computer Vision, Graphics and Image Processing, vol. 30, no. 1, pp. 32–46, 1985. View at Publisher · View at Google Scholar · View at Scopus