Abstract

Estimation of canopy density is necessary for ecological research and woodland management. However, traditional manual methods are time consuming and subject to interobserver variability, while existing photographic methods usually require expensive fish-eye lenses and complex analysis. Here we introduce and test a new method of digital image analysis, CanopyDigi. This allows user-defined threshold to polarise the 256 grey shades of a standard monochrome bitmap into dark “canopy” and light “sky” pixels (the threshold being selected using false-colour images to ensure its suitability). Canopy density data are calculated automatically and rapidly, and, unlike many other common methods, aggregation data are obtainable using Morisita’s index to differentiate closed (diffuse light) and open (direct light) canopies. Results were highly repeatable in both homogeneous and heterogeneous woodland. Estimates correlated strongly with existing (nondigital) canopy techniques, but quicker and with significantly lower interobserver variability (CV = 3.74% versus 20.73%). We conclude that our new method is an inexpensive and precise technique for quantifying canopy density and aggregation.

1. Introduction

Vegetation structural complexity comprises several parameters including the density and height of different vegetation layers, the percentage coverage of each layer, and species composition [1].In a woodland ecosystem, one of the most important aspects of overall structural complexity is the canopy density and aggregation [2, 3]. These parameters influence stand productivity and species composition, as well as the density and growth of the understorey, which is a key determinant of habitat [4]. Estimation of canopy coverage is important in studies of species-habitat relationships (e.g., [510]) and also for land managers given the increasing emphasis on sustainable woodland management. The effective use of canopy data in ecological research, or to inform woodland management, demands that measurements are accurate, precise, and obtainable using a quick, easy, and inexpensive method [11]. Estimates also need to have a high consistency and low interobserver variability to avoid generation of erroneous results [1214].

1.1. Canopy Measurements: Coverage and Density

Somewhat surprisingly given its importance, there is no standard method of measuring canopy on a relatively small scale (e.g., for specific woodland plots). Jennings et al. [15] distinguish two basic types of measurement of forest canopies; the percentage of canopy coverage—the area of ground covered by a vertical projection of the canopy—or the percentage of canopy density (also referred to as closure)—the area of ground covered by canopy when viewed from a single point. Methods to assess these will be considered briefly below (for a more in-depth review, see [15, 16]).

1.1.1. Coverage

The most straightforward method of estimating coverage is by eye. This can be estimated as a percentage or using a scale such as the generic DOMIN or Braun-Blanquet [17, 18]. These are quick and simple, but are also subjective and very variable [19]. Indeed more variability in visual estimates can be attributed to the evaluators than to actual differences in the vegetation cover [19]. To increase objectivity and decrease variability, canopy cover is often assessed using a sighting tube [16]. These either have a single central cross-wire or an acetate grid with several intersections, and recorders look up through the tube and note if the cross-points are aligned with vegetation or sky. The alignment ratio becomes the coverage estimate [6, 13]. Unfortunately numerous measurements are recommended at each sample location, and it is very difficult to keep the sighting tube vertical, which is essential to avoid overestimating the canopy [20].

1.1.2. Density

Canopy density is more closely linked to microclimate and light regime than canopy cover and is therefore generally preferred in ecological studies [16, 21]. It can be measured visually using the specific crown illumination index [22], using a spherical densiometer [23] or using a moosehorn [24]. A spherical densiometer is a mirrored dot-etched hemispherical dome. When placed under the canopy, the proportion of dots covered by the canopy’s reflection estimates its density. As well as this being expensive and specialised [4], there are concerns about its accuracy [15, 25] and it is subject to high interobserver variability and, therefore, a lack of precision [6, 26]. The moosehorn is a transparent dot-marked screen that held slightly away from the eye; the ratio of vegetation and sky alignments are recorded as before. Not only is this method cumbersome and the equipment fragile, but it also underestimates canopy density as approximately 23% of canopy gaps remain outside the field of view [12, 14].

It should also be noted that there is a considerable difference between a canopy with many small gaps, which allow diffuse light to penetrate (Figures 1(a) and 1(b)), and a canopy with one or more large gaps, which allow direct light to penetrate (Figures 1(c) and 1(d)) [27]. However, none of the above methods can discern this objectively, such that this important habitat parameter is disregarded in the vast majority of ecological studies. Indices to quantify structural complexity to distinguish different types of vegetation aggregation do exist, for example, gap fraction estimates, leaf area index, canopy openness, and sky factor [28, 29]. However, these are only obtainable using complex methods such as: (1) ground-level tripod-mounted scanning using light detection and ranging (lidar) units or terrestrial laser scanners, which can digitise structural information about the canopy very rapidly [28, 30]; or (2) canopy photography, both of which are often prohibitively expensive.

1.2. Canopy Photography

Determination of canopy density from photographs via estimation of light penetration has been suggested previously for canopy estimation [4, 31, 32]. Usually, this refers to hemispherical photography [33, 34], which gives an almost complete view of the canopy, and analysed using specially developed software such as Gap Light Analyzer [27] and HemiView [35]. However, there are several drawbacks to the use of hemispherical photography to quantify canopy parameters. The most important issue is the high cost (and thus low accessibility) of equipment [16], but hemiplots (images resulting from hemispherical photography) also: (1) include all vegetation present, including understory and field layers, which may not be desirable, (2) cannot be taken at ground level on slopes because the slope itself is photographed, and (3) require complex spherical geometric calculations as there is no direct relationship between canopy area and its photographic representation [36].

As a result of these limitations, particularly that of cost, canopy photography is often dismissed as being complex, time consuming, and expensive [12, 15] and there have been calls for the development of other methods that are cheaper and easier to use [16]. If tree canopy is the only structural complexity variable of interest, standard digital photographs (rather than hemiplots) can be taken using a standard camera and analysed much more easily. Some work on computer analysis of standard canopy photographs has been undertaken previously. Engelbrecht and Herz [21] examined canopy photographs using Adobe Photoshop (San Jose, USA) whereby a pure black-and-white image was created by maximising the image’s contrast, and the percentage of black (canopy) pixels was determined. This is a straightforward technique that uses commercial software. However, photographs need to be taken under uniformly overcast skies, which places considerable (often insurmountable) constraints on fieldwork. Moreover, although the method worked well on equatorial evergreen canopies, by the authors’ own admission, it does not appear to be transferable to other situations. Indeed, even when standard photographs of an artificial tree taken under laboratory conditions (uniform background, optimal lighting, etc.) were processed in this way, the resultant images were not a particularly accurate representation of the true conditions, with small areas of tree being regarded as background [28]. Purcell [37] and Richardson et al. [38] have used a conceptually similar approach to quantify ground vegetation cover from colour photographs by counting the number of green pixels within specific hue and saturation ranges. Again, however, there were constraints: plants in shadow were often not included, sun flecks could bias analysis, and the technique only worked with green vegetation (such that it would not work, for example, for trees in the autumn).

In this study, we introduce a new photographic method to assess percentage density via estimation of light penetration. This involves flexible digital image analysis of standard canopy photographs taken with a cheap digital camera (not hemiplots) with a user-selected threshold between dark (canopy) and light (sky) areas. Image analysis also includes quantification of canopy aggregation (patchiness) in order to quantify this important variable objectively. We test the new technique in two different situations (homogeneous and heterogeneous woodland) and then compare it to existing methods of quantifying canopy coverage and density in order to evaluate relative levels of interobserver variation. Finally, we consider the wider applicability of the new technique for surveying other aspects of vegetation structural complexity.

2. Materials and Methods

The following details conversion of monochrome canopy photographs to false-colour images using a series of threshold values to differentiate between dark “canopy” and light “sky” pixels to thus obtain information on canopy density (percentage quantity derived from estimation of light penetration) and aggregation (patchiness). Methods used to test the technique are then outlined. All aspects of determination and quantification are undertaken using a suite of computer programs written in Microsoft Visual Basic. This suite of programs, CanopyDigi, is included with this paper as electronic supplementary material, together with a user-friendly interface, full documentation, and help files (For further information, see Supplementary Material available online at http://dx.doi.org/10.5402/2012/619842).

2.1. Developing the Digital Image Analysis Technique

A threshold approach was used to obtain estimates of canopy density by digital analysis of canopy photographs held as monochrome bitmap (BMP) images. This made use of the fact that within any monochrome BMP, there is provision for 256 shades of grey, ranging from black (0) to white (255). Although in principle a canopy estimate could be obtained simply by counting the number of pixels in each grey shade using a demarcation value (threshold) at and above which pixels are regarded as canopy (an approach originally suggested by Anderson) [39], this was found to be problematic. Using a fixed threshold (e.g., [40]) placed too many restrictions on fieldwork to be practicable, while manual selection of a threshold was entirely subjective when there was no way of checking the demarcation value objectively [4, 11]. As preliminary testing showed that the most appropriate threshold varied significantly according to tree type, ambient light, and time of day, it was decided to undertake analyses on specially created false colour images with all “canopy” pixels being coloured blue and all “sky” pixels being coloured red. In this way, the suitability of any given threshold could be examined in a “determination” stage and canopy then analysed using that threshold in a “quantification” stage.

Theoretically, it should be possible to automate the determination stage, either fully or partially, using threshold-selection algorithms, and there have been several attempts to do this (e.g., [41, 42]). However, these algorithms are not always successful in determining an appropriate threshold [43], especially when the relative contrast between vegetation and sky changes across the image (e.g., sun-illuminated vegetation against white cloud in one part of the image and shaded vegetation against bright blue sky in another part of the image) [44] and add substantially to processing time. A quick examination of photograph-specific histograms demonstrates the problem. If canopy photographs were always bimodal in greyscale pixel frequency (i.e., if canopy equalled black and sky equalled white), it would be possible to choose either: (1) the midpoint between the two modes; or (2) the midrange value with the lowest pixel occurrence and make this the automated threshold point. Alternatively, if the histograms were always normally distributed, it would be possible to ensure that all values below the mode were treated as canopy and all values above this were treated as sky. Although both of these approaches would be easy to code, extensive testing showed that these approaches were, in reality, ineffective. Histograms were very location specific, some being near normal, others being bi- or multimodal, and others tending towards a uniform, U-shaped, or Poisson distribution (Figure 2). We therefore adopted a user-controlled determination stage as detailed above. This approach also has the huge advantage that the user can see the “fit” of the false-colour image to the original image. This means that if it is simply not possible to have one threshold to separate sky and canopy across the entire image (i.e., if a variable, context-specific, threshold is most appropriate, an approach which has been showcased for individual photographs but has yet to be developed into batch-processing software [44]) it is possible for the user to interpret canopy statistics for that image with caution. In other words, uncertainly can be recognised and allowed for explicitly rather than being hidden in the realms of computer code. Usefully, however, based on the results of testing the technique as reported here, we found that one threshold from a restricted (standardised) subset of options was appropriate in the majority of situations and that this could be used in the first (and often only) run through the determination stage (see Results). This has the twin advantages of making the threshold, if not automated, at least easy to select and quick to implement in the majority of situations, while retaining complete flexibility to enable analysis of atypical images.

2.1.1. Determination Stage

To determine the most appropriate threshold for any picture, eight false-colour images, each at a different default threshold, were created. The most appropriate threshold (i.e., that which created the “best fit” red/blue image) was then selected following simultaneous visual comparison of each false-colour image with the original photograph. In order to minimise the amount of time during which user input was required, batch processing was used whereby multiple photographs were run through the determination stage, without the need for user input. The most appropriate false colour image for each photograph was user selected in a rapid post hoc selection process. It should be noted that although the eight standard (default) thresholds were determined following extensive testing, the ability for users to change the preset thresholds for any pixel shade between 1 and 254 was retained. This ensured that the digital image analysis was entirely flexible [21], such that atypical photographs (e.g., photographs with deep shadows or light flecks) were still analysable. Importantly, the threshold value itself was immaterial as long as the false-colour image distinguished correctly between canopy and noncanopy areas of the photograph, such that photographs processed using different thresholds were still directly comparable.

2.1.2. Quantification Stage

To provide an estimate of canopy density for each false-colour image selected during the determination stage, the percentage of blue (canopy) pixels was calculated. Then, to provide an estimate of canopy dispersion, Morisita’s index, one of the most robust distribution measures [45], was quantified. To do this, each image was electronically divided into 12 subimages (equivalent to the plots or quadrates used in field applications of Morisita’s index (e.g., [46])), and the number of canopy pixels was calculated for each subpicture. The index was then calculated using𝐼𝑑𝑋=𝑛2𝑁𝑁(𝑁1),(1) where 𝐼𝑑 is Morisita’s index of dispersal, 𝑛 is the total number of subpictures, 𝑁 = total number of canopy pixels in the entire image and 𝑋2 is the square of the canopy pixels in each subpicture (𝑋12, 𝑋22, etc.). The index value increased with increasingly large gaps in the canopy, producing an interval variable for subsequent analysis.

Batch processing without the need for user input was used, and the final density and Morisita statistics for all original photographs in one batch were written out to a single CSV (comma-separated variable) file to facilitate further analysis (in Excel, SPSS, Minitab, etc.), as well as to individual photograph-specific text files.

2.2. Testing the Digital Image Analysis Technique

To assess the effectiveness of the digital image analysis technique, canopy photographs were taken at Nagshead Nature Reserve (Gloucestershire, UK) and Lady Park Wood (Monmouthshire, UK) (𝑛=295 and 58, resp.). Nagshead is a homogeneous wood, dominated by mature Pedunculate oaks (Quercus robur), which are around 200 years old and have an average height of about 25 m. Lady Park Wood is a heterogeneous, nonintervention, wood with a mixed-aged, mixed-height, canopy (60–200 years old; 15 m–30 m + in height) comprising oak (Q. robur), ash (Fraxinus excelsior), elm (Ulmus glabra), beech (Fagus sylvatica), small-leaved lime (Tilia cordata), and yew (Taxus baccata). To obtain images suitable for analysis (i.e., that captured the necessary information but were small enough to ensure rapid computer analysis), photographs were taken vertically upwards using a digital camera (Minolta Dimage Z1) set to monochrome format with a standardised image size of 640480 pixels (following [37]) (note that photographs could equally well have been converted to these settings after fieldwork using standard software such as Paintshop Pro or Matlab, or using shareware such as ReaConverter (http://www.reaconverter.com/), which all use batch processing to minimise the need for human interaction, and all of which would also convert photographs of a different aspect ratio than 1 : 1.333, if necessary). The resolution of 640480 was used because any higher resolution would not improve the accuracy of the technique, merely generating spurious precision and substantially increasing computer processing time; larger files can easily be converted (see above). The focal length at which a canopy photograph is taken will influence results since the focal length affects the actual area of the canopy being photographed (taking one image at, for example, 18 mm, and another at 50 mm would result in very different images, and, thus, in all likelihood, different canopy statistics would be obtained after analysis). There is no “right” focal length but if images need to be directly comparable—for example, in longitudinal monitoring at the same locations—it would be necessary to be standardised on one agreed focal length. In this case, focal length was set to 35 mm, the focus was set to infinity, and aperture and shutter speed were determined automatically. To avoid perspective influencing the canopy calculations, the verticality of the photographs was assured using a two-way spirit level affixed to the back of the camera. Following fieldwork, the images were converted from lossy JPEG format (the standard digital camera format) to lossless BMP format using ReaConverter (see above) to ensure that all photographs had an identical number of pixels, thereby facilitating analysis.

2.3. Threshold Sensitivity Analysis

To determine the effect of selecting a nonideal threshold on the resultant canopy density percentage and Morisita’s index data, a subset of 20 photographs, all taken in Lady Park Wood by one individual (AEG) and all judged to have an ideal threshold of 155, were selected randomly from the 54 such photographs matching these criteria. These were run through CanopyDigi at a threshold of 155, and four thresholds either side of this in increments of five greyscale values (i.e., 135, 140, 145, 150 and 160, 165, 170, and 175). The canopy percentage estimates were extracted for each threshold on a per-photograph basis and were run through a repeated measures ANOVA (after allowing for sphericity using the Greenhouse-Geisser correction) whereby each photograph (site) was treated as an individual case on the basis that nonsignificance would be good evidence that the technique is robust to relatively substantial deviations from the most suitable threshold. The same process was used for the Morisita values.

2.4. Comparison of the Digital Analysis Technique Using Photographs from Different Cameras

Although there is no reason to suppose that two images, taken consecutively at the same place and the same focal length but with different cameras, would give differing results, we tested this empirically. One of us (A. E. Goodenough) took one canopy photograph at each of 12 different locations with the camera used previously (a compact digital set to 35 mm) and, immediately thereafter, took a second photograph with a single lens reflex (SLR) camera (Canon 450D mounted with an 18–55 mm lens set to 35 mm equivalent). Photographs were taken under field conditions and without the aid of a tripod. We ran all 24 images through CanopyDigi and tested for differences in the recorded canopy density (percentage) and dispersion (Morisita’s index) using separate paired-samples t-tests (one for each of the two canopy parameters; percentage data were arcsine square root transformed to normalise them).

2.5. Comparing Methods: Interobserver Variation and Sampling Effort

To compare interobserver variation of canopy photographs with existing methods (1 = visual estimation, 2 = cross-wire sighting tube, and 3 = moosehorn), ten volunteers were asked to record the canopy at ten sample locations. All volunteers, drawn from both sexes and a variety of age ranges, were novices—they had no experience of taking canopy photographs or analysing the resultant images, nor any experience of the traditional methods used to estimate canopy. In each case a 5 m2 plot was laid to one side of a central tree. To prevent the use of other techniques influencing visual estimates, these were made first [47], with the remaining methods being employed in a random order [6]. Visual estimation (a percentage estimate to the nearest 5%), the moosehorn estimation, and the digital photograph were taken from the centre of the plot. To obtain the recommended 100 cross-wire sighting tube readings [20], the plot was conceptually divided into 100 squares (10 × 10), and a measurement was taken in the middle of each small square, the ratio of vegetation to sky alignments providing the coverage estimation. To eliminate potentially confounding variables, all fieldwork was undertaken in comparable lighting and weather conditions and within a 48-hour period. At each sample location, the volunteer took two photographs of the canopy. The first photograph was always analysed unless there was a problem (e.g., camera shake) when the second photograph was used instead [32]. The camera and settings were consistent and as described above. To ensure that the volunteers were not influenced by the estimates given by others, each completed their fieldwork in isolation [6]. The coefficient of variation (CV) was calculated for each sample location using the estimate of canopy density from each observer (𝑛=10) and expressed as a percentage (CV = (standard deviation/mean)100). Significant differences in the variability of the estimates derived using digital image analysis versus the other techniques were ascertained using a series of three 𝐹 tests. In all cases, square of the differences in all 10 canopy estimates (one at each sample location) was calculated on a per-individual basis to remove location-specific effects (as per [38]). The mean canopy calculation for each sampling site, according to each method, was calculated so that methods could be compared using correlation analysis. The sampling effort, a variable that is often ignored when canopy quantification methods are compared, was also recorded by quantifying the mean time taken to achieve estimates by each method, both in terms of fieldwork and postfieldwork calculations.

3. Results

3.1. Testing the Digital Image Analysis Technique

No problems were experienced in taking the canopy photographs. As regards processing, the same threshold value (greyscale value 155) was suitable for the majority of photographs (all 295 photographs taken at Nagshead Nature Reserve, the homogeneous site, and 54 of the 58 photographs taken at Lady Park Wood, the heterogeneous site). Of the Lady Park Wood photographs, two required a lower threshold (greyscale value 125) due to dark blue sky being regarded initially as canopy, while the other two needed a higher threshold (greyscale value 185) due to sun flecks on the vegetation being regarded initially as sky (this latter could potentially be minimised in camera with the use of a polarising filter attached for a single photograph that could then be processed as normal, or, alternatively, by taking one photograph with a polariser and one without a polariser and combining the images electronically to add in the sunlight-reflected pixels before use of CanopyDigi). The percentage data gave information on canopy density itself, while Morisita’s index quantified the aggregation of canopy gaps in each photograph (Figure 1). Digital analysis of the test photographs resulted in a wide range of percentage density estimates (minimum = 26%; maximum = 93%). Morisita’s index values provided an interval variable for subsequent analysis and ranged from 1.002 to 2.745.

3.2. Threshold Sensitivity Analysis

When 20 photographs, all with an ideal threshold of 155, were run through CanopyDigi at lower and higher thresholds (135, 140, 145, 150; 160, 165, 175, 175, resp.), neither the canopy percentages nor the Morisita values calculated at these different thresholds differed significantly from one another on a per-photograph basis (Greenhouse-Geisser repeated measures ANOVA: 𝐹 = 6.279, d.f. = 1, 𝑃=0.085 and 𝐹 = 1.266, d.f. = 1.022, 𝑃=0.324, resp.). This suggests that, while it is obviously best to select the ideal threshold to best distinguish canopy and sky areas of the photograph, selecting a threshold quite substantially lower or higher does not significantly confound the results.

3.3. Comparison of the Digital Analysis Technique Using Photographs from Different Cameras

There was no difference in either canopy density (percentage) or dispersion (Morisita’s index) estimates for the 12 locations photographed pairwise using two different cameras; the standard compact camera described above and as a digital SLR camera (paired-samples 𝑡-tests 𝑡 = 0.95, d.f. = 11, 𝑃=0.341 and 𝑡 = 0.389, d.f. = 11, 𝑃=0.705, resp.).

3.4. Comparing Methods: Interobserver Variation and Sampling Effort

Mean canopy density estimates derived using digital analysis correlated strongly and significantly with mean estimates from each of the traditional methods (correlation of photographically derived estimates with visual estimates 𝑟=0.600, 𝑛=10, 𝑃=0.033; with sighting tube estimates 𝑟=0.750, 𝑛=10, 𝑃=0.006; with moosehorn estimates 𝑟=0.680, 𝑛=10, 𝑃=0.015; Figure 3), indicating that the methods themselves were comparable. However, digital estimates had much lower interobserver variability than estimates made using traditional methods (Table 1; Figure 4). The mean difference between the highest and lowest canopy estimates at each sample location was 10 percentage points for digital image analysis and 44 percentage points for all other methods combined, with CV values being 3.74% and 20.73%, respectively. When compared statistically, canopy calculation using digital image analysis had significantly lower interobserver variability than visual estimates (𝐹 = 147; d.f. = 9; 𝑃<0.001), estimates derived from sighting tubes (𝐹 = 81; d.f. = 9; 𝑃<0.001), and estimates derived from using the moosehorn (𝐹 = 65; d.f. = 9; 𝑃<0.001). Indeed, the digital analysis method produced results with significantly lower variability even for the sample location (site 5) that had the smallest difference in interobserver variability between estimates made using the digital technique compared with those made using visual, sighting tube and moosehorn methods (𝐹 = 56, 41 and 17, respectively; d.f. = 9 and 𝑃<0.001 in all cases).

As regards sampling effort, visual estimation was the quickest method, requiring just 20 seconds per sampling location to record the information and later computerise it (Figure 5). Conversely, the sighting tube methods were very time consuming. The only way to achieve an acceptable sampling speed was to use two people, one taking the measurements, the other recording them. Even then, to survey each site took an average of 6.5 minutes in the field and a further 1.5 minutes to calculate the ratio of vegetation to sky alignments. The moosehorn method was quick in the field—about 30 seconds—but quite time consuming after fieldwork, taking an average of 3 minutes per site to process information. The photographic method was extremely quick in the field: just 10 seconds including camera set-up time. Because of automation and batch processing restricting user input in the determination stage to selecting the most appropriate threshold value for each picture, the actual time spent at the computer to achieve objective quantification of canopy density and dispersion at all 353 sites was about 30 minutes, or just over 5 seconds per site (the time needed to choose the correct threshold for each photograph). Computer processing time, during which no user input was necessary (to generate eight individual false-colour pictures for each photograph and to calculate statistical information from each selected false-colour picture), was about 3 hours, or around 30 seconds per picture—these processes were run overnight. Thus the overall time taken to complete the photographic after fieldwork analysis was 35 seconds per picture, the vast majority (>85%) of which was computer processing time during which no human intervention was necessary.

4. Discussion

CanopyDigi is a precise method of determining canopy density and enables objective mathematical calculation of canopy aggregation, providing useful quantification of canopy structure and differentiating between canopies with many small gaps, which allow diffuse light to penetrate (Figures 1(a) and 1(b)), and canopies with one or more large gaps, which allow direct light to penetrate (Figures 1(c) and 1(d)). The digital analysis method is much cheaper than advanced lidar scanning [28, 30] or hemispherical photography approaches [33, 34] and is comparable with existing commonly used techniques such as the use of a sighting tube or moosehorn, as demonstrated by a series of strong correlation coefficients. It should also be noted that previous work has suggested that analysis of standard canopy photographs correlates strongly with technology-heavy and expensive lidar scanning [28]. If the camera used an integrated global positioning system (GPS) functionality, or a GPS attachment is used, it would also be possible to create interactive maps of the canopy images tagged with the statistics produced by CanopyDigi (in a similar way to Stafford et al. [48]).

4.1. Comparing Methods: Interobserver Variation and Sampling Effort

Low levels of interobserver variability are important in all studies using data collected by multiple individuals. Use of digital image analysis provides canopy estimates that have a much lower level of interobserver variation than what occurs using traditional (visual, sighting tube, and moosehorn) methods, both overall and for every individual sample location. Nine of the ten observers found estimating coverage visually difficult, with most finding it especially hard for sample locations dominated by coniferous species (locations 6 and 8). Use of the sighting tube was found to be difficult on sloping ground, and constant use was found to be fatiguing. When using the moosehorn, any small movement of the canopy made it very challenging to assess whether vegetation was aligned with the dots, decreasing the consistency of the techniques between observers. Conversely, most people found the camera method comparatively easy, and the precaution of each volunteer taking two photographs at each sample location proved almost unnecessary as the first photograph was suitable for analysis (no camera shake) in all but one case out of 100.

In terms of sampling effort, assessing canopy visually is rapid and gives an immediate percentage estimate with no post-fieldwork effort; however, the precision and consistency of the estimates is very poor. The sighting tube method had very high sampling effort, such that the amount of time spent obtaining the readings would, in many cases, be prohibitively high. In contrast, digital photographs are easy and quick to take in the field, and use of automated batch processing decreases the amount of time for which user interaction is necessary in after fieldwork analysis. Total sampling effort is lower than for either the sighting tube method or the moosehorn, and, when the human-based processing time is considered, digital analysis takes the same time as visual estimation when the latter also includes the time taken to transcribe field notes to analysable electronic format—a process undertaken automatically in the digital analysis method—(15 seconds versus 20 seconds, resp.).

4.2. The Wider Applicability of the Digital Analysis Technique

The flexible digital image analysis technique described here is suitable for any situation where the purpose is to estimate the percentage cover and/or spatial pattern of one object that can be differentiated clearly from the background matrix in a monochrome picture. Thus it would be suitable for quantifying other aspects of vegetation structural complexity, including percentage ground cover using photographs taken from above and complexity within a vegetation layer (e.g., a scrub layer) using photographs taken horizontally within that layer against a white background (e.g., a sheet). Time-series photographs could also be used to assess the processes of leaf burst or defoliation. The technique could be used to quantify temporal vegetation colonisation of new surfaces (e.g., new quarry faces or mudflows).

4.3. Conclusions

The digital image analysis technique is a quick, inexpensive, and precise method of determining canopy density and dispersion objectively. Tests have shown that it performs to a very high standard in both homogeneous and heterogeneous woodland and that interobserver variability is vey low. In comparison with many past attempts to analyse canopy photographs digitally, this method is easier and quicker as spherical geometry is not needed [36]. Moreover, the flexible threshold selection means that there is no requirement to take photographs under uniformly overcast skies as in previous methods [11, 21]. The technique has wider applicability as a tool to survey other aspects of ecological complexity and to quantify temporal or spatial changes in relative abundance of a specific species if that species can be distinguished from its surroundings in a monochrome photograph.

Acknowledgments

Thanks are due to George Peterken for encouragement and suggesting the need to measure canopy patchiness and Rick Stafford for comments on a draft of this paper. Thanks also are due to the ten volunteers who trialled different methods to enable quantification of interobserver variability: A. Arthur, M. Goodenough, L. Hancock, M. Lee, A. Morss, P. Preston, A. Pyke, S. Pyke, R. Watkins, and T. Watkins.

Supplementary Materials

Dose response curves of petroleum ether, ethyl acetate, methanol and 70% ethanol extract of D. bulbifera and G. glauca against porcine pancreatic α-amylase, crude murine pancreatic, small intestinal, liver glucosidase and α-glucosidase are available in the supplementary material. Graphical data for HPLC fingerprinting of all the extracts of D. bulbifera and G. glauca are presented here.

  1. Supplementary Figures