1. Introduction
Under climate change, water stress has become a great challenge to maize products globally. In research on crop yield, non-destructive monitoring of crop structural traits is of great significance. As one of the most widely used structural traits, fractional vegetation cover (FVC), defined as the proportion of ground surface occupied by green vegetation [
1], plays an important role in monitoring vegetation growth status and estimating crop yields (e.g., evapotranspiration and above-ground biomass) [
2,
3,
4]. In addition, FVC is also a key parameter in the AquaCrop model, which is widely used to simulate crop yield response to water under different irrigation and field management practices [
5,
6]. Therefore, it is important to estimate FVC rapidly and accurately for annual crops under different irrigation treatments during crop growing seasons.
Visual estimation, direct sampling, and digital photography have been developed to measure FVC for agricultural applications [
7,
8]. With the development of sensor technology, many researchers have used Red-Green-Blue (RGB) digital cameras [
9,
10,
11] or near infrared (NIR) spectral sensors [
12] in the field of agriculture. However, RGB digital imagery has been used more widely than NIR spectral sensors in image segmentation, with the advantages of low cost and higher spatial resolution. With the availability of inexpensive high-quality digital cameras in agriculture applications, estimating FVC by image segmentation is becoming more common. In general, image-based FVC estimation can be grouped into two categories: (1) machine learning methods (e.g., K-means, Decision Tree, Artificial Neural Networks, and Random Forest) [
13,
14,
15] and (2) threshold-based methods [
16,
17]. Machine learning methods need a large amount of training data sets for the purposes of calibration. The generation of training data, affected by human intervention, has a great influence on model accuracy. Threshold-based methods, with their advantages of simplicity, high efficiency, and accuracy, play an important role in precision agriculture [
18], and have been successfully used for crops such as wheat, maize, cotton, and sugar beet [
12,
19,
20,
21].
Selecting an appropriate color feature for crop segmentation is a key step to obtain FVC estimations using threshold-based methods [
17,
22]. Excessive green index (ExG) [
23,
24], H channel of the Hue-Saturation-Value (HSV) color space [
25], and a* channel of the CIE L*a*b* color space [
26] are the three widely used color features. The visible spectral index, ExG, can be calculated directly from digital numbers in three components of RGB images, and represents the contrast of the green spectrum against red and blue [
24]. The H channel from the HSV color space uses an angle from 0° to 360° to represent different colors [
27]. The a* channel from the CIE L*a*b* color space is relative to the green-red opponent colors [
26]. In addition, to deal with the shadow effect in classification, the Shadow-Resistant LABFVC (SHAR-LABFVC) method was proposed [
28]. In the SHAR-LABFVC method, the CIE L*a*b* color space, from which the a* color feature was extracted, was transformed from the brightness-enhanced RGB. To distinguish it from the a* color space in other research [
26], the a* color feature in the SHAR-LABFVC method is denoted as the a*_I color feature in this study.
Another key step to obtain FVC estimations using threshold-based methods is searching for the appropriate classification threshold. Otsu’s method is one of the most widely used threshold techniques [
29], and has been used in many applications of image segmentation and plant detection [
30,
31,
32]. However, this method can produce under-segmentation in some circumstances [
18]. Another widely used threshold method was proposed by [
16] based on a Gaussian mixture model (GMM) of color features derived from images. Specifically, a Gaussian mixture model is a parametric probability density function represented as a weighted sum of Gaussian component densities. Classification thresholds for discriminating vegetation and backgrounds can be calculated from a GMM fitted on different color features. Fitting a GMM on an appropriate color feature contributes to accurate threshold calculation.
To evaluate the separability of two Gaussian distributions on different color features, the separability distance index (SDI) [
26] and the instability index (ISI), which is the reciprocal of SDI [
33], were proposed. SDI (Equation (1)) describes the separability of the distributions between vegetation and backgrounds,
where
and
are mean values of green vegetation and backgrounds, respectively; and
and
are standard deviations of green vegetation and backgrounds, respectively. The larger the SDI is, the easier it is to fit these two Gaussian distributions. However, whether the SDI is an appropriate criterion for selecting the right color feature to produce the best performance of GMM fitting is still uncertain and has not been explored in previous research. It is necessary to further study the applicability of SDI in selecting color features.
With GMM fitted on an appropriate color feature, applying an optimal approach to obtain the classification threshold is essential for calculating FVC. Once the parameters for both Gaussian distributions of vegetation and backgrounds have been estimated, classification thresholds can be calculated by two classic methods, namely, the intersection method, and the equal misclassification probability method. In the intersection method, the intersection of the two Gaussian distributions is used as the classification threshold (hereinafter referred to as T
1). In the equal misclassification probability method (hereinafter referred to as T
2), the threshold makes the green vegetation and backgrounds have an equal misclassification probability [
26]. Both two classic methods derive thresholds based on the Gaussian distributions of green vegetation and backgrounds in each image.
However, both the two abovementioned threshold methods have limitations. Firstly, both perform well when the bimodality is clearly seen in the distributions of images on a color feature; however, when FVC is extremely low or the canopy is nearly closed, the thresholds obtained by these methods may be inaccurate. Secondly, in semi-arid and arid areas, where climate is characterized by long periods of drought with decreasing projected rainfall, the high possibility of crops suffering from water stress may raise a new challenge for FVC estimation caused by spectral changes. To cope with water stress, crops usually exhibit adaptive mechanisms at the leaf level to reduce light absorption and dissipate excess absorbed energy, such as the decrease in chlorophyll concentration and down-regulation of photosynthesis, and increase in the concentration of deep oxidized xanthophyll cycle components [
34,
35,
36,
37]. The spectral changes of crops caused by the pigment changes at the leaf level may have a great influence on the accuracy of threshold calculation using the two methods. Therefore, RGB images collected across the entire crop growing season with a wider FVC range and different levels of water stress conditions need to be further studied regarding the performance of FVC estimation using threshold methods based on GMM. In addition, in our previous research for cotton FVC estimation [
21], an interesting phenomenon was found: a fixed classification threshold exists in the a* channel. Therefore, further study is needed to determine if there is a fixed classification threshold for maize.
In this study, there were two main objectives: (1) to determine if a fixed-threshold method exists and is applicable for images covering the full range of FVC from open to nearly closed; (2) to explore if the fixed-threshold method outperforms the two classic threshold-based methods (T1 and T2 methods) in estimating FVC for images with high reference FVC or captured from deficit irrigation plots. This research could provide a more practical, efficient, and accurate way to estimate the FVC of field maize under water stress based on RGB imagery.