1. Introduction
Fraction vegetation coverage (FVC) refers to the ratio of the vertical projection of the canopy area to the ground surface area per unit area. FVC is expressed as a fraction or percentage of the reference area and is influenced by field management practices such as nitrogen application and irrigation. FVC is also closely related to crop yield [
1], and to crop physiological activities such as photosynthesis and transpiration and can reflect the growth status of crops [
2]. For tuber crops like potatoes, monitoring FVC can provide timely insights into the growth of underground tubers and the nutritional status of the crops [
3] and provide data support for irrigation and fertilizer application. Monitoring crop vegetation cover is highly important for crop management [
4]. To date, high-resolution remote sensing images have been widely used in many fields to obtain spatial information, providing a more accurate information source for vegetation monitoring [
5]. Remote sensing images facilitate the extraction and monitoring of vegetation information at the field scale.
Common remote sensing technologies used for monitoring crop growth status include proximal remote sensing, aerial remote sensing, and satellite remote sensing. Proximal remote sensing is not suitable for field-scale crop growth monitoring due to limitations in measurement speed [
6]. Satellite remote sensing can obtain large-scale crop field information and is therefore often used in agriculture. Ma et al. used Sentinel-2 remote sensing data to invert the leaf area index of wheat and adopted the SP-UCI optimization algorithm to establish a wheat yield estimation model [
7]. Worrall et al. combined satellite remote sensing data with a domain-oriented neural network to estimate the growth stages of maize [
8]. Shamsuzzoha et al. utilized Landsat-8 image datasets to determine rice RGVI changes caused by cyclones and established machine learning models for predicting major land use phenological changes in rice crops [
9]. Although satellite remote sensing can process large-area crop data, its practical application in crop phenotypic information like vegetation coverage is limited due to long measurement cycles, low resolution of captured data, and complex data processing.
Drone remote sensing is widely applied in crop growth monitoring due to its high image resolution and flexible advantages. Drone remote sensing technology can be categorized into hyperspectral remote sensing, multispectral remote sensing, and visible light remote sensing. Hyperspectral and multispectral remote sensing can acquire more detailed crop information compared to visible light remote sensing [
10]. However, these technologies come with higher costs and more complex data processing requirements [
11]. Although drone visible light remote sensing images contain less information, which limits their application in crop growth monitoring, visible light cameras are more cost-effective than hyperspectral and multispectral cameras, and most mainstream drones are equipped with visible light cameras [
12]. Therefore, exploring methods to monitor crop growth using drone visible light remote sensing data is of significant importance for reducing agricultural production costs and promoting the use of drones in the agricultural sector.
Estimating crop FVC based on UAV remote sensing technology commonly uses machine learning and threshold-based methods [
13]. Although machine learning methods provide higher accuracy in estimating crop FVC, they often require a large number of samples for training to improve the accuracy in large-scale crop coverage estimation. This makes sample selection inefficient and highly susceptible to the influence of human sample selection [
14]. Threshold-based methods are characterized by their simplicity, efficiency, and accuracy; however, the effectiveness of threshold extraction greatly affects the estimation accuracy of crop FVC, making the precision of threshold extraction crucial for the threshold-based estimation of FVC [
15]. The Otsu method is a commonly used technique for determining thresholds and has been widely applied in crop classification, but it can result in under-segmentation in certain situations [
16]. Additionally, the resolution of UAV multispectral remote sensing images is relatively low, making the Otsu method unsuitable. The bimodal histogram method and maximum entropy threshold method are commonly used for estimating crop FVC [
17]. For remote sensing images containing only crops and soil background, there is usually a single peak distribution, whereas for complex images with more than two peak distributions, the bimodal histogram method is not suitable. Noise in the images can affect the maximum entropy value, leading to lower accuracy in FVC estimation [
18]. Therefore, selecting a threshold extraction method that is stable, accurate, and easy to operate is essential for estimating crop FVC.
The Gaussian mixture model (GMM) threshold method can effectively address the issues present in the aforementioned threshold methods [
19]. This method assumes that the target object and background follow a GMM distribution in a color feature and uses the intersection of the GMM as the classification threshold [
20]. The GMM threshold method can achieve ideal crop FVC estimation accuracy to some extent. However, using color features for threshold determination often results in low FVC estimation accuracy when crop plants are small or during the near-closure period. Additionally, the GMM threshold method requires appropriate samples to determine the classification threshold, and relying solely on manually selected samples reduces the efficiency of the algorithm [
21]. Combining machine learning with the GMM threshold method can achieve rapid FVC estimation and better accuracy [
22]. The relationship between the number of samples selected by machine learning and the balance between FVC estimation accuracy and efficiency has not yet been explored, and whether the texture features of crop plants and the background conform to a GMM distribution remains unknown. Therefore, further research on the GMM threshold method is needed to propose a stable, efficient, and high-accuracy crop FVC estimation method.
The main research objectives are as follows: (1) to construct a new vegetation index capable of achieving high-precision vegetation cover extraction for potatoes throughout their entire growth period by combining machine learning and threshold extraction methods to determine its classification threshold; (2) to select vegetation indices and texture features that have a high correlation with FVC and achieve high-precision extraction of potato vegetation cover based on the classification threshold; (3) to establish a vegetation cover estimation model for the entire growth period of potatoes.
2. Materials and Methods
2.1. Experimental Design
To increase sample diversity and improve the reliability of the FVC estimation model, two experimental areas were designed for field experiments. Experiment Area A is located in the Yangling Demonstration Zone, Shaanxi Province (latitude 34°18′10″ N, longitude 108°5′14.33″ E). The soil pH, organic matter, and nitrogen concentration in this experimental area were 7.2, 27.31 g/kg, and 1.96 g/kg, respectively. Potatoes of the Jinxu 16 variety were planted in May 2021, with planting spacing, row spacing, and planting depths of 0.6 m, 0.5 m, and 10 cm, respectively, and the potatoes were harvested in September 2021. Five nitrogen levels were set (N1: 0 kg/ha; N2: 75 kg/ha; N3: 150 kg/ha; N4: 225 kg/ha; and N5: 300 kg/ha). Each treatment was repeated five times, totaling 25 plots, each measuring 7.5 m in length and width. Five ground control points were systematically set around the experimental area for the georeferencing of multiperiod images and the preprocessing of stitched images.
Experiment Area B is located in Ningtiaoliang town, Jingbian County, Yulin city, Shaanxi Province (latitude 37°33′55.60″ N, longitude 108°22′7.46″ E). The soil pH, organic matter, and nitrogen concentration in this experimental area were 7.7, 16.35 g/kg, and 0.89 g/kg, respectively. Three early maturing potato varieties (Hisen 6, 226, and V7) were used in this experiment, with five nitrogen levels (N1: 0 kg/ha; N2: 60 kg/ha; N3: 120 kg/ha; N4: 180 kg/ha; and N5: 240 kg/ha). Each treatment was replicated three times, totaling 15 plots, each measuring 112 m
2 (14 m × 8 m). The remaining treatments were the same as those in Experiment Area A. Potatoes were planted in May 2022 and harvested in August 2022, with a total growth cycle of approximately 90 days. The information about Experiment Area B is shown in
Figure 1.
2.2. Visible Light Image Acquisition and Preprocessing
Considering the impact of sensors on estimating potato phenotypic information and nitrogen nutrition diagnosis and to further explore the applicability of the model, Experiment Area A utilized a DJI Phantom 4 RTK drone (SZ DJI Technology Co., Ltd., Shenzhen, China) to capture visible light images during the tuber formation and tuber enlargement stages of potatoes. The drone has an aperture range from f/2.8 to f/11, an ISO range of 100–3200 (automatic) and 100–12,800 (manual), a maximum photo resolution of 5472 × 3648, a flight time of 30 min, and a positioning accuracy (Real-Time Kinematic positioning mode) of 1.5 cm + 1 ppm vertically and 1 cm + 1 ppm horizontally. DJI GS RTK software was used for drone flight route planning, with the flight altitude set at 30 m and the flight speed set at 3 m/s. The drone captured images vertically downwards with an overlap of 85% in both heading and sideways. All flights were conducted at noon under clear, windless conditions. During the flights, five ground control points were visible in the images, and their GCP values were measured using a differential global positioning system (DGPS) to further correct the image positioning information. The ground resolution of the images is 0.95 cm/pixel.
Experiment B utilized the DJI Phantom 4 Pro drone to capture visible light images of potato canopies during the tuber formation, tuber enlargement, and tuber maturity stages. The drone has a shutter speed range of 1/2000–1/8000 s, with other parameters identical to those of the DJI Phantom 4 RTK drone. Atrtizure software was used for drone flight path planning, with flight parameters identical to those used in Experiment A.
Preprocessing of remote sensing images is an indispensable process for monitoring crop growth and nutrient diagnosis [
23]. Common preprocessing methods for UAV remote sensing technology include image stitching, image registration, and cropping [
24]. This study utilized Pix4D mapper software developed by the Swiss company Pix4D AG to stitch all visible light images. The main workflow was as follows: (i) importing images; (ii) importing control point coordinates for point cloud generation; (iii) applying a one-click automatic process for point cloud extraction and 3D model generation; and (iv) generating point cloud data, the DSM, and the DOM [
25]. ENVI software Version 5.3 (Exelis Visual Information Solutions, Boulder, CO, USA) was used for image registration of visible light, multispectral orthoimages, and elevation images. The corresponding experimental field areas for the entire growth period of potatoes were cropped using ENVI software.
2.3. Selection and Construction of Remote Sensing Feature Elements
Experiment B included more potato data covering various growth stages and a richer variety of potato species. The potential impact of variety on the accuracy of vegetation cover extraction was considered. This study focused on researching vegetation cover estimation methods using Experiment B and validated them using Experiment A. To reduce the influence of noise on data processing, convolutional low-pass filtering was applied to denoise the visible light images of potatoes captured by drones throughout their growth period. The convolution kernel size was set to 5, with a weighted return value of 0. After denoising, vegetation index calculations and texture feature extraction were performed.
2.3.1. Selection and Construction of Visible Light Band Vegetation Indices
Previous studies have constructed numerous vegetation indices based on visible light images to assess crop vegetation cover. This study selected eight vegetation indices, namely, the GRVI, EXG, RGBVI, MGRVI, NGRVI, NGBDI, GLI, and TRVI [
26,
27,
28,
29]. However, the construction process of these common vegetation indices in the visible light band did not adequately consider the interrelationships among the three bands. Changes in individual bands often had a significant impact on the constructed vegetation indices. Therefore, the construction of new vegetation indices using multiband combinations can significantly improve the effectiveness of cover detection.
The visible light remote sensing images of the tuber enlargement stage of potatoes in Experiment B were visually interpreted. The experimental field encompassed various land cover types, contributing to a complex field environment. To effectively distinguish between these land cover types, 100 regions of interest (ROIs) were selected for each type using ENVI software. Subsequently, the characteristic values of the blue, green, and red bands for each land cover type were calculated and summarized, as shown in
Table 1.
As shown in
Table 1, there is an overlap between the grayscale values of potato plants in the blue band and shadows and between the grayscale values in the green band and those in the drip irrigation tape band. Additionally, there is a partial overlap in the grayscale values in the blue band with both shadows and drip irrigation tape. It is difficult to distinguish potato plants using a single band. Scatterplots of red-green-blue for potato plants, soil, shadows, and drip irrigation tape (
Figure 2) were generated.
Figure 2 clearly shows that there are distinct boundaries between potato plants and other land cover categories under the red-green and blue-green combinations. Potato plants are mainly concentrated in the lower-left region of the scatterplots, while other land cover categories are predominantly located in the upper-right region. However, the blue-red combination was unable to identify potato plants within the shadows.
A total of 25 points were selected from the scatterplot boundary lines between plants and shadows and between soil in
Figure 2a,b for linear fitting. The fitting results are shown in
Figure 3a,b.
The fitted boundary functions were then used to construct new combination indices: the combination index of blue and green bands (BGCI) and the combination index of red and green bands (RGCI). The corresponding formulas are shown as Equation (1) and Equation (2), respectively, as follows:
2.3.2. Texture Feature Extraction
Texture is a visual characteristic that reflects the homogeneous appearance of crops, embodying the structural arrangement attributes of periodic changes on the surface of objects in the image [
30]. Currently, texture features have been widely applied in crop classification and prediction [
31]. The gray level co-occurrence matrix (GLCM) is a common method for calculating image texture features, mainly by extracting texture features through the calculation of conditional density probability functions of the grayscale levels of objects in the image [
18]. In this study, the ENVI software was used to calculate eight common texture features, namely, the mean, variance, homogeneity, contrast, dissimilarity, entropy, angular second moment, and correlation, within a 7 × 7 window.
2.4. Fraction Vegetation Coverage Extraction Method
2.4.1. Determination of Extraction Thresholds
The vegetation index intersection method can effectively achieve vegetation coverage extraction. However, the common applications and vegetation indices of the vegetation index intersection method are currently designed for specific environments. Further investigations are needed to determine the vegetation coverage of potatoes throughout their entire growth period [
32]. In this study, the threshold determination process based on the vegetation index intersection method was divided into two main parts; one part involved clipping the entire experimental potato field and performing supervised classification on the clipped area. The other part combined the supervised classification results with the Gaussian mixture model to further determine the extraction threshold for potato vegetation coverage. This study presents solved thresholds for the constructed RGCI, BGCI, and eight other common vegetation indices solved using the vegetation index intersection method.
The support vector machine (SVM) is currently widely used in remote sensing image classification and effectively addresses issues such as small samples, nonlinearity, and high dimensionality. It also has strong generalization capabilities, with the type of kernel function being the radial basis function. To efficiently process sample data, the SVM was used for supervised classification of the cropped images. The cropping range of the region was set to 7 m × 7 m. The specific process for supervised classification using the SVM was as follows:
- (1)
The orthophoto images of the potato field were cropped and 40 regions of interest (ROIs) of potato plants and the background on the orthophoto images were selected. The separability of the selected samples was calculated.
- (2)
Based on the separability of the samples from the three growth stages of potatoes, the reasonableness of the selected samples was assessed.
- (3)
Additionally, 30 ROIs of potato plants and the background were selected to verify the SVM classification results using a confusion matrix.
2.4.2. Validation of Vegetation Coverage Extraction Accuracy
Using the determined thresholds, the vegetation coverage of potatoes was extracted for each growth stage. In this study, 300 regions of interest (ROIs) of potato plants and the background were selected outside the cropped areas of the potato experimental field using visual interpretation. A confusion matrix was used to validate the vegetation coverage extracted by the vegetation index intersection method for the three potato growth stages. The validation was evaluated using the kappa coefficient and overall accuracy.
2.5. Establishment of the Fraction Vegetation Coverage Estimation Model
2.5.1. Vegetation Index and Texture Feature Selection
Various potential covariates can be used for vegetation coverage estimation, but applying all available covariates can reduce the data processing efficiency. Therefore, it is necessary to filter out important features before establishing the estimation model.
Based on the best potato vegetation coverage results from Experiment B, three sampling areas were selected from each of the fifteen experimental plots to measure vegetation coverage during the three growth stages. The visible light image consists of three bands, red, green, and blue, each containing eight texture features. To facilitate presentation, the 24 texture features are shown in
Table 2. The mean values of the corresponding 10 vegetation indices and 24 texture features were calculated for each sampling area.
To improve the computational efficiency of the model, this study utilized random forest and Pearson correlation coefficient methods to select 10 vegetation indices and 24 texture features and obtained important remote sensing factors. Random forest (RF) is an algorithm that measures feature importance by randomly replacing each feature [
33]. When a feature is highly important, the prediction error rate of the RF model will increase. The change in the error rate of the out-of-bag data before and after feature replacement is applied to evaluate each feature, and their importance scores are obtained. This study used the Python 3.6 library (Scikit Learn package) to implement this algorithm and selected, repeated tenfold cross-validation to optimize the accuracy of the RF model. Subsets with minimal impact on the random forest model were used as subsets of predictors.
The Pearson correlation coefficient was used to detect the linear correlation between continuous variables, with values ranging from −1 to 1; positive and negative values indicate positive and negative correlations, respectively. The larger the absolute value is, the greater the linear correlation.
2.5.2. Establishment and Validation of the Fraction Vegetation Coverage Estimation Model
This study used RF and Pearson correlation coefficients to assess the importance of vegetation indices and texture features. Based on the top six features selected by these algorithms, a model for estimating the FVC was constructed. For each potato growth stage, a total of 45 vegetation coverage extraction results were obtained. Thirty randomly selected data points were used to construct the vegetation coverage estimation model, while the remaining fifteen data points were used for model validation. Linear fitting was used to establish a vegetation coverage estimation model for potato plants and validate its performance.
Common model evaluation metrics include the coefficient of determination (R
2) and root mean square error (RMSE) [
34]. A higher R
2 value and a smaller RMSE value indicate higher model accuracy. In this study, the evaluation metrics R
2 and RMSE were used to validate the performance of the models. The formulas for calculating R
2 and RMSE are shown in Equations (3) and (4), respectively, as follows:
In the equations, represents the measured values; represents the predicted values; represents the mean of the measured values; and represents the sample size.
3. Results
3.1. Extraction of FVC Thresholds Based on the Vegetation Index Intersection Method
To determine the effectiveness of the SVM in classifying plants and the background in images, the classification performance of the SVM was validated using a confusion matrix, as shown in
Figure 4. The overall classification accuracy for the experimental field during the tuber formation stage was 99.922%, with a kappa coefficient of 0.9982. During the tuber enlargement stage, the overall classification accuracy was 99.9871%, with a kappa coefficient of 0.9997. For the maturity stage, the overall classification accuracy was 99.8899%, with a kappa coefficient of 0.9977. The validation results from the confusion matrix indicate that the SVM achieved high classification accuracy in the small-potato regions across all three stages. The overall classification accuracy for the three growth stages of potatoes was greater than 99%, and the kappa coefficient was greater than 0.99. Therefore, this method can be used to determine the FVC extraction threshold for potatoes.
The extraction thresholds for various vegetation indices were calculated based on the vegetation index intersection method. Since the newly constructed vegetation indices, RGCI and BGCI, are similar in principle to the EXG vegetation index, this study focused on comparing the threshold extraction results of these three indices. As shown in
Figure 5,
Figure 6 and
Figure 7, the EXG histograms exhibit irregularities during the three growth stages of the potatoes. In particular, in the tuber formation stage, multiple intersections in the EXG histogram prevent the effective determination of the extraction threshold for the potato vegetation coverage. The newly constructed RGCI and BGCI vegetation indices effectively address the irregularity issues of the EXG index and can better determine the classification thresholds for the background and potato plants. The coverage extraction thresholds of the other selected vegetation indices can also be determined via the vegetation index intersection method. The extraction threshold results for each vegetation index corresponding to the three growth stages of potatoes are shown in
Table 3.
3.2. Extraction Results of Potato FVC
The determined thresholds were applied to extract the vegetation coverage of the entire experimental field for the corresponding three growth stages of potatoes. Pixels with values greater than the threshold were classified as potato plant pixels, while those with values less than the threshold were classified as background pixels. Further calculations were performed using Formula 5 to calculate the vegetation coverage of the potatoes. The extraction results of potato vegetation coverage for the three stages are shown in
Figure 8.
In the formula, , , and represent potato vegetation coverage, the number of potato plant pixels, and the number of soil pixels, respectively.
Figure 9 shows the accuracy verification results of potato vegetation coverage extraction based on the vegetation index intersection method. During the three growth stages of the potatoes, except for the GRVI, which had lower accuracy in extracting vegetation coverage during the tuber formation stage with an overall accuracy below 90% and a kappa coefficient below 0.5, the remaining vegetation indices achieved higher accuracy in estimating vegetation coverage.
During the tuber formation stage, the BGCI achieved the highest accuracy in terms of vegetation coverage extraction among all the vegetation indices, with an overall accuracy of 99.6079% and a kappa coefficient of 0.9898. Similarly, during the tuber enlargement stage, the BGCI achieved the highest accuracy in terms of vegetation coverage extraction, with an overall accuracy and kappa coefficient of 98.8405% and 0.9753, respectively. In the potato maturity stage, the RGCI achieved the highest accuracy in estimating vegetation coverage, with an overall accuracy and kappa coefficient of 98.6336% and 0.9712, respectively.
Through comparison, the RGCI and BGCI demonstrated better accuracy in extracting potato vegetation coverage based on the vegetation index intersection method than did the other vegetation indices. Comparing the vegetation coverage extraction results of the first two stages, the overall accuracy of the BGCI was higher than that of the RGCI by 0.1741% and 0.4866%, respectively. However, in the potato maturity stage, the overall classification accuracy of the BGCI was lower than that of the RGCI by 0.7719%.
3.3. Potato Vegetation Coverage Estimation Model
To ensure the effective estimation accuracy of the potato FVC estimation model throughout the entire growth period, this study selected the BGCI combined with the vegetation index intersection method to extract vegetation coverage during the tuber formation and tuber enlargement stages of the potatoes. The RGCI combined with the vegetation index intersection method was chosen to extract vegetation coverage during the maturity stage. The extraction results are shown in
Figure 10.
During the feature selection process for individual potato growth stages, the limited amount of data from each stage may lead to nonrepresentative feature selection results. Therefore, this study conducted feature selection for the entire growth period of the potato plants. As shown in
Figure 11, the Pearson correlation coefficient method over the entire potato growth period revealed that the BGCI and RGBVI exhibited the highest correlation with FVC (R = 0.92), followed by the EXG and RGCI, with correlation coefficients of 0.91 for both indices. Next in line were the NGRVI (R = 0.90) and NGBDI (R = 0.89).
Among the 24 texture features, B-correlation was the most highly correlated feature with FVC (R = 0.73), but its correlation coefficient was still less than 0.75. Therefore, the results from the Pearson correlation coefficient method indicated that texture features are not suitable for establishing the FVC estimation model for the tuber formation stage of potatoes.
The results of the random forest feature selection also showed that the BGCI obtained the highest score, with an importance of 0.1969. Next in line were the NGBDI, RGCI, RGBVI, NGRVI, and EXG, with importances of 0.1241, 0.0894, 0.0554, 0.0483, and 0.0332, respectively. These scores accounted for less than 0.1 of the total importance. Although the feature importance ranking from random forest selection was not entirely consistent with that of the Pearson correlation coefficient method, both methods identified the MGRVI, BGCI, TRVI, NGRVI, RGCI, and NGRVI as the six remote sensing feature factors with relatively high correlations to potato FVC. These selected vegetation indices were used to estimate potato vegetation coverage during the tuber formation stage.
Using the six selected vegetation indices, the estimation of the vegetation coverage during the potato tuber formation period was carried out, and the estimation results were analyzed and evaluated. Evaluating the accuracy of FVC estimation of the model using R
2 and RMSE values, it was observed that with the increase in vegetation coverage data volume throughout the potato growth period, the BGCI and RGCI achieved better accuracy in FVC estimation. As shown in
Figure 12, the BGCI obtained the highest FVC estimation accuracy during the entire potato growth period (R
2 = 0.9116, RMSE = 5.7903), followed by the EXG (R
2 = 0.9065, RMSE = 5.8669) and RGCI (R
2 = 0.8987, RMSE = 5.8633). The NGRVI had the lowest FVC estimation accuracy (R
2 = 0.7175, RMSE = 9.7841).
5. Conclusions
This study presents an analysis of the differences in grayscale values between potato plants and the background in the visible light spectrum in the field via the construction of RGCI and BGCI vegetation indices based on the relationships in the red-green and blue-green bands. The computation processes of eight common visible light spectrum vegetation indices and twenty-four texture features were described. The potato FVC was extracted using the vegetation index intersection method, maximum entropy thresholding method, and bimodal histogram method. The accuracy of 10 vegetation indices combined with three FVC extraction methods was validated using a confusion matrix, and the best vegetation index and potato FVC extraction method were further selected. The Pearson correlation coefficient method and random forest feature selection were used to screen the aforementioned vegetation indices and 24 texture features. Based on the top six selected features, potato FVC estimation models were established. The main conclusions are as follows:
(1) This study newly constructed two vegetation indices, BGCI and RGCI, and successfully obtained classification thresholds using the SVM combined with vegetation indices, which can effectively differentiate between the background and potato plants. The extraction thresholds for the BGCI during the three growth periods were −13.0583, 10.1801, and −4.3000, respectively. For the RGCI, the extraction thresholds during the three growth periods were 2.5892, 23.0584, and 16.9357, respectively. The BGCI and RGCI could effectively distinguish potato plants from the background under the above thresholds.
(2) The BGCI and RGCI combined with the vegetation index intersection method both achieved excellent results in the extraction of potato vegetation coverage throughout the entire growth period. During the potato tuber formation and expansion stages, the BGCI combined with the vegetation index intersection method achieved the highest vegetation coverage extraction accuracy, with overall accuracies of 99.61% and 98.84%, respectively. The RGCI combined with the vegetation index intersection method achieved the highest accuracy, 98.63%, in terms of vegetation coverage extraction during the maturation stage. Overall, the RGCI combined with the vegetation index intersection method obtained the most ideal vegetation coverage extraction results throughout the entire potato growth period.
(3) This study screened multiple vegetation indices and texture features and successfully established a highly accurate potato vegetation coverage estimation model. Using the Pearson correlation coefficient method and random forest feature selection, six vegetation indices highly correlated with potato FVC (BGCI, NGBDI, RGCI, RGBVI, NGRVI, and EXG) were selected, and corresponding vegetation coverage estimation models were constructed. Among these models, the vegetation coverage estimation model based on the BGCI exhibited the highest accuracy (R2 = 0.9116, RMSE = 5.7903). In the validation of model generality, the FVC estimation models based on the newly constructed BGCI and RGCI also achieved good accuracy. The RGCI had an R2 value of 0.9497 and an RMSE of 2.2548, while BGCI had an R2 value of 0.9557 and an RMSE of 2.1486. This study proposed two novel vegetation information indices, and the potato vegetation coverage models based on these indices demonstrated good model accuracy and generality. The high accuracy of FVC estimation and the general applicability of the model are important for the real-time detection of potato vegetation coverage. Through high-precision coverage estimation at any time, the actual growth condition of potato can be understood in time, so that the input of potato water and fertilizers can be controlled in a timely and effective manner. This is important for guiding the field management of tuber crops to improve potato yields.