Next Article in Journal
Fault Diagnosis of Rolling Bearings in Agricultural Machines Using SVD-EDS-GST and ResViT
Previous Article in Journal
Crop Density and Sowing Timing Effect on Tan Spot Occurrence in Spring Wheat
Previous Article in Special Issue
Spectral-Frequency Conversion Derived from Hyperspectral Data Combined with Deep Learning for Estimating Chlorophyll Content in Rice
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Refinement of Cropland Data Layer with Effective Confidence Layer Interval and Image Filtering

1
SNARS Laboratory, School of Instrumentation and Optoelectronic Engineering, Beihang University, Beijing 100191, China
2
Meteorological Department of West Azerbaijan Province, Iran Meteorological Organization (IRIMO), Orumiyeh 670056, Iran
3
School of Mechanical Engineering, Zhejiang University, Hangzhou 310030, China
*
Author to whom correspondence should be addressed.
Agriculture 2024, 14(8), 1285; https://doi.org/10.3390/agriculture14081285
Submission received: 25 June 2024 / Revised: 30 July 2024 / Accepted: 2 August 2024 / Published: 4 August 2024

Abstract

:
Various systems have been developed to process agricultural land data for better management of crop production. One such system is Cropland Data Layer (CDL), produced by the National Agricultural Statistics Service of the United States Department of Agriculture (USDA). The CDL has been widely used for training deep learning (DL) segmentation models. However, it contains various errors, such as salt-and-pepper noise, and must be refined before being used in DL training. In this study, we used two approaches to refine the CDL for DL segmentation of major crops from a time series of Sentinel-2 monthly composite images. Firstly, different confidence intervals of the confidence layer were used to refine the CDL. Secondly, several image filters were employed to improve data quality. The refined CDLs were then used as the ground-truth in DL segmentation training and evaluation. The results demonstrate that the CDL with +45% and +55% confidence intervals produced the best results, improving the accuracy of DL segmentation by approximately 1% compared to non-refined data. Additionally, filtering the CDL using the majority and expand–shrink filters yielded the best performance, enhancing the evaluation metrics by about 1.5%. The findings suggest that pre-filtering the CDL and selecting an effective confidence interval can significantly improve DL segmentation performance, contributing to more accurate and reliable agricultural monitoring.

1. Introduction

The agricultural sector has historically met food production demands; however, concerns about its ability to sustain this role for an expanding global population persist in food security discussions [1]. While significant strides have been made in reducing hunger globally, substantial challenges remain [2,3]. Achieving effective situational awareness in food production necessitates a comprehensive understanding of the earth as an integrated system, a capability that is uniquely provided by satellite technology [4]. Satellites offer the ability to map farmland areas and crop types, estimate planted areas and yields and provide early warnings for droughts and floods. Consequently, numerous monitoring systems have been developed in recent years to utilize these satellite data. Among these, Cropland Data Layer (CDL), published annually by the United States Department of Agriculture (USDA), stands out by offering 30 m resolution crop type maps derived from extensive labeled training samples [5,6,7]. A comprehensive overview of various global agricultural monitoring systems is available in the work of Fritz, et al. [8].
The CDL is widely regarded as the primary source for ground-truth data in machine learning model training owing to its advanced system architecture and adherence to geospatial web services standards within a publicly accessible web environment [9]. Numerous studies affirm the CDL’s credibility as a reliable ground-truth for crop mapping [10,11,12,13]. Currently, the CDL represents the most comprehensive national crop type map available for free use and download, offering exceptional overall accuracy for major crops, particularly corn and soybeans, with accuracy rates exceeding 95% [5]. However, several studies have highlighted defects within the CDL, primarily due to its reliance on remote sensing datasets, which raises concerns regarding the quality of CDL products [9,14,15,16,17,18]. To address these concerns, the USDA National Agricultural Statistics Service (NASS) has provided an accuracy rating since 2008 and introduced the confidence layer in 2017 to enhance data reliability [5].
The use of various CDL confidence layer percentages has been explored in some studies to utilize the CDL as ground-truth data [13,19]. For instance, Hao et al. [13] employed a 95% confidence interval, rejecting all CDLs below this threshold. However, there is no clear justification for selecting the 95% threshold, and similar studies lack a rationale for their chosen confidence intervals. Additionally, a review of the CDL reveals significant errors, some resembling salt-and-pepper noise, even within the 100% confidence interval. These inaccuracies negatively impact deep learning (DL) data outcomes, necessitating the removal of erroneous CDLs through refining techniques. While some studies [9,14] have attempted to address this issue, fully accurate results have yet to be achieved, indicating a need for further research and the development of alternative approaches.
To refine the CDL and improve DL training outcomes, this research had two primary objectives. The first objective was to establish effective confidence intervals for refining the CDL to enhance DL segmentation accuracy. Various confidence interval values were applied to the CDL of major crops in the study area, and the resulting data were used as the ground-truth in a DL model. The effectiveness of these intervals was analyzed using multi-criteria decision analysis (MCDA) to determine the best approach for future applications. Additionally, the generalizability of these confidence intervals was tested in a different study area, confirming the robustness of our method across different geographical regions. The second objective was to evaluate the effectiveness of different image filtering techniques in mitigating errors within the CDL. By applying these filters, we assessed their contribution to improving the quality of the CDL and subsequently enhancing DL model accuracy for mapping croplands using Sentinel-2 imagery. This research also involved comparing our methods with the refined CDL (R-CDL) from Lin et al. [9], further demonstrating the efficiency and practicality of our approach.

2. Materials and Methods

2.1. Study Area

In this study, we focused on a specific region within the Mississippi Delta, highlighted in Figure 1, selected for its agricultural diversity and distinct seasonal phenological stages [20]. In 2021, soybeans were predominantly cultivated in this area, covering 31.7% of the landscape, along with significant portions of corn, cotton and rice, illustrating its agricultural variety. This diverse region, covering approximately 575,908 square kilometers and featuring various land cover types and surface features, provides an ideal environment for evaluating CDL refinements and the effectiveness of image filtering techniques in improving DL segmentation accuracy for precise agricultural monitoring.
To evaluate the generalizability of our method and conclusions, we conducted additional testing in a different geographical region named T14TNK, identified by its UTM Zone 14 within the Military Grid Reference System (MGRS). This area, covering approximately 100 km by 100 km, is located between Nebraska and Kansas, defined by the coordinates 97°42′ W–99°0′ W and 39°40′ N–40°38.5′ N, and includes parts of the Republican River. The major crops in this region are corn and soybeans, making it an ideal candidate for validating our methodology across different agricultural landscapes. By applying the same confidence interval and image filtering techniques, we aimed to assess the consistency and reliability of our approach. The results from this additional region further support the applicability of our method to diverse agricultural settings.

2.2. Materials

2.2.1. Sentinel-2 Imagery

To utilize satellite images, we selected multispectral images from the European Sentinel-2 sensors. These images are widely used in large-scale vegetation extraction, both in research and various application systems, due to their comprehensive optical spectrum for crop identification and acceptable spatial resolution compared to the CDL [21,22]. The red (B4), green (B3), blue (B2) and near-infrared (B8) bands have a spatial resolution of 10 m. Additionally, the three red-edge bands (B5, B6 and B7), two short-wave infrared bands (B11, B12) and one narrow NIR band (B8a) offer spatial resolutions that can exceed 20 m, adding extra band attributes for classification [21,22,23,24]. Bands 1, 9 and 10 correspond to coastal aerosols, water vapor and cirrus clouds, respectively [22]. These bands have a 60 m resolution, which makes accurate recording of the reflectance of minor surface objects challenging. Therefore, after excluding bands 1, 9 and 10, we selected a total of 10 bands for our investigation.
Utilizing Google Earth Engine (GEE) [21,23], we processed monthly minimum composite datasets from May to September, comprising 872 tiles in Mississippi Delta and 29 tiles in T14TNK. This period aligns with the peak growing season in the study area, as indicated by the Normalized Difference Vegetation Index (NDVI) [25] in Figure 2. To provide a comprehensive view of vegetation dynamics, NDVI curves for the entire year of 2021 were generated, contrasting growing and non-growing periods. The NDVI for each major crop was calculated using Sentinel-2 satellite imagery, with cloud masking applied to remove cloudy pixels. The data were composited into 15-day intervals over 2021, and NDVI was computed. Stratified random sampling selected 10,000 sample points for each major crop based on the CDL with 100% confidence interval, extracting NDVI values to provide a detailed temporal profile for each crop.

2.2.2. Cropland Data Layer (CDL)

Training and validation ground-truth data were derived from the CDL. The USDA—NASS produces the CDL, which offers an annual crop type for the United States at a 30 m resolution with more than 100 crop classes [6]. Since 2006, the state-level CDL classification has been conducted using the supervised RuleQuest See5.0 [26] decision tree classification approach [27,28]. Each US state’s metadata include producer accuracy (PA), user accuracy (UA) and Kappa score, and the CDL is evaluated using a subset of ground-truth data. According to Boryan et al. [27], the primary crops in the CDL have excellent classification accuracy ranging from 85% to 95% [5]. The chart in Figure 3 displays the distribution of various land cover types in Mississippi Delta, highlighting major crops such as soybeans, corn, cotton and rice. Additionally, we included the “other” class to account for non-crop areas and different crop types from major crops.
To compare the effectiveness of our method, we utilized refined CDL data based on the methodology outlined in Lin et al.’s study [9]. This study aimed to identify and resolve inaccuracies in CDL crop classifications by employing a decision tree method to detect and refine questionable pixels using spatial and temporal crop information. The refined CDL (R-CDL) data were validated against high-resolution satellite images and USDA’s official acreage estimates, ensuring their robustness. By incorporating this refined dataset, we were able to benchmark our proposed methods against a well-validated standard, providing a comprehensive evaluation of our approach in relation to existing techniques.

2.2.3. CDL Confidence Layer

The confidence layer is an additional layer provided alongside the CDL, which was evaluated in this study. This layer geographically indicates the expected confidence associated with each output pixel based on the classification rules applied. It offers a visual representation of the distribution and magnitude of classification confidence or error, which is highly beneficial for users. The confidence layer represents classification accuracies as a confidence interval, where a rating of 100 signifies very high confidence, and a rating of 0 signifies very low confidence. Essentially, the confidence value measures how well a pixel fits into the decision tree ruleset, rather than the absolute accuracy of a specific pixel [5].
By utilizing the confidence layer, users can better assess the reliability of the CDL for different applications. In this study, we employed various confidence interval thresholds to refine the CDL, which was then used as the ground-truth for training DL models. The impact of these confidence intervals on the quality of training data and DL performance was systematically evaluated to identify the most effective threshold for accurate crop segmentation from Sentinel-2 imagery.

2.3. Research Approaches

This section delineates the research methodology, with a flowchart in Figure 4 summarizing the essential steps, inputs and outputs. The inputs include monthly minimum composites of Sentinel-2 time series during the growing season, while the outputs encompass various DL segmentation evaluation metrics and maps of major crops under different conditions of confidence layer data and CDL refinements. To enhance the quality of the CDL for use as the ground-truth in the DL, two approaches are proposed: utilizing different confidence intervals of the CDL confidence layer and applying various image filters to mitigate noise and errors. Initially, CDL crop data are refined using specified confidence intervals and overlaid with Sentinel-2 imagery to create a comprehensive database. The DL model is then trained on these datasets, with accuracy metrics compared through MCDA. In the second approach, different image filters are applied to the major crops of the CDL. These refined datasets are subsequently used to train the DL model, and the results are compared to identify the most effective refinement technique.

2.3.1. CDL Refinement Using the Confidence Layer

Given the focus on major crops, the initial step involves extracting crop data with varying confidence levels by querying the CDL under different confidence level values. For example, a +75% confidence layer includes crop data with confidence intervals of 75% and higher (e.g., 95%) while excluding data with confidence values below 75%. This method retains higher quality data and discards lower quality data. However, removing low-quality data can sometimes weaken the dataset for DL applications, whereas using all available data may negatively impact the DL by incorporating erroneous data. Numerous studies have explored varying confidence intervals within the confidence layer [13,19], while others have opted to utilize the entire dataset [20]. A critical aim of this research is to determine the most effective confidence interval for refining the CDL as ground-truth data in DL cropland segmentation. To address this, an experiment was designed using crop layer data with confidence intervals from 0% to 100%.

2.3.2. CDL Refinement Using Image Filters

Despite the CDL having defined confidence levels, substantial amounts of erroneous data persist, even at the 100% confidence level [9]. Visual inspection of data in respect of different crops reveals defects such as salt-and-pepper noise and errors [14,17]. These issues are apparent across various confidence intervals. Additionally, the CDL erroneously classifies field borders and roads as croplands, contributing to significant noise and inaccuracies. To address the pervasive issues of noise and errors within the CDL, the next step involves refining the CDL using an image filter to establish a more accurate ground-truth for DL applications. The aim of this process is to eliminate noise, incorrect border areas between fields and other unknown defects. Initially, each major crop from the CDL was separated, and various image filters were applied individually, including the following:
  • Aggregation [29]: This filter smooths the data by combining small, isolated patches of classified pixels into larger, contiguous regions, thereby reducing salt-and-pepper noise.
  • Boundary cleaning [30]: This filter simplifies the raster by smoothing zone boundaries using expansion (dilation) and shrinking (erosion) techniques. This process involves evaluating each input cell based on its orthogonal and diagonal neighbors to achieve the desired smoothing effect.
  • Expand [31]: The expand filter increases the size of designated zones within a raster by a certain number of cells. These designated zones are treated as the foreground, while the rest are considered the background, allowing the foreground to grow into the surrounding background areas.
  • Shrink [32]: The shrink filter designates specific zone values as the foreground, enabling their expansion into surrounding background zones. This process effectively reduces noise and minor misclassifications, resulting in a more accurate and conservative delineation of crop extents.
  • Majority [33]: This filter reclassifies each pixel based on the majority class of its neighbors, helping to correct misclassified pixels by considering local context.
  • Expand–shrink: This technique involves initially shrinking the classified regions in an image to remove noise and small misclassified areas and then expanding the remaining areas back to their original size. This process helps in retaining the core areas of the classified regions while minimizing the impact of noise and erroneous small regions.
After filtering, the refined major crop datasets were merged into a single comprehensive layer, similar to the original CDL. This process generated six distinct filtered datasets, each treated with different image filters. These improved data serve as a more reliable input for DL models, thereby enhancing the precision of cropland segmentation. These filtering techniques were selected for their ability to mitigate specific types of errors and noise present in the CDL.

2.4. Data Pre-Processing

A robust pre-processing workflow was applied to each monthly Sentinel-2 dataset using the GEE platform to optimize data for crop segmentation analysis. The initial step involved a cloud-and-shadow masking technique utilizing the satellite’s Scene Classification Layer (SCL) [34] and cloud probability data. This step effectively excluded atmospheric obstructions, enhancing image clarity. Following this, the workflow generated minimum composites for each month, selecting the pixel with the lowest reflectance value from all clear observations to minimize cloud interference. Despite a potential increase in shadow presence, this method prioritizes reducing cloud cover due to its significant impact on spectral data accuracy. Simultaneously, the major crops in the CDL were selected to focus on corn, cotton, rice and soybeans. This involved selectively masking the CDL to retain only these crops and clipping it to match the spatial extent of the Sentinel-2 composite. The 256 × 256 pixel grid sampling generated 977 samples in the study area and 211 in the test area, filtering out less informative samples to optimize data richness. The refined samples were randomly divided into training and validation groups in an 80–20% split and organized into batches of 64 for efficient DL model training.

2.5. Model Architecture and Training

For the crop segmentation task using Sentinel-2 imagery, the Pytorch-based DeepLabV3Plus [35] model with a ResNet34 encoder was employed. This model was selected due to its proven efficacy in semantic segmentation across various applications and ResNet34’s robust feature extraction capabilities [36]. The DeepLabV3Plus architecture, with its standard decoder and batch normalization, was adapted to suit the unique requirements of this dataset. The inputs consisted of 5 monthly minimum composites of Sentinel-2 imagery from May to September, each comprising 10 bands, resulting in a 50-layer input. This configuration captures the phenological changes in major crops. The model output was designed to classify five distinct classes corresponding to major crops (corn, cotton, rice, soybeans and other), optimizing it for detailed segmentation in the study area. The Adam optimizer [37] was used with a learning rate of 0.01. During training, cross-entropy loss was employed, with early stopping implemented after 50 epochs of no improvement to prevent overfitting.

2.6. Accuracy Metrics and Comparative Analysis

Overall accuracy (OA), user’s accuracy (UA), producer’s accuracy (PA) and F1-score (F1) [38,39] accuracy metrics were employed to evaluate the approaches.
O A = T P + T N Totali   Instances  
U A = T P T P + F P  
P A = T P T P + F N  
F 1 = 2 × U A × P A U A + P A  
where TP represents the number of true positives; FP represents the number of false positives; FN represents the number of false negatives; and TN represents the number of true negatives.
Also, a confusion matrix (error matrix) [38,40] is used to evaluate the performance of our DL models in classifying different crop types. A confusion matrix is a table that outlines the performance of the model by comparing the actual classes to the predicted classes. Each cell in the matrix represents the proportion of instances (expressed as a percentage) that were correctly or incorrectly classified by the model. Values range from 0 to 100, indicating the percentage of instances that fall into each category. For instance, a value of 95 in the matrix indicates that 95% of the instances of a particular class were correctly predicted, whereas a value of 0.005 might indicate a small proportion of misclassifications. These values provide a detailed insight into the model’s accuracy, UA, PA and overall performance in distinguishing between different crop types.
In addition, MCDA in this context involved evaluating multiple criteria (accuracy metrics) across various confidence interval alternatives in the DL metric results. Let A represent the set of confidence interval alternatives and C represent the set of criteria such as UA, PA and F1 [38,39]. Each criterion c C was normalized to bring different accuracy metrics onto a comparable scale. The normalization process for a criterion c for alternative a was calculated as
n a c = x a c a A x a c 2
where x a c is the original value of criterion c for alternative a, and n a c is the normalized value. Weights w c were assigned to each criterion c, reflecting its importance in the decision-making process:
w c = 1 C
The weighted normalized value for each criterion of each alternative was calculated as:
v a c = n a c   ·   w c    
Identification of the ideal best (A+) and ideal worst (A−) was as follows:
A c + = m a x a A   v a c   a n d   A c = m i n a A   v a c  
Calculation of the distance to ideal best (D+) and ideal worst (D−) was as follows:
D a + = c C ( v a c A c + ) 2   a n d   D a = c C ( v a c A c ) 2
The relative closeness of each alternative to the ideal solution was calculated as
R C a = D a D a + + D a  
Alternatives were then ranked based on their relative closeness to the ideal solution, with a higher R C a indicating a closer proximity to the ideal, and thus, a higher preference. MCDA systematically analyzes DL model performance across different confidence intervals and image filters, integrating metrics like OA, UA, PA and F1.
For the generalizability analysis, significance analysis experiments were conducted to quantify the similarity and statistical significance of the results between the Mississippi Delta and T14TNK areas. The Pearson correlation coefficient [40,41] and paired t-tests [39] were employed to compare the performance of different refining methods between these two regions. The Pearson correlation coefficient measures the strength and direction of the linear relationship between two variables. The paired t-test is used to determine whether the means of two paired samples differ significantly. The p-value associated with the T-statistic indicates whether the difference between the pairs is statistically significant. To ensure the comparison focuses on trends rather than absolute values, data normalization is performed.
T - s t a t i s t i c = d ¯ s d n
In the T-statistic, d ¯ is the mean of the differences between paired observations; s d is the standard deviation of the differences; and n is the number of pairs.
P e a r s o n   c o r r e l a t i o n   c o e f f i c i e n t = x i x ¯ y i y ¯ x i x ¯ 2 y i y ¯ 2
where x i and y i are the individual sample points, and x ¯ and y ¯ are the means of the sample sets x and y, respectively. A Pearson correlation coefficient close to 1 indicates a strong positive correlation, while a value close to −1 indicates a strong negative correlation.

3. Results

To implement the proposed approaches, the research was divided into three phases. The first phase involved evaluating the approaches using the CDL with varying confidence levels. In the second phase, different image filters were applied to the CDL, and the DL segmentation results were subsequently evaluated. In the third phase, the methodologies were tested on an additional test area, T14TNK, to compare and validate the effectiveness of the approaches in a different geographical and crop context.

3.1. Impact of CDL Confidence Intervals on Segmentation

The confidence interval percentages applied to the major crop of the CDL result in a progressive reduction in the number of crop pixels within the image. As the confidence interval increases toward 100%, the number of pixels representing crops correspondingly decreases. Figure 5 quantifies this reduction, displaying the area (in hectares) occupied by different crops at each confidence interval. Notably, in some scenarios, the pixel count can be reduced by up to two times. It is crucial to understand the impact of this reduction on machine learning applications, as it directly affects the accuracy and reliability of DL segmentation outcomes. Figure 6 illustrates the distribution of major crop pixels at various confidence intervals, showing how increased confidence thresholds progressively refine the CDL.
Table 1 displays the outcomes (OA, UA, PA and F1) of the DL implementation using different CDL datasets at varying confidence interval percentages. Additionally, the MCDA relative closeness values are provided to indicate the effectiveness of each confidence interval. The best overall performance can be observed at confidence intervals of +45%, +55% and +65%, with OAs of 90.5%–90.6%, indicating that these intervals effectively balance the inclusion of high-quality data while excluding erroneous data. Conversely, the lowest performances are seen at +85%–100% confidence intervals, with OAs less than 88.6%, indicating that excluding too much data adversely impacts the model’s performance.
For crop-specific performance, corn and soybeans exhibited the highest F1 at the 55% confidence interval, while cotton and rice achieved the highest F1 at the 45% confidence interval. The MCDA values corroborate these results, showing the highest relative closeness at the 45% and 55% intervals. This reinforces the conclusion that mid-range confidence intervals provide the effective balance by retaining high-quality data and excluding noise. These findings suggest that the 45% and 55% confidence intervals for the CDL strike the best balance, significantly enhancing the DL model’s accuracy and reliability for agricultural mapping applications.
The results of the DL segmentation, trained using data with various confidence intervals and the entire CDL dataset, are illustrated in Figure 7. In the segmentation results utilizing the entire dataset without the confidence layer, noise is effectively mitigated, and the original fields are clearly delineated. However, the method encounters challenges in accurately identifying roads and gaps between agricultural fields. In several instances, the segmentation erroneously classified non-cropland areas as cropland, which represents a significant limitation. Conversely, the results for the +85% and higher confidence intervals, while successfully removing noise, also erroneously exclude numerous agricultural fields.
Conversely, the segmentation results using the +35% to +75% confidence intervals are highly satisfactory. These datasets accurately depict the fields in their entirety, correctly identify gaps between the fields and eliminate non-agricultural areas. There are slight variations between the +45% and +55% confidence intervals; in some cases, the +45% interval outperforms, while in others, the +55% interval yields better results. This suggests that a confidence level within this range could provide the most effective segmentation performance, balancing accuracy and reliability.
The confusion matrices for different confidence intervals of the CDL in Figure 8 reveal various impacts on the performance of DL models in segmenting major crops. The All CDL matrix shows balanced performance across crops but contains some misclassifications, particularly in distinguishing corn and soybeans. As the confidence interval increases to +45% and +55%, there is a slight improvement in the identification of cotton and rice, evidenced by higher UA values in these categories. The +55% confidence level shows a noticeable enhancement in OA, especially for corn and soybeans, reducing false positives and false negatives significantly. The +65% interval continues this trend, with improved accuracy for cotton and rice. However, at the +95% and 100% confidence levels, there is a marked decline in performance across all crops, with increased misclassification rates, particularly for soybeans, indicating that overly stringent confidence levels may exclude valuable data, thereby impairing model accuracy.

3.2. Evaluating Filtered CDL

To address the discrepancies in CDL quality, various image filters were applied to enhance the accuracy of the DL models. The aim of these filters is to reduce noise and correct errors within the CDL dataset, ensuring a more reliable ground-truth for DL training. Table 2 presents the outcomes for each filter in terms of OA, UA, PA and F1 for the major crops. The “No Filter” and “R-CDL [9]” results in Table 2 are included for comparison. It should be noted that “No Filter” corresponds to “All CDL” in Table 1.
Applying different image filters to the CDL significantly improved DL model performance over the No Filter baseline. The majority filter and expand–shrink filter each enhanced the OA and F1 by approximately 2%, with similar improvements in user and producer accuracy across all crops. Both filters demonstrated superior effectiveness in refining the CDL for DL applications. The boundary clean filter provided notable enhancements, particularly in OA and F1, showing an approximate 1% improvement over the No Filter results. Although the shrink filter also delivered improvements, it was slightly less effective than the majority and expand–shrink filters, with a modest increase in the OA and F1. The results for R-CDL from Lin et al.’s study [9] show that it achieves competitive performance, indicating its effectiveness in improving CDL quality. Conversely, the expand filter resulted in the lowest performance, highlighting the importance of choosing appropriate filters to avoid amplifying noise. The application of these filters, especially the majority and expand–shrink filters, proved crucial in enhancing the quality of the CDL, thereby improving the DL model’s accuracy in segmenting agricultural fields.
The visual comparison of the segmentation results using different image filters on the CDL, as shown in Figure 9, reveals several insights. The No Filter column depicts the raw CDL, displaying significant noise and misclassification, especially along the boundaries of crop fields and roads. The aggregate filter, while effective at smoothing out some noise, still maintains a distinct separation between crop fields and adjacent features, suggesting good performance in boundary areas. Conversely, the shrink filter, although it reduces noise, tends to exaggerate the separation between crop fields and roads, sometimes over-segmenting these areas. The expand filter performs the poorest, as it amplifies noise and leads to over-segmentation, failing to accurately delineate crop fields from non-crop areas. The boundary clean filter, intended to enhance edge detection, shows weaknesses in accurately defining the boundaries of fields, often merging them with surrounding areas, which reduces its effectiveness in precise segmentation.
Overall, the majority and expand–shrink filters stand out. The majority filter shows the highest accuracy in maintaining the integrity of crop fields and effectively reducing noise, making it the best performer overall. The expand–shrink filter also performs well, particularly in maintaining clear boundaries between crop fields and roads, indicating a robust capability in handling complex boundary definitions. These filters significantly improve the segmentation results compared to the No Filter approach, with noticeable enhancements in the clarity and accuracy of the segmented crop fields. This visual analysis aligns with the numerical results, confirming that appropriate filtering techniques can substantially enhance the quality of the CDL for DL applications.
In Figure 10, the confusion matrices for different image filters applied to the CDL reveal distinct variations in the performance of DL models across major crops. The majority filter shows robust performance, particularly in identifying rice and soybeans, with higher UA and PA values compared to the other filters. This filter effectively reduces false positives and false negatives, making it highly reliable for segmenting various crop types. The expand–shrink filter also performs well, particularly in accurately identifying corn and cotton, though its performance slightly lags behind the majority filter.
Conversely, the expand filter exhibits the weakest performance, as evidenced by its higher rates of misclassification, particularly for corn and rice. The boundary clean filter shows moderate effectiveness, with relatively good accuracy for soybeans but less consistency for corn. The aggregate filter performs reasonably well but struggles with distinguishing between crops and non-crop areas, particularly in boundary regions. The majority and expand–shrink filters stand out as the most effective in improving DL segmentation accuracy for the major crops, highlighting their value in enhancing the quality of the CDL for agricultural applications.

3.3. CDL Refinement in T14TNK Area

To ensure the robustness and generalizability of our proposed methodologies in a different geographical and agricultural context, we conducted an additional experiment in the T14TNK test area. The performance metrics of the DL models, trained with the CDL at different confidence intervals, are summarized in Table 3. Notably, the results indicate that the 45% and 55% confidence intervals yield the highest accuracy, suggesting that these intervals effectively balance data quality and noise reduction. This result aligns with our findings in the Mississippi Delta study area, confirming the effectiveness of these confidence intervals across different regions. Conversely, higher confidence intervals, such as 95% and 100%, exhibit the worst results, likely due to excessive exclusion of valid data, thus supporting the importance of selecting appropriate confidence levels to enhance model performance.
The performance metrics of DL models using different image filters on the CDL for the T14TNK area are summarized in Table 4. The results demonstrate that the majority and expand–shrink filters significantly enhance model accuracy, with the OA reaching 93.5% and 93.1%, respectively. The majority filter showed the highest effectiveness, achieving the best results across most metrics. The expand–shrink filter also performed well, indicating that it is a reliable method for improving CDL data quality. In contrast, the expand filter yielded the lowest performance metrics, highlighting the importance of selecting appropriate filtering techniques. The results of the R-CDL from Lin et al.’s study [9] also show competitive accuracy, supporting the robustness of our proposed approaches. These findings align with our previous results in the Mississippi Delta, confirming the generalizability and effectiveness of our methodology across different regions.

4. Discussion

4.1. CDL Refinement Using Confidence Layer

The superior performance observed with the +45% and +55% confidence intervals, as indicated by the highest F1 and MCDA relative closeness, suggests that this level effectively balances the inclusion of high-quality data while minimizing noise. This interval likely provides a robust training set that enhances the model’s ability to generalize accurately across different validation samples. In contrast, the decline in performance at the +85%–100% confidence intervals can be attributed to the excessive exclusion of data, which results in a training set that is too limited to capture the full variability in the croplands, thereby impairing the model’s performance.
The MCDA relative closeness plays a crucial role in explaining these results by providing a comprehensive evaluation that integrates multiple criteria, such as OA, UA, PA and F1. This holistic approach allows for a more nuanced assessment of the model’s performance across different confidence intervals. The high MCDA scores for the +45% and +55% intervals underscore their overall effectiveness, balancing various performance metrics more successfully than other intervals. The insights gained from MCDA highlight the importance of selecting an effective confidence interval that not only maximizes individual accuracy metrics but also achieves a harmonious balance across multiple evaluation criteria, leading to a more reliable and robust model for cropland mapping. This emphasizes the necessity for a strategic approach in choosing confidence intervals to enhance the quality and effectiveness of DL segmentation models in agricultural applications.

4.2. CDL Refinement Using Image Filters

The results of the DL segmentation using various image filters on the CDL underscore the importance of data pre-processing in improving model accuracy. The majority and expand–shrink filters emerged as the top performers due to their ability to effectively reduce noise and preserve the integrity of crop boundaries. The success of the majority filter can be attributed to its ability to smooth out irregularities and enhance the homogeneity of crop areas, thereby improving the model’s ability to accurately classify different crop types. This filter’s effectiveness in maintaining clear distinctions between crops and non-crop areas, particularly along field boundaries, is crucial for accurate segmentation. Similarly, the expand–shrink filter’s ability to manage the complexities of field boundaries and reduce noise without over-segmenting highlights its robustness in handling diverse agricultural landscapes.
However, the poorer performance of the expand filter highlights the risks of amplifying noise during the pre-processing stage. This filter’s tendency to over-segment and misclassify non-crop areas as crops underscores the need for careful selection of filtering techniques. The boundary clean filter’s moderate performance points to its partial success in edge detection, which, while beneficial, may not be sufficient alone for accurate segmentation. The results indicate that a nuanced approach to data filtering, which balances noise reduction and boundary preservation, is essential for optimizing DL model performance in crop segmentation tasks.

4.3. Comparative Analysis of CDL Refinement Methods

Among the various methods tested, DL segmentation trained with data processed by the majority filter yielded the best results. The findings indicate that approaches utilizing filtered data outperform those using different confidence intervals, with the second-best results achieved using the expand–shrink filter. Conversely, the approaches using data with +95% and 100% confidence intervals produced the worst results, despite their frequent use in other studies. This suggests that high confidence intervals may exclude too much data, leading to poor model performance. Similarly, the DL model trained with expanded data showed significant weaknesses in correct segmentation, with results comparable to the worst-case confidence interval scenarios.
The substantial difference in accuracy between the filter-based approaches and the confidence-interval-based approaches, often around 1%, highlights the importance of pre-processing the CDL with appropriate filters before training DL models. This finding suggests that the application of noise reduction filters on the CDL should be a preliminary step to improve DL performance. Furthermore, combining confidence intervals and filter approaches could potentially enhance CDL quality even further. However, this combined approach requires careful evaluation, as the simultaneous use of filters and confidence intervals might have unforeseen negative effects on DL results.
In comparing the results of our best-performing approach with the R-CDL from Lin et al.’s study [9], our findings demonstrate notable improvements. The R-CDL, which involves sophisticated steps to refine the CDL, was outperformed by the expand–shrink and majority filter approaches, which scored around 1% higher in various accuracy metrics. While our methods using the +45% and +55% confidence intervals resulted in slightly lower accuracies, they still demonstrated significant efficacy. This comparison, as illustrated in the accompanying chart in Figure 11, underscores that our simpler filter-based approaches can achieve superior or comparable accuracy metrics without the complexity of the R-CDL method, thus validating the robustness and effectiveness of our methodologies.

4.4. Generalizability Assessment

To assess the generalizability of our method across different study areas, we conducted a detailed comparative analysis using the F1-score. We used the F1-score as the primary accuracy metric for comparison because it balances precision and recall, providing a comprehensive measure of model performance, especially in the context of imbalanced classes. The study areas included the Mississippi Delta and the T14TNK region. We first normalized the F1-scores to ensure that the trends were the main focus rather than the absolute values. The Pearson correlation coefficient and paired t-tests were employed to quantify the similarity and statistical significance of the results between these areas. This approach allowed us to understand whether the trends observed in one region could be replicated in another, thus validating the robustness of our method.
The comparison yielded promising results. For all the approaches, the Pearson correlation coefficient was 0.82, indicating a strong positive correlation between the two study areas. The paired t-test resulted in a T-statistic of 3.14 × 10−14 and a p-value of 1, suggesting no significant difference between the means of results in the two regions. When focusing on the confidence-interval-based methods, the correlation coefficient was an impressive 0.99, with a T-statistic of 3.68 × 10−15 and a p-value of 1, reinforcing the high similarity and consistency of the results. For the filter-based methods, the correlation coefficient was 0.76, and the paired t-test yielded a T-statistic of 1.04 × 10−14 and a p-value of 1, indicating a robust but slightly less strong correlation compared to the confidence interval methods. This is likely due to the poor performance of the shrink filter in the T14TNK area, which could be attributed to the major crops in that region being corn and soybeans. The shrink filter may have had a detrimental effect on accurately segmenting these crops.
A p-value of almost 1 indicates no statistically significant difference between the results, suggesting that the methods’ performance in the Mississippi Delta and T14TNK regions is statistically indistinguishable. This reinforces the robustness and generalizability of the methods, showing consistent performance across different regions. Although the trend is similar, the results for T14TNK are generally 1–2% better, except for the shrink and expand filters. This improvement might be due to the DL model having fewer classes to differentiate, allowing it to perform better in T14TNK. Additionally, the major crops in T14TNK, corn and soybeans, may be easier to classify accurately in this region due to their distinct spectral signatures. The chart in Figure 12 visually represents these trends, confirming that our approach is highly generalizable across different geographical regions. To thoroughly analyze the generalizability of our approach, it is essential to test the methods in multiple study areas, which should be undertaken in future research.

4.5. Implementation Prospects and Limitations

By integrating the refined CDL with Sentinel-2 imagery, this approach provides a robust framework for creating more accurate and reliable cropland maps. These advancements could play a critical role in agricultural monitoring, enabling better resource management and decision-making processes. The potential to combine these methods with advanced pre-processing techniques further amplifies their utility, potentially leading to more precise segmentation models that can benefit a wide range of agricultural applications.
However, the study has several limitations that must be addressed in future research to fully validate and optimize these methods. A significant constraint is the reliance on the CDL as the sole ground truth, which limits the ability to independently verify the results. Access to additional ground-truth data from different sources would enable a more comprehensive validation of the refined CDL and the effectiveness of the applied filters. Additionally, testing the methods in multiple study areas is essential to ensure that our findings are robust and applicable across different geographical regions. Furthermore, while the current filtering methods have shown efficacy, the development and integration of more advanced filtering techniques could further enhance data quality and segmentation accuracy. Future research should focus on exploring and validating these advanced methods, as well as establishing collaborations to obtain diverse ground-truth datasets, to overcome these limitations and improve the robustness of DL-based cropland segmentation models.

5. Conclusions

Numerous studies have leveraged the CDL and satellite imagery for the automation of cropland classification and segmentation using machine learning models. Despite its utility, the CDL is marred by errors such as salt-and-pepper noise, which adversely affect the performance of DL models in cropland segmentation. Consequently, refining the CDL to mitigate these inaccuracies is crucial. In this research, we investigated two primary approaches for refining the CDL: applying varying confidence intervals to the CDL and employing different image filters. For the time series data, five monthly minimum composites of Sentinel-2 imagery from the Mississippi Delta were used as satellite data for training and testing. To further validate our methods, we tested the approach in a new study area (T14TNK).
The analysis indicates that the CDL with confidence intervals ranging from +45% to +55% yielded the best results in terms of OA, enhancing the accuracy by approximately 1% compared to using the unfiltered CDL. The best results were observed with the CDL processed using the majority and expand–shrink filters, which improved the evaluation metrics by about 1.5%. The application of MCDA further corroborated these findings by providing a comprehensive evaluation across multiple criteria, underscoring the overall efficacy of the majority filter.
The results from T14TNK confirmed the generalizability of our method, showing similar trends in accuracy improvements. The performance metrics in T14TNK were generally around 1%–2% higher compared to the Mississippi Delta, likely due to fewer crop classes in T14TNK, which may have allowed the DL model to perform better. The Pearson correlation coefficient and paired t-tests were employed to quantify the similarity and statistical significance of the results between these areas. The Pearson correlation coefficient for confidence interval methods was 0.99, indicating a very strong correlation, while the paired t-tests yielded a T-statistic close to 0 and a p-value of 1 after normalization, suggesting no significant difference between the performance in the two regions. This reinforces the robustness and generalizability of our methods across different geographical regions.
Based on the findings of this research, it is recommended to filter the CDL with image filters such as majority or expand–shrink before employing it for DL segmentation training to identify crops. Additionally, selecting data with confidence intervals of +45% prior to applying these filters can further enhance the quality and reliability of the training data. Future research should explore the synergistic effects of combining filtering and confidence interval techniques and investigate the potential of integrating advanced CDL pre-processing methods to augment the performance of DL models in agricultural applications. This holistic approach could lead to more precise and dependable segmentation models, thereby benefiting various agricultural mapping and monitoring tasks.

Author Contributions

R.M.: conceptualization, methodology, formal analysis, validation, investigation, data curation, writing—original draft; F.W.: Conceptualization, methodology, validation, resources, writing—review and editing, supervision, projection administration, funding acquisition; A.O.: validation, investigation, writing—review and editing; L.F.: software, validation, writing—review and editing; G.Y.: writing—review and editing, supervision, project administration. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The Sentinel-2 imagery can be accessed freely using the ESA Copernicus Open Access Hub, and the CDL is available from the USDA NASS website. The R-CDL dataset can be downloaded from Zenodo repository (https://zenodo.org/doi/10.5281/zenodo.5565596, accessed on 17 July 2024). While the Sentinel-2 and CDL are publicly accessible, the derived data supporting the findings of this study are available only upon request due to data sharing restrictions imposed by the license. All of these datasets were accessed and processed using GEE.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Čerkasova, N.; White, M.; Arnold, J.; Bieger, K.; Allen, P.; Gao, J.; Meki, M.; Kiniry, J.; Gassman, P. Field scale SWAT+ modeling of corn and soybean yields for the contiguous United States: National Agroecosystem Model Development. Agric. Syst. 2023, 210, 103695. [Google Scholar] [CrossRef]
  2. FAO. The State of Food Insecurity in the World—Meeting the 2015 International Hunger Targets: Taking Stock of Uneven Progress; FAO: Rome, Italy, 2015; pp. 1–54. [Google Scholar]
  3. FAO; IFAD; UNICEF; WFP; WHO. The State of Food Security and Nutrition in the World: Transforming Food Systems for Affordable Healthy Diets; FAO: Rome, Italy; IFAD: Rome, Italy; UNICEF: Rome, Italy; WFP: Rome, Italy; WHO: Rome, Italy, 2020. [Google Scholar]
  4. Zhu, Y.; Wu, S.; Qin, M.; Fu, Z.; Gao, Y.; Wang, Y.; Du, Z. A deep learning crop model for adaptive yield estimation in large areas. Int. J. Appl. Earth Obs. Geoinf. 2022, 110, 102828. [Google Scholar] [CrossRef]
  5. USDA; N.A.S.S. Statewide Agricultural Accuracy Report. Available online: https://www.nass.usda.gov/Research_and_Science/Cropland/metadata/meta.php (accessed on 17 July 2024).
  6. Ebrahimy, H.; Zhang, Z. Per-pixel accuracy as a weighting criterion for combining ensemble of extreme learning machine classifiers for satellite image classification. Int. J. Appl. Earth Obs. Geoinf. 2023, 122, 103390. [Google Scholar] [CrossRef]
  7. Zhang, C.; Di, L.; Lin, L.; Li, H.; Guo, L.; Yang, Z.; Yu, E.G.; Di, Y.; Yang, A. Towards automation of in-season crop type mapping using spatiotemporal crop information and remote sensing data. Agric. Syst. 2022, 201, 103462. [Google Scholar] [CrossRef]
  8. Fritz, S.; See, L.; Bayas, J.C.L.; Waldner, F.; Jacques, D.; Becker-Reshef, I.; Whitcraft, A.; Baruth, B.; Bonifacio, R.; Crutchfield, J.; et al. A comparison of global agricultural monitoring systems and current gaps. Agric. Syst. 2019, 168, 258–272. [Google Scholar] [CrossRef]
  9. Lin, L.; Di, L.; Zhang, C.; Guo, L.; Di, Y.; Li, H.; Yang, A. Validation and refinement of cropland data layer using a spatial-temporal decision tree algorithm. Sci. Data 2022, 9, 63. [Google Scholar] [CrossRef] [PubMed]
  10. Cai, Y.; Guan, K.; Peng, J.; Wang, S.; Seifert, C.; Wardlow, B.; Li, Z. A high-performance and in-season classification system of field-level crop types using time-series Landsat data and a machine learning approach. Remote Sens. Environ. 2018, 210, 35–47. [Google Scholar] [CrossRef]
  11. Wang, S.; Azzari, G.; Lobell, D.B. Crop type mapping without field-level labels: Random forest transfer and unsupervised clustering techniques. Remote Sens. Environ. 2019, 222, 303–317. [Google Scholar] [CrossRef]
  12. Zhong, L.; Hu, L.; Zhou, H. Deep learning based multi-temporal crop classification. Remote Sens. Environ. 2019, 221, 430–443. [Google Scholar] [CrossRef]
  13. Hao, P.; Di, L.; Zhang, C.; Guo, L. Transfer Learning for Crop classification with Cropland Data Layer data (CDL) as training samples. Sci. Total Environ. 2020, 733, 138869. [Google Scholar] [CrossRef]
  14. Lark, T.J.; Schelly, I.H.; Gibbs, H.K. Accuracy, Bias, and Improvements in Mapping Crops and Cropland across the United States Using the USDA Cropland Data Layer. Remote Sens. 2021, 13, 968. [Google Scholar] [CrossRef]
  15. Reitsma, K.D.; Clay, D.E.; Clay, S.A.; Dunn, B.H.; Reese, C. Does the US cropland data layer provide an accurate benchmark for land-use change estimates? Agron. J. 2016, 108, 266–272. [Google Scholar] [CrossRef]
  16. Li, R.; Wei, C.; Afroz, M.D.; Lyu, J.; Chen, G. A GIS-based framework for local agricultural decision-making and regional crop yield simulation. Agric. Syst. 2021, 193, 103213. [Google Scholar] [CrossRef]
  17. Zhang, C.; Di, L.; Hao, P.; Yang, Z.; Lin, L.; Zhao, H.; Guo, L. Rapid in-season mapping of corn and soybeans using machine-learned trusted pixels from Cropland Data Layer. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102374. [Google Scholar] [CrossRef]
  18. Zhang, C.; Yang, Z.; Di, L.; Lin, L.; Hao, P. Refinement of cropland data layer using machine learning. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2020, 42, 161–164. [Google Scholar] [CrossRef]
  19. Wang, Y.; Feng, L.; Zhang, Z.; Tian, F. An unsupervised domain adaptation deep learning method for spatial and temporal transferable crop type mapping using Sentinel-2 imagery. ISPRS J. Photogramm. Remote Sens. 2023, 199, 102–117. [Google Scholar] [CrossRef]
  20. Ge, S.; Zhang, J.; Pan, Y.; Yang, Z.; Zhu, S. Transferable deep learning model based on the phenological matching principle for mapping crop extent. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102451. [Google Scholar] [CrossRef]
  21. Yang, G.; Yu, W.; Yao, X.; Zheng, H.; Cao, Q.; Zhu, Y.; Cao, W.; Cheng, T. AGTOC: A novel approach to winter wheat mapping by automatic generation of training samples and one-class classification on Google Earth Engine. Int. J. Appl. Earth Obs. Geoinf. 2021, 102, 102446. [Google Scholar] [CrossRef]
  22. ESA. MSI User Guide. Available online: https://sentinel.esa.int/web/sentinel/user-guides/sentinel-2-msi (accessed on 17 July 2024).
  23. Saad El Imanni, H.; El Harti, A.; Panimboza, J. Investigating Sentinel-1 and Sentinel-2 Data Efficiency in Studying the Temporal Behavior of Wheat Phenological Stages Using Google Earth Engine. Agriculture 2022, 12, 1605. [Google Scholar] [CrossRef]
  24. Tomíček, J.; Mišurec, J.; Lukeš, P.; Potůčková, M. Retrieval of Harmonized LAI Product of Agricultural Crops from Landsat OLI and Sentinel-2 MSI Time Series. Agriculture 2022, 12, 2080. [Google Scholar] [CrossRef]
  25. Wang, C.; Zhang, H.; Wu, X.; Yang, W.; Shen, Y.; Lu, B.; Wang, J. AUTS: A Novel Approach to Mapping Winter Wheat by Automatically Updating Training Samples Based on NDVI Time Series. Agriculture 2022, 12, 817. [Google Scholar] [CrossRef]
  26. Bartschat, A.; Reischl, M.; Mikut, R. Data mining tools. WIREs Data Min. Knowl. Discov. 2019, 9, e1309. [Google Scholar] [CrossRef]
  27. Boryan, C.; Yang, Z.; Mueller, R.; Craig, M. Monitoring US agriculture: The US Department of Agriculture, National Agricultural Statistics Service, Cropland Data Layer Program. Geocarto Int. 2011, 26, 341–358. [Google Scholar] [CrossRef]
  28. Rahman, M.S.; Di, L.; Yu, E.; Zhang, C.; Mohiuddin, H. In-Season Major Crop-Type Identification for US Cropland from Landsat Images Using Crop-Rotation Pattern and Progressive Data Classification. Agriculture 2019, 9, 17. [Google Scholar] [CrossRef]
  29. Dong, W.; Woźniak, M.; Wu, J.; Li, W.; Bai, Z. Denoising aggregation of graph neural networks by using principal component analysis. IEEE Trans. Ind. Inform. 2022, 19, 2385–2394. [Google Scholar] [CrossRef]
  30. Ilesanmi, A.E.; Ilesanmi, T.O. Methods for image denoising using convolutional neural network: A review. Complex Intell. Syst. 2021, 7, 2179–2198. [Google Scholar] [CrossRef]
  31. Wade, T.; Wickham, J.; Nash, M.; Neale, A.; Riitters, K.; Jones, K. A Comparison of Vector and Raster GIS Methods for Calculating Landscape Metrics Used in Environmental Assessments. Photogramm. Eng. Remote Sens. 2003, 69, 1399–1405. [Google Scholar] [CrossRef]
  32. Olivas, E.S.; Guerrero, J.D.M.; Sober, M.M.; Benedito, J.R.M.; Lopez, A.J.S. Handbook of Research on Machine Learning Applications and Trends: Algorithms, Methods and Techniques—2 Volumes; IGI Publishing: Hershey, PA, USA, 2009. [Google Scholar]
  33. Soares, D.; Galvão, L.; Formaggio, A. Crop area estimate from original and simulated spatial resolution data and landscape metrics. Sci. Agric. 2008, 65, 459–467. [Google Scholar] [CrossRef]
  34. European Space Agency. Copernicus Sentinel-2 Collection 1 MSI Level-2A (L2A). Available online: https://sentinels.copernicus.eu/web/sentinel/sentinel-data-access/sentinel-products/sentinel-2-data-products/collection-1-level-2a (accessed on 17 July 2024).
  35. Liu, M.; Fu, B.; Xie, S.; He, H.; Lan, F.; Li, Y.; Lou, P.; Fan, D. Comparison of multi-source satellite images for classifying marsh vegetation using DeepLabV3 Plus deep learning algorithm. Ecol. Indic. 2021, 125, 107562. [Google Scholar] [CrossRef]
  36. Yojana, K.; Thillai Rani, L. OCT layer segmentation using U-NET semantic segmentation and RESNET34 encoder-decoder. Meas. Sens. 2023, 29, 100817. [Google Scholar] [CrossRef]
  37. Zhang, D.; Pan, Y.; Zhang, J.; Hu, T.; Zhao, J.; Li, N.; Chen, Q. A generalized approach based on convolutional neural networks for large area cropland mapping at very high resolution. Remote Sens. Environ. 2020, 247, 111912. [Google Scholar] [CrossRef]
  38. Congalton, R.G.; Green, K. Assessing the Accuracy of Remotely Sensed Data: Principles and Practices; CRC Press: Boca Raton, FL, USA, 2019. [Google Scholar]
  39. Larsen, A.E.; Hendrickson, B.T.; Dedeic, N.; MacDonald, A.J. Taken as a given: Evaluating the accuracy of remotely sensed crop data in the USA. Agric. Syst. 2015, 141, 121–125. [Google Scholar] [CrossRef]
  40. Abdali, E.; Valadan Zoej, M.J.; Taheri Dehkordi, A.; Ghaderpour, E. A Parallel-Cascaded Ensemble of Machine Learning Models for Crop Type Classification in Google Earth Engine Using Multi-Temporal Sentinel-1/2 and Landsat-8/9 Remote Sensing Data. Remote Sens. 2024, 16, 127. [Google Scholar] [CrossRef]
  41. Bahrami, H.; Homayouni, S.; Safari, A.; Mirzaei, S.; Mahdianpari, M.; Reisi-Gahrouei, O. Deep Learning-Based Estimation of Crop Biophysical Parameters Using Multi-Source and Multi-Temporal Remote Sensing Observations. Agronomy 2021, 11, 1363. [Google Scholar] [CrossRef]
Figure 1. Geographic overview and data layers used in the study. The top-left map shows the study area within the Mississippi Delta (blue rectangle) and T14TNK test area (red rectangle). The top-right image displays Sentinel-2 composite imagery from May 2021. The bottom-left map illustrates the distribution of crops in the CDL. The bottom-right map depicts the CDL confidence layer, indicating the confidence values associated with the CDL.
Figure 1. Geographic overview and data layers used in the study. The top-left map shows the study area within the Mississippi Delta (blue rectangle) and T14TNK test area (red rectangle). The top-right image displays Sentinel-2 composite imagery from May 2021. The bottom-left map illustrates the distribution of crops in the CDL. The bottom-right map depicts the CDL confidence layer, indicating the confidence values associated with the CDL.
Agriculture 14 01285 g001
Figure 2. NDVI profiles for the major crops in the study area throughout the year 2021.
Figure 2. NDVI profiles for the major crops in the study area throughout the year 2021.
Agriculture 14 01285 g002
Figure 3. The 2021 CDL land cover distribution in the study area.
Figure 3. The 2021 CDL land cover distribution in the study area.
Agriculture 14 01285 g003
Figure 4. Flowchart summarizing the research approach for major crop mapping from Sentinel-2 imagery using various CDL confidence levels and image filters.
Figure 4. Flowchart summarizing the research approach for major crop mapping from Sentinel-2 imagery using various CDL confidence levels and image filters.
Agriculture 14 01285 g004
Figure 5. Distribution of crop pixels refined by different confidence levels of major crops within the study area for the year 2021.
Figure 5. Distribution of crop pixels refined by different confidence levels of major crops within the study area for the year 2021.
Agriculture 14 01285 g005
Figure 6. Impact of varying confidence intervals on the refinement of the CDL. The percentages indicate the confidence thresholds used to refine the CDL.
Figure 6. Impact of varying confidence intervals on the refinement of the CDL. The percentages indicate the confidence thresholds used to refine the CDL.
Agriculture 14 01285 g006
Figure 7. DL segmentation results of croplands from Sentinel-2 imagery. The DL models were trained using different CDL confidence layer intervals.
Figure 7. DL segmentation results of croplands from Sentinel-2 imagery. The DL models were trained using different CDL confidence layer intervals.
Agriculture 14 01285 g007
Figure 8. Confusion matrices illustrating the performance of DL models trained with different CDL confidence intervals in segmenting major crops. Diagonal numbers represent the percentage of correctly classified instances for each crop, while non-diagonal numbers indicate the percentage of misclassified instances between different crops.
Figure 8. Confusion matrices illustrating the performance of DL models trained with different CDL confidence intervals in segmenting major crops. Diagonal numbers represent the percentage of correctly classified instances for each crop, while non-diagonal numbers indicate the percentage of misclassified instances between different crops.
Agriculture 14 01285 g008
Figure 9. Filtered CDL and corresponding DL segmentation results for major crops using various image filters. The “No Filter” and “R-CDL” results are included for comparison.
Figure 9. Filtered CDL and corresponding DL segmentation results for major crops using various image filters. The “No Filter” and “R-CDL” results are included for comparison.
Agriculture 14 01285 g009
Figure 10. Confusion matrices illustrating the performance of DL models using different image filters on the CDL for major crop segmentation. The values within the matrix represent the percentage of correctly and incorrectly classified instances, ranging from 0 to 100.
Figure 10. Confusion matrices illustrating the performance of DL models using different image filters on the CDL for major crop segmentation. The values within the matrix represent the percentage of correctly and incorrectly classified instances, ranging from 0 to 100.
Agriculture 14 01285 g010
Figure 11. Comparison of DL accuracy metric results using different CDL refinement methods, including the R-CDL from Lin et al.’s study [9].
Figure 11. Comparison of DL accuracy metric results using different CDL refinement methods, including the R-CDL from Lin et al.’s study [9].
Agriculture 14 01285 g011
Figure 12. F1-score trends comparing the performance of different refining methods between the Mississippi Delta and T14TNK areas. The chart illustrates the accuracy metrics for various confidence intervals and filtering techniques, highlighting the generalizability and robustness of the methods across different geographical regions.
Figure 12. F1-score trends comparing the performance of different refining methods between the Mississippi Delta and T14TNK areas. The chart illustrates the accuracy metrics for various confidence intervals and filtering techniques, highlighting the generalizability and robustness of the methods across different geographical regions.
Agriculture 14 01285 g012
Table 1. Performance metrics of the DL model using CDL datasets at different confidence interval percentages. The metrics include overall accuracy (OA), user’s accuracy (UA), producer’s accuracy (PA) and F1-score (F1) and MCDA relative closeness. “All CDL” refers to using the entire CDL without any confidence interval filtering, serving as a baseline for comparison.
Table 1. Performance metrics of the DL model using CDL datasets at different confidence interval percentages. The metrics include overall accuracy (OA), user’s accuracy (UA), producer’s accuracy (PA) and F1-score (F1) and MCDA relative closeness. “All CDL” refers to using the entire CDL without any confidence interval filtering, serving as a baseline for comparison.
Confidence IntervalOverallCornCottonRiceSoybeansOther
OAUAPAF1UAPAF1UAPAF1UAPAF1UAPAF1UAPAF1MCDA
All CDL89.489.589.789.688.788.188.488.891.990.393.391.492.385.686.085.891.391.191.285.8
+5%89.589.289.989.589.686.888.287.093.290.091.692.892.287.184.685.890.792.091.484.8
+15%89.790.389.790.089.787.788.790.191.991.093.092.392.689.183.085.989.593.691.586.4
+25%89.990.290.090.189.788.088.889.792.291.093.092.792.888.983.586.189.993.491.687.5
+35%90.090.589.890.290.887.789.290.291.891.094.091.592.786.585.986.291.192.391.788.0
+45%90.590.790.790.790.189.789.990.891.691.293.093.893.488.385.686.991.492.892.192.3
+55%90.690.091.490.790.789.390.089.093.291.091.095.793.386.887.787.292.691.291.990.8
+65%90.689.891.290.488.691.089.888.692.690.690.695.292.889.484.687.091.592.391.990.1
+75%90.389.090.689.886.592.489.487.191.889.392.291.591.987.285.986.592.191.491.786.2
+85%88.686.287.786.984.588.286.384.185.184.587.492.789.984.481.883.190.990.790.868.1
+95%84.777.079.678.173.681.677.471.575.173.277.187.381.973.564.568.789.189.489.28.9
+100%84.976.379.577.872.781.877.070.475.172.677.382.679.871.369.270.289.988.989.48.4
Table 2. Performance metrics of DL models using filtered CDL with various image filters. “R-CDL” refers to the CDL refined using the method described in Lin et al.’s study [9]. “No Filter” is equivalent to “All CDL” in Table 1.
Table 2. Performance metrics of DL models using filtered CDL with various image filters. “R-CDL” refers to the CDL refined using the method described in Lin et al.’s study [9]. “No Filter” is equivalent to “All CDL” in Table 1.
Overall CornCottonRiceSoybeansOther
FiltersOAUAPAF1UAPAF1UAPAF1UAPAF1UAPAF1UAPAF1MCDA
Aggregate90.189.589.089.287.388.888.188.788.988.891.891.191.489.283.586.390.692.891.752.9
Boundary Clean90.391.690.090.792.985.288.992.691.992.294.591.993.286.388.887.591.492.191.866.0
Expand86.787.985.486.588.478.583.189.485.087.290.487.188.785.681.883.785.494.789.814.7
Expand–Shrink91.791.591.791.691.790.391.090.193.691.893.293.993.590.787.389.092.093.592.884.0
Majority91.491.891.891.890.292.091.192.593.292.894.094.294.190.986.088.491.493.892.683.8
Shrink91.489.489.489.389.585.387.486.889.087.990.093.091.487.485.986.693.293.893.551.3
R-CDL [9]90.790.891.391.090.090.390.290.094.592.293.592.593.088.886.187.491.492.892.175.9
No Filter89.489.589.789.688.788.188.488.891.990.393.391.492.385.686.085.891.391.191.258.1
Table 3. Performance metrics of DL models trained with the CDL at different confidence intervals for the T14TNK area.
Table 3. Performance metrics of DL models trained with the CDL at different confidence intervals for the T14TNK area.
Confidence IntervalOverallCornSoybeansOther
OAUAPAF1UAPAF1UAPAF1UAPAF1MCDA
All CDL92.791.992.292.192.490.391.488.791.890.294.594.694.692.0
+5%92.792.092.192.091.690.991.290.190.690.394.594.794.692.2
+15%92.792.092.292.193.289.791.488.891.890.394.195.194.691.2
+25%92.792.192.192.192.890.091.489.491.390.394.195.094.592.3
+35%92.892.092.392.291.991.191.589.391.590.494.894.494.693.6
+45%93.192.392.692.592.291.591.889.991.790.894.994.694.896.1
+55%92.992.192.592.391.892.192.089.991.290.694.794.094.394.2
+65%92.691.891.991.991.991.991.989.890.390.093.893.693.791.0
+75%91.890.491.691.091.091.491.286.491.288.793.992.293.182.1
+85%90.488.889.189.088.289.989.185.785.785.792.591.892.263.6
+95%87.382.983.483.180.781.781.277.077.877.491.190.690.86.0
+100%87.582.784.183.478.884.881.777.477.577.591.990.090.99.4
Table 4. Performance metrics of DL models using various image filters on the CDL for the T14TNK area. The R-CDL [9] results are included for comparison, demonstrating the effectiveness of different filtering techniques in enhancing CDL quality.
Table 4. Performance metrics of DL models using various image filters on the CDL for the T14TNK area. The R-CDL [9] results are included for comparison, demonstrating the effectiveness of different filtering techniques in enhancing CDL quality.
Overall CornSoybeansOther
FiltersOAUAPAF1UAPAF1UAPAF1UAPAF1MCDA
Aggregate92.290.791.591.190.591.691.187.290.088.694.293.093.670.3
Boundary Clean92.592.491.591.992.296.294.194.088.591.291.189.790.474.8
Expand85.686.085.084.680.898.688.897.573.183.579.683.481.545.6
Expand–Shrink93.191.893.392.595.792.794.291.393.892.588.293.690.877.7
Majority93.592.793.293.095.494.294.891.893.292.591.192.291.680.7
Shrink85.679.891.283.798.779.588.172.997.283.367.896.979.841.8
R-CDL [9]92.792.292.192.192.190.991.590.590.990.793.994.494.175.5
No Filter92.791.992.292.192.490.391.488.791.890.294.594.694.674.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Maleki, R.; Wu, F.; Oubara, A.; Fathollahi, L.; Yang, G. Refinement of Cropland Data Layer with Effective Confidence Layer Interval and Image Filtering. Agriculture 2024, 14, 1285. https://doi.org/10.3390/agriculture14081285

AMA Style

Maleki R, Wu F, Oubara A, Fathollahi L, Yang G. Refinement of Cropland Data Layer with Effective Confidence Layer Interval and Image Filtering. Agriculture. 2024; 14(8):1285. https://doi.org/10.3390/agriculture14081285

Chicago/Turabian Style

Maleki, Reza, Falin Wu, Amel Oubara, Loghman Fathollahi, and Gongliu Yang. 2024. "Refinement of Cropland Data Layer with Effective Confidence Layer Interval and Image Filtering" Agriculture 14, no. 8: 1285. https://doi.org/10.3390/agriculture14081285

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop