Next Article in Journal
TTSNet: Transformer–Temporal Convolutional Network–Self-Attention with Feature Fusion for Prediction of Remaining Useful Life of Aircraft Engines
Previous Article in Journal
In Shift and In Variance: Assessing the Robustness of HAR Deep Learning Models Against Variability
Previous Article in Special Issue
Evaluating UAV-Based Remote Sensing for Hay Yield Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Combining a Standardized Growth Class Assessment, UAV Sensor Data, GIS Processing, and Machine Learning Classification to Derive a Correlation with the Vigour and Canopy Volume of Grapevines

1
Department of Viticulture and Oenology, DLR (Dienstleistungszentrum Ländlicher Raum) Mosel/Steillagenzentrum, Gartenstraße 18, 54470 Bernkastel-Kues, Germany
2
Department of Computer Science, University of Koblenz, Universitätsstraße 1, 56070 Koblenz, Germany
3
Department of Environmental Remote Sensing and Geoinformatics, Trier University, Universitätsring 15, 54296 Trier, Germany
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(2), 431; https://doi.org/10.3390/s25020431
Submission received: 7 November 2024 / Revised: 29 December 2024 / Accepted: 31 December 2024 / Published: 13 January 2025
(This article belongs to the Special Issue Remote Sensing for Crop Growth Monitoring)

Abstract

:
Assessing vines’ vigour is essential for vineyard management and automatization of viticulture machines, including shaking adjustments of berry harvesters during grape harvest or leaf pruning applications. To address these problems, based on a standardized growth class assessment, labeled ground truth data of precisely located grapevines were predicted with specifically selected Machine Learning (ML) classifiers (Random Forest Classifier (RFC), Support Vector Machines (SVM)), utilizing multispectral UAV (Unmanned Aerial Vehicle) sensor data. The input features for ML model training comprise spectral, structural, and texture feature types generated from multispectral orthomosaics (spectral features), Digital Terrain and Surface Models (DTM/DSM- structural features), and Gray-Level Co-occurrence Matrix (GLCM) calculations (texture features). The specific features were selected based on extensive literature research, including especially the fields of precision agri- and viticulture. To integrate only vine canopy-exclusive features into ML classifications, different feature types were extracted and spatially aggregated (zonal statistics), based on a combined pixel- and object-based image-segmentation-technique-created vine row mask around each single grapevine position. The extracted canopy features were progressively grouped into seven input feature groups for model training. Model overall performance metrics were optimized with grid search-based hyperparameter tuning and repeated-k-fold-cross-validation. Finally, ML-based growth class prediction results were extensively discussed and evaluated for overall (accuracy, f1-weighted) and growth class specific- classification metrics (accuracy, user- and producer accuracy).

Graphical Abstract

1. Introduction

Detecting relevant vineyard objects, including vine rows, grapevine stems, and quantifying vine health properties, like vigour, are essential for effective vineyard management [1,2,3]. Therefore, extracting vines and correlating important viticultural parameters, like grape yield and vigour, with sensor-derived image features is essential for high-precision vineyard management and automatization approaches. For this purpose, various sensor technologies and methods from sensor-based precision viticulture are at the forefront [4]. These technologies include different sensor-system data processed with new digital techniques, including classical image processing to more recent machine learning and deep learning techniques [5,6,7]. According to the vineyard-specific geometric structure and size, high-resolution UAV (Unmanned Aerial Vehicle)-derived images and proximal sensors are favored compared to aerial and satellite sensor data in viticulture applications [8,9]. Based on these sensor systems, many approaches have been developed over recent decades for extracting vineyard objects, especially vine rows and single vines. For vine rows, these approaches include manual GIS (Geographic Information System)-based extraction methods in combination with DTM/DSM (Digital Terrain Model/Digital Surface Model) data [10], threshold-based GIS- methods [11,12], object-based image segmentation techniques [13,14,15,16], HSV (Hue Saturation Value)-based methods [17], texture-based techniques [18], geometric structure extraction from stereo image techniques [19], ultrasonic sensor detection techniques [20], LiDAR (Light Detection And Ranging)- point cloud processing techniques [21,22,23] deep learning methods [24], and combinations of various methods [15,17,25]. Moreover, several vine extraction approaches were developed based on data from different sensor types and processing routines [26,27,28,29,30,31,32].
The extracted objects comprised different crops [33,34], including vines [35], subsequently joined and correlated to the extracted image features from the sensor data. For example, some authors derived canopy parameters like the LAI (Leaf Area Index) [36], health status of vines [37], or the canopy volume [27] from different sensor types and data sources. In the case of 2D multispectral image data, these features comprise essentially spectral features like Vegetation Indices (VI) [38], structural features like CHM (Canopy Height Model), or texture features [18]. The extracted features are correlated to the ground truth data [39,40,41,42,43], which were measured and labeled in the field to generate the training and evaluation datasets as basis for model training and evaluation. The selection of the prediction technique depends on the specific data structure and type of prediction variable. Therefore, several methods are used, from classical regression methods [43] to more sophisticated modern machine learning and deep learning-based approaches, depending on those constraints [44,45].
Under consideration of recent research in sensor-based precision agri- and viticulture, the research objective of this study is to evaluate the potential of combining canopy spectral, structural, and texture vine canopy features for growth estimation of vines based on machine learning model predictions of specific vine growth classes, applying the growth class assessment of Porten [43] as ground truth. Hitherto, no previous study used this specific assessment and UAV- image data geoprocessing methodology for machine learning classifier-based predictions of vine vigour classes.
To reach this objective, the following research questions and objectives of this study were formulated:
  • Is there a significant difference between spectral, structural, and texture input features and their combinations for model prediction metrics and the robustness of the results?
  • Which classifiers after hyperparameter tuning provide the best overall prediction performance?
  • Which growth classes have the highest positive prediction matches, and which have the lowest matches? Is there a pattern in class-specific metrics that can be addressed in the model prediction process and the selected input feature groups?
  • Based on the findings from the previous points, what are the advantages and limitations of the growth class assessment after Porten [43], and how could the evaluation, database, methodology, and classification process be optimized to improve the prediction and separability of the growth classes in the future?
To answer these research questions, the generation and extraction of different canopy feature types (spectral, structural, texture) of single vines based on UAV-multispectral sensor data, Digital Terrain and Surface Models (DTM/DSM), and derived geodata (e.g., CHM- Canopy Height Model) were investigated as prediction variables for growth classifications. To achieve this goal, the most suited variables, such as Vegetation Indices (VI) (spectral features), height measures (e.g., Canopy Height Model-structural features), and texture features (GLCM- Gray-Level Co-occurrence Matrix), were selected, computed and considered as model training input to predict the growth classes after [43], which were used as target classes. To fit the training data, the most promising machine learning classifiers were selected based on extensive literature research in the fields of precision agriculture, viticulture, and environmental sciences using sensor-based remote and proximal sensor techniques for crop, yield and biomass estimation. Moreover, selected ML classifier hyperparameters were selected and tuned with a grid-search-based repeated-k-fold-cross-validation approach. Next, the quality of class separation due to the input feature groups, classifier selection, and robustness of the prediction results was examined, based on chosen model metrics and the current literature on biomass, crop, and vine volume and vigour estimation. Thereafter, the class-specific accuracy (confusion matrix) was computed and evaluated for a more differentiated evaluation of class-specific influences, related to the input feature types and input feature groups. Finally, the growth classification assessment after Porten [43] and the methodology of this study will be briefly evaluated in terms of the potential for machine learning-based growth predictions of grapevines, including future perspectives, for enhancing the classification scores and improving the growth class separation procedure.

2. Materials and Methods

The following subsections outline the study’s investigation area, materials, and methods as the foundation for subsequent processing and classification routines.

2.1. Test Site and Field Data Acquisition (Ground Truth Data)

The investigation area is located in the so-called arena, a test vine field of the DLR (Dienstleistungszentrum Ländlicher Raum, a service center for rural area based in Bernkastel-Kues) Mosel, in the northwest of Bernkastel-Kues in Rhineland, Palatinate, (Germany) (see Figure 1) (49.91334° N, 7.04960° E (EPSG: 4326- WGS 84)). The approximately 2 ha sized area consists of regularly sloping vine rows in fall line (slope orientation) with some individual vine row tips.
The area has an average slope of approx. 30° and a predominantly SSW orientation at an average altitude between 183–208 m above sea level. The vine variety under investigation is White Riesling (Vitis vinifera subsp. vinifera), a typical grapevine variety of the Moselle region. The form of training is espalier training with two pendulum arches (see Figure 2).
The average distance between vine rows is approximately 2.0 m, and the average distance between adjacent vines in the same row is approximately 1.20 m (standing space 2.4 m2/vine). During ground truth labeling (see Section 2.4) and UAV-dataset generation (see Section 2.5.1 and Section 2.5.2) at véraison stage (03.08.2022), the mean temperature was recorded with 30.5 °C, with a relative humidity of 34% and a wind speed of around 1 m/s at the local weather station (see source link in Supplementary Materials). Typical grapevine infections within the Moselle area are the vine leaf disease Bois Noir [46], the grapevine trunk disease Esca [47], and virus infections [48], which can affect vital parameters like vigour.

2.2. Vine Growth and Growth Classification Assessment After Porten (2020)

The vigour of grapevines is influenced by several key factors, such as soil composition, soil chemistry, soil management, vine variety, fertilization type and amount, stand space, rootstem, and gate level [49]. The concept of terroir and its relation to vine quality is more explicitly discussed by [50,51,52]. Therefore, vine vigour as an explicit expression of vine health depends on many interacting factors, including the vegetative and reproductive phenological growth stages. Moreover, vigour is a good indicator of vegetative and reproductive parameters like canopy volume, height, leaf area, grape yield, and related vineyard management and cultivation strategies [29,53,54]. Therefore, due to the potentially high correlation of vigour with physiological parameters, the growth class assessment after [43] was selected as basis for the ground truth target classes for later correlation and ML classification with the processed, extracted, and aggregated UAV-sensor image data. The specific growth classification assessment after Porten [43] has not been used before, for correlation and classification with UAV-sensor derived image data. Previously, less differentiated assessments were correlated with UAV-sensor data [55,56,57,58,59,60] on whole vine sites, but without considering the growth stages on the spatial resolution of each single grapevine stem. Nevertheless, for the effective correlation with UAV-sensor image data, a differentiated vine growth assessment is necessary to describe the internal structure of each vine described by [61]. Under consideration of these previous studies and requirements, the growth class assessment after [43] was developed, extending previously used assessments for correlation with UAV-sensor data. For validation, [43] tested the assessment on several vine fields by correlating the labeled vines with relevant viticulture vegetative and reproductive parameters including dry wood weight, one-year wood, leaf chlorophyll content, or the number of bunches measured on the same vines. The correlation and statistical significance of the growth class labels and the relevant parameters were investigated using several statistical methods, including factor analysis and statistical significance tests like one-way ANOVA (Analysis of Variance) and the Mann–Whitney-U-test. For comparison, similar correlations and spatial interpolations of viticulture parameters with different remote and proximal sensor types for investigating vineyard variability were achieved by [62]. Therefore, the growth class assessment after [43] is proven to be scientifically valid and can be used as a template for ground truth data labeling regarding remote- and proximal-sensor-based growth predictions. The specific structure and coding of the assessment are described in the following section.

2.3. Growth and Infection Classification Assessment After Porten (2020)

The “growth” in growth classes can be seen as a synonym for gradual vine vigour categories, from no growth (0) to excellent growth (9), with the intermediate growth classes (1,3,5,7) in between. For each growth class a color code was used, for visualizing the grapevine stem positions together with the growth classes categories, determined in the field (see first column in Table 1).
The gap between growth classes (3 to 5) does not necessarily express a linear relationship to vine vigour; rather, it provides space for intermediate classes if modifications are required to adapt or extend the assessment. For the visual standardization of the growth classes, template pictures of each growth class from a side view were taken (see Figure 3). This is especially important to reduce labeling error sources, like intra- and interpersonal errors, as much as possible. Within the classification process, only growth classes without infection type and infection strength were considered. Nevertheless, the infection has a strong impact on vine vigour [43] but was not the main focus of attention in this investigation. Accordingly, the specific characteristics of the individual infections are not described in detail here because this was not the focus of this study.

2.4. Manual Ground Truth Labeling of Vines

Based on the standardized classification scheme after [43], the vines were labeled during the véraison stage (3 Auguest 2022) and merged with the vines position described in the method section (see Section 2). The total number of vines labeled amounts to 1657, with a growth class count distribution of growth class 0 = 206, growth class 1 = 190, growth class 3 = 343, growth class 5 = 470, growth class 7 = 402, growth class 9 = 46.
For this study, only growth class labels without the infection additive and infection strength were considered as target classes for machine learning predictions due to the overall weak correlation and explanatory power of infection to vine growth concerning all vines. The véraison was chosen as reference time for labeling due to the most significant differentiation of viticultural parameters concerning the individual growth classes [43], especially during the onset of bunch’ ripening. Labeling of the vines was done simultaneously to UAV-sensor data acquisitions and terrestrial field data measurements (soil water content, early leaf water potential) for meaningful correlation. The growth class ground truth was used as target variable for the training-test data split previous to fitting the preprocessed sensor image data to the ML classifiers. In terms of UAV-sensor-based ML classification, ground truth data included vine yield measurements [8], leaf water potential [60], stem water potential [41], and stomatal conductance [42], not considering growth as the target variable, as in this study. Within this study, the early morning leaf water potential was measured before sunset at 2–3 am. Besides bivariate correlation the soil water content and leaf water measurement data were not further considered for ML classification of the growth classes due to only a few sampling sites and inappropriate measurement effort.

2.5. UAV Data Acquisition, Preparation, and Processing

2.5.1. Localization and Spatial Reference System of the Geodata

Precise georeferencing is essential in all sensor-based precision agriculture studies [8]. All geodata (vector data: vines, raster data: DTM/DSM—Digital Terrain Model/Digital Surface Model, multi-channel raster orthomosaics) were referenced to Coordinate Reference System (CRS) EPSG (European Petroleum Survey Group) code 25,832 ETRS (European Terrestrial Reference System) 89/UTM (Universal Transverse Mercator) zone 32N. As further geolocation equipment, a ground station (DJI D-RTK Real Time Kinematic) and 2 Mobile Station GNSS (Global Navigation Satellite System) reference stations were used for absolute geolocalization of the drone and camera position during the flight. Additionally, several Ground Control Points (GCPs) were installed and precisely located with a portable Emlid Reach RS2- GPS (Global Positioning System) device to improve georeferencing, geometric rectification and reduce image distortions during postprocessing. The SAPOS (Satellite Positioning Service of German Land Surveying) guaranteed geometric corrections necessary for high-precision absolute location measurements. The open-source GIS software QGIS 3.22 (Quantum GIS, QGIS Development Team, Open Source Geospatial Foundation) was used for geodata visualization and all cartographic layouts.

2.5.2. UAV Sensor Data Acquisition

Multispectral camera image data were recorded on five distinct spectral bandwidths (Red, Green, Blue, Red-Edge, Near Infrared) using the UAV Black Snapper XL Pro (quadrocopter) equipped with a MicaSense Red-Edge-MXTM (Micasense Inc., Seattle, WA, USA) multispectral camera. An overview of the spectral bandwidth of each spectral band is summarized in Table 2. A calibration panel (Micasense Inc., Seattle, WA, USA) with known reflectance values was used for radiometric calibration. In agricultural application scenarios, multispectral sensors were often used for crop yield predictions [63,64], but also in more specific use-cases, like stress detection of vines [6,65,66]. Moreover, a LiDAR scanner integrated into a DJI Zenmuse L1 (RGB camera incl. gimbal) was used during UAV-sensor data acquisitions, installed on a DJI Matrice 300 RTK UAV, to generate point clouds, necessary for the generation of the DTM and DSM.
The flight path for UAV- sensor data acquisition was created in block shapes within the flight planning software DJI flight Planner. The recording altitude above ground was set at 40 m, and the overlap for the recorded sensor images was set at 80%. The flight was accomplished midday with preferably cloud-free sky and perpendicular sunlight conditions to keep radiometric interferences as low as possible. After each flight, the recorded sensor image data were checked and saved on a prepared external hard drive.

2.5.3. Photogrammetry

The photogrammetry software Agisoft Metashape Professional (version 1.7.0) was utilized in several processing steps to create the DSM, DTM, and multispectral orthomosaics [13] with a SfM (Structure from Motion)-based approach [9,67,68]) on the basis of the recorded images of the UAV- sensor data acquisitions (see previous Section). For this procedure, the imported images were adjusted, externally georeferenced, aerotriangulated, and readjusted. Georeferencing and image rectification were performed using GNSS coordinates of the five installed Ground Control Points (GCPs), matched in Agisoft Metashape professional version 2.1.4. Subsequently, the photogrammetric point cloud was derived to generate the multispectral orthomosaics, DSM and DTM. The orthomosaic, DTM, and DSM generation process involved three additional main steps: (1) initial processing for key point computation and image matching, (2) radiometric calibration for converting raw digital numbers into reflectance values, and (3) DSM generation for surface elevation representation and DTM generation through filtering of the ground classes. Finally, the orthomosaic was generated from the DSM and the calibrated spectral information. The georeferenced raster images were saved as TIFF (Tagged Image File Format) files (GeoTiff) and exported to QGIS version 3.22 for further geospatial processing.

2.5.4. Single Vine Geopositions

The grapevines’ geoposition were manually set, based on the high- precise RGB orthomosaics from canopy-free season generated from acquisitions during November 2021. To ensure a unique identifier for each grapevine stem, the vines were systematically numbered by each row from west to east and north to south in slope line, from top of the slope to foot of the slope, to ensure systematic geographical sorting and unique identification. Progressively, during ground truth labeling the open-source geographical digitalization App QField was used to label the vines according to the growth classification assessment after [43]. Subsequently, the attributed point-shapefiles were re-exported to QGIS version 3.22. A map of the growth class and health condition labeled grapevines after Porten [43] is shown in Figure 4. The localized and ground truth labeled vines are the basis for all subsequent geoprocessing and ML classification procedures (Section 2.6 and Section 2.7).

2.6. Geoprocessing Workflow

Figure 5 summarizes the methods and workflow for post UAV- data acquisition geoprocessing and iterations. The processing workflow was mainly developed in QGIS version 3.22 (PyQGIS) and SAGA-GIS version 7.8.2, with integrated geoprocessing tools like GDAL (Geospatial Data Abstraction Library). Different authors emphasized the value of GIS and geospatial tools for agriculture and viticulture [39,69,70,71]. The following Section 2.6.1,Section 2.6.2,Section 2.6.3,Section 2.6.4,Section 2.6.5,Section 2.6.6,Section 2.6.7,Section 2.6.8,Section 2.6.9 and Section 2.6.10 describe the multiple geoprocessing steps in more detail.

2.6.1. Vine Row Mask for Feature Extraction

QGIS 3.22, SAGA- GIS 7.8.2, together with the Python API (PyQGIS), were used to calculate the spectral features (VI), structural features (e.g. CHM), and texture features (GLCM). Image processing and segmentation were tested and performed using SAGA- GIS 7.8.2 and Python (libraries scikit-image). The usefulness of differentiating image features into spectral, structural, and textural features is reviewed in [39] and applied, for example, to oat yield predictions [33] or vine yield predictions [35]. The specific spectral, structural, and texture features were selected according to studies in agri- and viticulture combined with correlation analysis in accordance to the growth class ground truth.
The following subsections describe the generation of the different feature types. For the spectral features, those often used in viticulture context were selected. In the case of the structural and texture features, those generally applied for crops, biomass, and vegetation volume predictions were chosen. The general assumption was that spectral, structural, and texture features used in agriculture and, more specifically, in viticulture studies should be suitable as classification input for machine learning models concerning growth class and vigour predictions.

2.6.2. Spectral Feature Type for Classification (Vegetation Indices)

Threshold values of VI for vine row extraction were set and compared to each other to select the best extraction result. In addition to standard Vegetation Indices like NDVI (Normalized Difference Vegetation Index), NDREI (Normalized Difference Red Edge Index), OSAVI (Optimized Soil Adjusted Vegetation Index), and TSAVI (Transformed Soil Adjusted Vegetation Index), more unusual indices such as the NDWI (Normalized Difference Water Index) or GNDVI (Green Normalized Vegetation Index) were also calculated for correlation with the growth classes and as potential feature inputs for the machine learning models (see Table 3). All VI were calculated in QGIS 3.22 with the native raster calculator using the spectral bands of the multispectral orthomosaic. The VI were mainly selected based on the comprehensive review of [38] and their different application fields, primarily biomass, yield, vine row extraction, or water stress detection and classification.

2.6.3. Structural Feature Types for Classification

The CHM was calculated as the difference between DSM and DTM (DSM-DTM). The CHM and grapevine geolocations were subsequently exported to QGIS version 3.22 /ArcGIS (ArcMap version 10.5) to calculate structural height features around each grapevine stem, including average height (Hmean), maximum height (Hmax), minimum height (Hmin), median height (Hmedian), and height standard deviation (Hstd) (see Table 4). This was done as part of the zonal statistic spatial aggregation process (see Section 2.6.9). Moreover, the CHM volume was calculated and later used as additional structural input feature for ML growth class prediction. The derived structural features were also correlated with the spectral features, texture features, and the growth class ground truth, to identify the most important parameters for model input feature selection.

2.6.4. Texture Feature Types for Classification

Texture features were computed using the GLCM texture algorithm, introduced by Haralick et al. [78] and applied by several authors [33] in agricultural application fields. Basicallly, the GLCM provides information on the spatial relationship of pixel pairs in an image and is widely for image texture- based feature extraction and analysis in remote sensing. Using SAGA- GIS 7.8.2, different textural features were calculated for each of the five bands (R-G-B-RE-NIR), including variance (VAR), mean (ME) homogeneity (HOM), dissimilarity (DIS), contrast (CON), entropy (ENT), angular second moment (ASE), and correlation (COR). A 3 × 3 moving window was set for the calculations applied to each band of the multispectral orthomosaic. Due to the high correlation of the GCLM texture measures with the vegetation- sensitive spectral bands of the orthomosaic (R- NIR) the texture features of these channels were further investigated (correlation coefficient greater than 0.4 after Pearson [79]). Further details about these texture measures respectively features can be found in [80]. Table 5 presents the list of GLCM textural measures investigated in this study with the applied formula. The texture parameters were chosen as one more input feature group for the ML models to compare the model performance with the spectral feature group (VI) and the structural feature input group (e.g. CHM volume) and their combinations.

2.6.5. Extraction Mask Generation

The extraction mask of the vine rows was created by fusing a pixel-based threshold mask with the a mask created by using an object-based segmentation approach. The generated combined mask was utilized to extract the vine row canopy and the associated spectral, structural, and texture image features necessary for ML model input and the classification process.

2.6.6. Pixel-Based Mask

The OSAVI is especially useful for correcting spectral soil influences by optimizing the soil line (sl = 0.16). Various authors have discussed the Vegetation Indices mainly used for health monitoring and isolating the vine rows from the background primarily for extracting vine rows [81]. Under consideration of these studies, the OSAVI was selected from the VIs to generate the pixel-based mask (OSAVI mask). The OSAVI threshold (OSAVI > 0.7) was applied to create the binary OSAVI mask for vine row extraction in combination with visual validation. Alternatively, the threshold was determined using Otsus’s histogram-based method [82]. Moreover, threshold values from literature for vegetation, biomass, and crops were tested [83]. Compared with other selected VIs, OSAVI showed the best performance for vine row extraction and intermediate to high correlation with growth classes, according to the assessment of [43]. Technically, the mask was created using the raster calculator in QGIS 3.22.

2.6.7. OBIS (Object-Based Image Segmentation)-Based Mask

The OBIS-based segmentation methods use a different approach compared to classical pixel-based methods: they group the pixels rather than classify each pixel separately with machine learning techniques [84,85].
For the object-based image (unsupervised) segmentation of the OSAVI image, the OBIA (Object-Based Image Analysis) algorithm after Adam and Bischof [85] with the neighborhood classification of Moore from the SAGA- GIS Toolbox (Image Analysis- Image Segmentation GUI (Graphical User Interface)) with k-means clustering postprocessing was applied. Furthermore, variance in feature space was set to zero, variance in position space was set to zero, the similarity threshold was set to zero, the generalization was set to 1, the number of clusters was set to 6, and the clustered split option was checked.
The technical approaches of the OBIA techniques included in SAGA-GIS are explained in detail in [86,87,88] and other related studies. This OBIA technique has been used in multiple remote sensing studies in which complex images have been segmented. According to the documentation, the SAGA- GIS methodology of OBIS (Object-Based Image Segmentation) was applied [87]. The two methods analyze cells on an image using an approach similar to the lattice network. The two approaches tested in this study include the Neumann approach with Neighborhood 4 and the Moore approach with Neighborhood 8. The algorithms determine the most effective object features in OBIA, ensuring high separability among landscape features [89]. The Moore neighborhood is one of these methods based on a 2D square lattice with a central cell and eight neighbor cells. Moreover, the algorithm allows four more cells as ‘neighbors of neighbors’ to be considered as the extended neighborhoods of cells. Using the neighborhood analysis approach approved for pixels on an image, the machine can quickly group pixels based on the similarity of the spectral brightness of their neighbors. Both methods of neighborhood analysis are well-known techniques used for analyzing the pixels in cell automata of Object-Based Image Segmentation [86]. Finally, the vine row mask was manually selected and rasterized for subsequent feature extraction and spatial aggregation.

2.6.8. Merging Vine Row Masks

The final vine row mask was generated by fusing the pixel-based OSAVI mask (Section 2.6.6) with the OBIA mask (Section 2.6.7). The generated mask was utilized to extract the calculated spectral, structural, and GLCM texture features. These steps were technically achieved by using the native QGIS 3.22 raster calculator, which ensured that only vine canopy-exclusive pixels were considered for the spatial aggregation of spectral, structural, and textural features (following section).

2.6.9. Spatial Data Aggregation (Zonal Statistics)

The previously classified and extracted vine row pixels were subsequently used to aggregate the previously calculated features (spectral, structural, texture) using zonal statistics. In the process the mean, standard deviation, variance, minimum, maximum, range, and pixel count around each vines’ location were computed. Several studies have used zonal statistics for pixel volume calculation in different research fields of viticulture [28,90]. Technically, zonal statistics were performed in QGIS 3.22 with the Python API PyQGIS. The corresponding pixel values were aggregated within a rectangle measuring 1 m width (average width of vine rows) and a length of 2 × 0.60 m (1.20 m) (approx. average distance between individual vines in the same row for previous and subsequent grapevine stem in row) located around the individual vines (see red sampling rectangles around vines’ geopositions in Figure 6). The sampling rectangles were adjusted according to the average aspect of the slope derived from the DTM. For aggregation itself, only features intersecting the sampling rectangles and the vine row mask (see previous section) were used for spatial aggregation. The results of the zonal statistic aggregation were automatically joined onto the grapevine point shapefile as additional columns within the attribute table of the shapefile. The aggregated values around the single vines were used to calculate the CHM volume (see Section 2.6.10). Most importantly, the prepared vine canopy extracted spectral, structural and texture features for each grapevine, presented as one row in the attribute table, were now ready as input for the ML-based growth classifications.

2.6.10. Calculation of the Canopy Height Model (CHM) Volume for Single Vines

The CHM-based pixel volume of the single vines was calculated by combining the CHM with the different extraction masks (extracted features) to consider only pixels previously detected as part of the vine row. The CHM was calculated as the difference between the DSM and the DTM within the vine row mask. Subsequently, the volume (CHM volume) of the individual vines was estimated from the mask’s product-extracted CHM with the area and number of vine row pixels within the zonal statistics sampling rectangles intersecting the vine row mask around each vine. Due to its high correlation with the growth classes, the calculated single vine volume was chosen as an additional input parameter for the model training (see Table 6). The Mann–Whitney-U-test was used to quantify the separability of the growth classes after [43] based on the CHM volume for each growth class combination pair. Similar approaches have been applied to estimate the pixel volume of vine rows and stems [28]. Moreover, based on the overlapping of the canopy of adjacent vines, the total real canopy volume of each vine cannot be 100% precisely computed. Therefore, the calculated values have a variance error due to this overlapping effect besides other methodological errors that sum up to the total uncertainty of the absolute vine volume estimation.

2.7. Growth Class Estimation Modeling

Following the geoprocessing and image processing steps (Section 2.6.1,Section 2.6.2,Section 2.6.3,Section 2.6.4,Section 2.6.5,Section 2.6.6,Section 2.6.7,Section 2.6.8,Section 2.6.9 and Section 2.6.10), the growth class prediction, based on two selected ML classifiers (RFC, SVM), was an essential step to derive overall and growth-class-specific prediction metrics to evaluate the suitability of the extracted features for predicting growth classes according to the assessment of [43]. For the ML classification process the Python library sklearn was used to fit the training data to the ground truth target classes with the ML classiifers. Before model fitting, the feature groups (spectral, structural, and texture) and their combinations (1–7) (see Table 6) were selected, scaled (standard scaling), and cleaned for NoData values.
The generation of the test and training datasets of the input features were performed with repeated-k-fold- cross-validation (repeats = 5, splits = 5, random state = 1, n jobs = −1) by splitting the input data (input features and growth class ground truth) into train and test data sets. From GLCM derived features, those with the highest correlation after Pearson (calculated with RStudio version 4.3.3), to the growth classes were selected as input features in addition to the spectral and structural features. Mostly band 3 (red band) features were included in this category due to the high correlation between the red band and vegetation vigour. The RFC and SVM classifiers were chosen to fit the training data to the growth class ground truth data. These classifiers belong to the most efficient classification methods in agriculture crop and biomass prediction, and detection, including grapevine yield estimations [8]. Subsequently, the hyperparameters of the models were optimized with a hyperparameter-tuned grid search [33]. The essential steps for the model generation are summarized in Table 7. The model’s output generated n = 25 single metric outputs for each growth class. These values were used for univariate statistics and the evaluation of the consistency of the metrics’ output. For model evaluation, the validation metrics were calculated for the train and test data to search for overfitting and underfitting tendencies. The models and the validation metrics (accuracy, f1-weighted) were selected based on literature sources, primarily, biomass estimation and crop modeling research. Nevertheless, the specific geometrical properties of vine canopies and class frequency balance of the label dataset were also considered by several authors [91,92,93], adapting the classification procedure accordingly. These aspects are also shortly considered in the discussion section (see Section 4.4). Furthermore, the f1-weighted scores were calculated to test for the impact of imbalance on the generated dataset [93,94,95]. To quantify the statistical difference of the average model prediction metric results, the Mann–Whitney-U-test was applied for all model combinations (1–7) and metric pairs (accuracy vs f1-weighted). Moreover, the class-specific accuracy (confusion matrix) was calculated for the model output with the best overall accuracy score. Additionally, user and producer accuracy were calculated. The user accuracy for each growth class was calculated as the ratio of the sum of all correctly identified growth classes for this class, divided by the sum of all other incorrectly labeled growth classes for the same growth class and expressed in %. For producer accuracy, for each growth class, the sum of all correctly identified growth classes for this growth class was divided by the sum of all other incorrectly (ML model) predicted growth classes for the same growth class and also expressed in %. This approach allowed for a more specific discussion of the classification results and trends of the individual growth classes (see Section 4.6).

3. Results

The following subsections present the CHM volume calculation and the results of the growth class ML model predictions based on the geoprocessing workflow from Section 2.6.1,Section 2.6.2,Section 2.6.3,Section 2.6.4,Section 2.6.5,Section 2.6.6,Section 2.6.7,Section 2.6.8,Section 2.6.9,Section 2.6.10 and Section 2.7.

3.1. CHM Volume Results

The results of the CHM model calculation are shown Figure 7. The stars (*) connected by the brackets indicate the strength of significant differences between the growth class grouped CHM volume calculations, based on the paired Mann–Whitney-U-test (Figure 7). The graph shows that all neighbor growth classes significantly differ to each other according to the CHM volume. This CHM volume growth class separation is more pronounced than spectral feature only approaches (e.g. OSAVI). Subsequently, the CHM volume calculation result was used as one additional structural input feature for ML classifications.

3.2. Overall Growth Class Prediction

In the following Section 3.2 the major results of the classification metrics are presented according to the different input feature types and their combinations.

3.2.1. Spectral Feature Group Estimation

The best model performance for the spectral input feature group (1) of the test data set was received from the SVM 1 model with an accuracy = 40.65% and f1-weighted of 37.48.
On the other hand, the RF 1 model yielded the lowest accuracy (accuracy = 30.41%, f1-weighted = 20.98%). For the train data the best model performance for the spectral feature group was received from the SVM 1 model with an accuracy of 50.48% and f1-weighted of 46.76%. For the train dataset, the RF 1 model predicted the lowest accuracy (accuracy = 31.51%, f1-weighted = 22.25%) (Table 8).

3.2.2. Structural-Feature-Based Growth Class Estimation

Higher estimation accuracies were obtained from models based on the structural input feature group (2). The best model performance for the structural feature group of the test data was received from the SVM 2 model with an accuracy = 47.38% and f1-weighted of 46.21%. On the other hand, the RF 2 model yielded an accuracy of 40.4% and f1-weighted of 31.57%. The best model performance for the structural input feature group of the train data was received from the SVM 1 model with an accuracy of 50.34% and f1-weighted of 51.23%. For the train dataset, the RF 2 model yielded the lowest accuracy (accuracy = 41.58%, f1-weighted = 32.66%) (compared in Table 8). The structural feature types influences prediction metrics more strongly than non-structural features, including models 1, 3, and 5. The higher bivariate correlation coefficients indicate the improved metrics scores compared to the spectral features (see Supplementary Materials).

3.2.3. Texture-Feature-Based Growth Class Estimation

The best model performance for the texture-feature-based group (3) of the test data was received from the SVM 3 model with an accuracy = 37.98% and f1-weighted of 47.72. On the other hand, the RF 1 model yielded an accuracy of 33.17% and f1-weighted of 28.32%. The best model performance for the train data texture input feature group was received from the SVM 1 model with an accuracy of 40.13% and f1-weighted of 52.49%. For the train dataset, the RF 3 model yielded an accuracy of 34.10% and f1-weighted of 29.10% (compared in Table 7). The texture feature group input predictions are less favorable than the structural feature inputs and are comparable to the spectral feature input results.

3.2.4. Summary of Impact of Feature Types and Feature Type Combination on Model Classification

In comparison, model predictions show significant differences according to the chosen input feature types, feature group, and selected ML classifier. Structural features like CHM and the calculated CHM volume led to improved classification metrics compared to spectral features (Table 8). For the third texture-features-type only input group (3), model predictions were slightly better than for the spectral feature only input group (1) of RFC. In contrast, texture features they are even less favorable in the case of the SVM classifier compared to spectral features. Generally, the spectral- and texture-only feature groups differ only slightly and have the least desirable classification metrics for both classifiers. Moreover, combining all input feature groups and feature types leads to the best and most robust classification results for growth class prediction, also for both classifiers (RFC, SVM). Nevertheless, the RF classifier showed decreased prediction metrics for most input feature groups compared to the SVM classifier (Table 8). In summary, selecting and combining the extracted input features is essential for model prediction results concerning growth class predictions.

3.2.5. Overall Growth Class Prediction

The metrics for the train and test dataset predictions for the repeated k-fold cross-validation (n = 25) ML classifications (1-7) are summarized in Table 8 and visualized for the accuracy of the SVM and RFC classifier models as boxplots with significance stars derived from the Mann–Whitney-U-test in Figure 8 and Figure 9.
ML models including the structural input feature group (2,4,6,7), are marked in gray to discuss the influence of the structural features compared to the other feature types in Section 4. The highest accuracy of the test data prediction with an accuracy = 48.51% and f1-weighted = 45.50% was obtained from the SVM classifier with the combination of spectral, structural, and texture features as input for model training fit (SVM 7). The highest accuracy of the train data prediction (accuracy = 53.7%, f1-weighted = 50.11%) was also obtained from SVM modeling with the combination of all feature type groups (SVM 7) as input features. For the RFC classifier, the highest test data accuracy (accuracy = 40.0%, f1-weighted = 31.5%) was also achieved with the combined spectral, structural, and texture feature input (RF 7), followed closely by the structural-feature input-only group accuracy (accuracy = 40.65%, f1-weighted = 37.5%) (RF 2). Regarding training data accuracy, the structural feature group (2) received the best accuracy of 41.58% and f1-weighted of 32.66% for the RFC classifier (RF 2). Moreover, all models containing structural feature inputs (2,4,6,7) have, on average, improved classification metrics compared to the non-structural-feature-containing models (1,3,5) for both classifiers (RFC, SVM). The worst prediction metrics are obtained from the RF classifier models, which contained no structural feature inputs for training (RF 1, R F 3, RF 5). Within the SVM classifier group, the models without structural features (SVM 1, SVM 3, SVM 5) showed decreased classification metrics compared to models with structural feature input (SVM 2, SVM 4, SVM 6, SVM 7). Generally, the SVM classifier obtained consistently higher accuracy and f1-weighted scores on the average order of 10- 20% of absolute accuracy. Moreover, the standard deviations for the metrics also showed less range for the models with the best model mean scores (SVM 4, SVM, 6, SVM 7) compared to the less successful prediction models (RF1, RF 3, RF5). Therefore, the SVM models, especially including the structural-feature-type group as input, seem more robust and consistent than the other RFC models without these feature types.

4. Discussion

The following section discusses the results presented in Section 3 within the context of vine row extraction, correlation of the growth classes to the calculated CHM pixel volume, and the classification metrics of the ML models.

4.1. Extraction Masks

The pixel-based and OBIA-based vine- row extraction approaches are compared and discussed in multiple remote sensing studies and reviews [10]. The general approach in this study used a combined pixel- and OBIA-generated mask for spectral, structural, and texture feature extraction. Moreover, the merged vine row mask metrics were evaluated visually and calculated for a short segment of a vine row according to the OBIA accuracy assessment from [96,97]. The intersection with the digitized ground truth vine row is approximately 89% for the combined mask, indicating a relatively good fit with the template mask. Nevertheless, the focus of this study was not to optimize the vine row mask; therefore, this aspect will not be discussed in more detail here.

4.2. Structural Parameters and CHM Volume

The results of the CHM volume calculation (Figure 7) reflect the high correlation and influencing power of the CHM volume for growth class separation. Other studies have also detected and discussed the high correlation power of structural features like the CHM or the CHM volume [10]. Moreover, these studies also showed a strong correlation between the CHM and the CHM volume for biomass and crop yield estimation in different agricultural scenarios. CHM models were primarily used for vine row extraction. Moreover, several authors used the CHM, extracted pixels for volume calculation, and compared them with ground truth measurements [81]. Several authors discussed the structural parameters’ power as prediction parameters and input features for other regression analyses and classifications, especially ML-based classifiers like SVM or RF [33]. In most cases, structural parameters showed superior biomass and growth prediction ability compared to spectral features. This is also indicated by the high correlation coefficients and significance tests, which revealed more vital effect sizes for CHM volume than spectral features like the OSAVI against the growth classes. A high correlation between the vine’s volume, biomass yield, and structural parameters (CHM) is discussed by several authors [10,28,98]. Nevertheless, due to the complex internal structure of grapevines, the CHM and also manual measurement methods can only approximate the actual volume of vines [3]. The influence of structural parameters on machine learning model classification in this study is further discussed in Section 4.4.2. Due to the highly positive correlation of the CHM volume to the growth classes after [43], the CHM volume is used as an additional input feature for ML model training.

4.3. Model Validation

Evaluating ML model predictions is essential for interpreting classification metrics, robustness, and reproducibility of the growth class prediction [92]. Comparing the generally higher training to test metric score results hints at overfitting of the model predictions due to the limited amount of available training data, including ground truth data, multispectral orthomosaics, and CHMs during sampling time (compare train and test data metrics from Table 8 and Figure 10). Especially the SVM models show a higher difference between the train and test data (higher train than test score) values of the repeated cross-fold validated overall accuracy (OA), which is probably a result of the sensitivity of kernel-based classification models. The tendency of overfitting for regression, machine learning, and deep learning models was also observed by [99].
Moreover, to test for imbalance, the OA is compared with the f1-weighted score (Figure 11 and Figure 12). The f1-weighted score considers the sample size for each growth class about the total sample size. Pronounced differences between the overall accuracy and the f1-weighted score indicates an imbalance of the dataset due to the underrepresentation of some growth classes (e.g. growth classes 0, 1, 9) compared to more frequent growth classes (e.g. 3, 5, 7). This issue is also considered in the discussion part of Section 4.6. In ongoing studies, more training data from multiple sampling dates and different vine fields should be incorporated to improve the model predictions’ fit, consistency, and reproducibility. Moreover, more samples from less frequent growth classes should be integrated to avoid imbalance bias in the prediction metrics. Therefore, a batch workflow to calibrate and process the data from different data acquisitions should be developed in future studies.

4.4. Model Performance Evaluation

The impact and contribution of the essential feature types and input feature groups is evaluated in the following section, considering their hierarchy in classifying vine growth and vine vigour.

4.4.1. Influence of Spectral Features

Spectral features’ correlation and classification power for predicting biomass production and crop yield has been observed, with varying correlation strength considering the structure, growth characteristics of vegetation type, sensor type, and applied spectral indices [35,67,100,101,102,103,104,105,106,107,108,109,110,111,112,113,114,115,116]. Generally, the spectral feature type based classification was less successful, compared to models with structural feature type inputs (Table 8). This indication can result from the intermediate correlation of the growth classes with spectral features due to saturation effects and the complex internal geometric structure of the vine canopy [35]. Moreover, the SVM models predictions show improved classification metrics than the RF classifier with the same input features. However, spectral features are essential for predicting different types and origins of photosynthetic materials. Many studies used spectral features as input for yield predictions in various agriculture and viticulture application scenarios [64,116,117,118]. It is worth noting that models based on spectral features underestimated biomass samples with higher values. One reason for this could be optical saturation effects of the VI. Similar observations were observed in winter cover crop biomass estimation [119] and soybean biomass and LAI estimation [120]. Especially, VI based on NIR-band and red-band ratios (e.g. NDVI) tend to saturate for highly dense canopies [121,122], which resulted in the poor performance of predictive models due to saturation effects. Prabhakara et al. [119] reported that the NDVI showed asymptotic saturation in the higher range of rye biomass (>1500 kg/ha). VI are also environment- and sensor type-sensitive and -specific [123]. The surface sensor data derived from nadir’s view often do not reflect internal 3D canopy structure and geometrical patterns, depending on the crop or plant type. These factors distort the correlation between the growth class ground truth and the “real” volume of vines. Due to the complex internal architecture of vines [31,43], these effects are relatively pronounced for sensor- based viticultural investigations.

4.4.2. Influence of Structural Features

Many studies have validated the potential of structural features in biomass estimation [124,125,126] and vine row volume calculation [10]. Bendig et al. [127] reported that canopy height derived from the Crop Surface Model (CSM) is a suitable indicator of biomass in barley. Moreover, Shu et al. [128] used the CHM and the canopy coverage to estimate maize above-ground biomass. In this study, the structural input features provided increased classification metrics for both classifiers (RFC and SVM) compared to the spectral feature type-only model results (see Table 8). A possible explanation of the superior prediction performance based on structural features over those based on spectral features is that canopy structural features can provide more information about the three-dimensional canopy structure and, therefore, also show higher correlations to the growth classes and the CHM volume. The superiority in estimating biomass based on structural or combined feature types is also claimed by the study of Michez et al. [98] by combining LiDAR-derived height measures with VI.

4.4.3. Influence of Texture Features

Several authors used texture canopy features and related vegetation components as input for biomass estimation or yield prediction of different crop types [10,33,39,129]. Delenne et al. [18] presented, developed, and discussed different texture extraction techniques for vine row detection. Therefore, texture features can potentially add explanatory power in characterizing the growth of vines besides grapevines’ spectral and structural signature [130,131]. Wengert et al. [132] recognized that GLCM-based textural features improved the estimation of barley dry biomass and Leaf Area Index (LAI). Similar results were also documented in the above-ground biomass estimation of legume grass mixtures [111], rice [105], and winter wheat [106]. A possible explanation for the improved estimation accuracy in some of these studies is that texture features consider the spatial variation in the pixels and provide additional information about the physical structure of canopies, including canopy edges, and the overall canopy architecture [123,133]. Nevertheless, in this study, the growth class prediction results are comparable to the spectral-feature-input-only predictions and inferior to the structural input feature models (see Table 8). This is probably a reflection of the same issue mentioned with the spectral features, capturing only surface nadir reflectance properties of the vine canopies, without capturing the complex internal structure. Moreover, texture features are based on spectral bands; thus, they show some collinearity with spectral features. However, in contrast to spectral features, texture features can characterize canopy architecture and structure patterns to some extent [130], as well as weaken saturation issues and suppress soil background effects. The only light difference between the spectral and texture features in this study does not hint at a superior performance of textural features over spectral features, like in the study of [131], where forest biomass was predicted. This non-agreement is probably a result of different vegetation types and associated structure, which play a significant role in sensor data correlation and varying types of prediction variables (continuous biomass yield vs. categorical growth classes). This may also reflect the texture features’ inability to consider minor differences in the canopy texture between the growth classes.

4.4.4. Combining Different Feature Types

Numerous studies have highlighted the potential of combining structural features like CHM with spectral (VI) or texture features rather than using them separately [64,106,115,130,131,134,135,136,137,138,139,140], which resulted in robust and improved estimations of biomass yield. Moreover, [137,138,139,140] have tested the combination of multiple information sources by integrating non-spectral features (3D thermal sensors) or textural features with spectral features for evaluating grassland biomass and cultivated crops. Consistent with previous studies, the results in this work also show that a combination of spectral, structural, and texture features (7) resulted in improved accuracies and f1-weighted scores than spectral or structural features alone (see Table 8). Primarily, canopy structural features can provide valuable information about canopy architecture not supplied by spectral features and can, to some extent, overcome the saturation problem of spectral features. Nevertheless, the highly varying porosity of the canopy cannot be perfectly modeled based on this 2D assumption rather than an overall estimation. For vines, Matese and Di Gennaro [35] used spectral and structural input features derived from UAV images of vine fields as input for yield prediction. Compared to classical regression methods, they showed the advantage of combining spectral features (NDVI) with structural canopy features and the increased prediction metrics of machine learning classifiers, including RF, SVM, and GPR (Gaussian Process Regressor). Apparently, in this study, combining different input feature groups mainly including structural features (e.g. Canopy Height, Canopy Surface) leads to the best prediction results and most robust models with relatively low standard deviations for the growth classes after [43]. Shu et al. [99] also concluded that combining spectral, structural, and texture features provides superior prediction metrics compared to classical regression methods like PLSR (Partial Least Square Regression).

4.5. Comparison of Machine Learning Classifiers (RFC vs. SVM)

The classification results and trends based on the input features are similar for both classifiers (RFC, SVM), with improved prediction metrics for models with structural features that are exclusive (2) or a combination of features, including structural feature types (4, 6, 7) (Table 8). Generally, the SVM models showed slightly higher accuracies and f1-weighted scores than the RFC model predictions (higher accuracy, f1-weighted, lower Accuracystd and f1-weightedstd). This lower prediction quality is probably a result of the better performance of SVM when distinct classes are predicted, the sample size of the input data is relatively small, and the correlation between the input features and the growth classes is primarily linear. Moreover, the data size of the input features for each growth class could be too small, especially for the RF classifier, which generally requires more training samples for a sufficient model fit than SVM. Moreover, SVM could be more suitable due to the linearity of the growth classes and the margin optimization of the classes relative to each other. Otherwise SVM needs more time for model fitting, and it is more challenging to handle the hyperparameter tuning compared to RFC. Moreover, SVM is better at handling fewer samples with high input dimensions than RF, whereas RF is particularly effective and precise for many samples with high dimensions and target classes. Adugna et al. [141] discuss such discrepancies for the model classifiers, especially for the SVM and RF classifiers. Sheykhmousa et al. [142] compared RFC and SVM classifiers in the remote sensing applications of landscapes, considering many external factors like image resolution, sensor type, type of target classes, complexity of the landscape, and internal characteristics of the classifier, like, hyperparameter optimization such as kernel selection and number of training features for model fitting, influencing the prediction quality and metrics. Therefore, including more growth class samples as training input and widening the feature space could shift the more favorable classifier from SVM to RFC in future classifications, with extended datasets. A discussion comparing the prediction results and model robustness of different machine learning classifiers is also provided recently by [143].

4.6. Class-Specific Accuracies of the Best Model

To discuss growth class-specific prediction matches, a confusion matrix was calculated for the ML model with the best overall accuracy concerning the test data (SVM 7) (see Table 9). The pattern of the class-specific accuracy is similar for all models and classifiers, differing mostly in relative accuracy values due to feature inputs, classifier type, and selected hyperparameter. For discussing the class-specific accuracies, the confusion matrix with the highest overall accuracy was calculated (SVM 7, compare Table 8).
Class-specific accuracies indicate a more or less pronounced range of accuracy predictions for all growth classes, considering direct neighbor classes as matches, with overall prediction accuracies of more than 80% for most growth classes (see color code marks in the confusion matrix, Table 9). Therefore, strongly pronounced prediction outliers (e.g. labeled growth class 3 predicted as growth class 7 (5.01%)), or even higher discrepancies between ground truth and model prediction are rare or even not apparent (e.g. labeled growth class 9 predicted as growth class 3). This indicates the general suitability of the selected input feature groups for correlation and classification concerning the growth class assessment after [43]. The most substantial overlap with neighbor growth classes occurs for growth class 9 and growth class 0. This is probably a result of different factors, including more robust labeling errors, sensor data effects, or separation issues of the interrow area from the vine row section (especially growth class 0 and 1). A visualization of the difference between the growth class ground truth and model prediction for each grapevine position is presented in Figure 13. Especially, the intermediate growth classes 5 (67.7%) and 7 (57.1%) showed the highest match rate, which may result from less overlap and interference with vine row vegetation and soil background compared to the lower growth class spectrum (growth classes 0, 1, 3), on the one hand and fewer saturation effects compared to the highest growth class 9, on the other hand. Other reasons might be the varying number of ground truth samples for each growth class (imbalance of the dataset) combined with ground truth labeling errors. Moreover, differences in user and producer accuracies in Table 9 and the train and test data predictions (higher train than test data accuracy) indicate model overfitting. The difference in user and producer accuracy also hints at a lack of sensitivity and specificity compared to positive and negative precision (compare [144]). Therefore, user accuracy (indicating better training data prediction) provides slightly higher matches when summed up for all classes than producer accuracy (lower test data prediction), which is also indicated in Figure 10, for the overall prediction metrics.
On the one hand, the model predictions overestimate the low-growth classes (0,1), which results in negative values, most likely resulting from difficulties separating the vine row from interrow vegetation. Therefore, a substantial overlap with growth classes 1 and 3 is apparent. On the other hand, higher end-growth classes (7,9) are, on average, more underestimated by model predictions (see Table 9). This indication is probably a result of different effects and error sources, including saturation effects due to the complex internal structure of the vines, sensor view direction, sensor type, and labeling errors of the ground truth data. Such effects have also been observed in other remote sensing studies, like those of Zhen et al. [145]. Moreover, landscape features, including different grapevine varieties and vine stages, are more challenging to differentiate when structurally similar to each other, concerning the sensor data image features (spectral, structural, and texture features) [146,147]. Despite these limitations, the growth class assessment after [43] is basically suited for ML classification based on UAV-sensor data. The shortcomings are mostly the sum of different interacting factors mentioned above (labeling errors, sensor type, sensor direction, feature types, distortions), which could be overcome with other sensor types, respectively, with fusion of different sensor types combined with a reduction in ground truth labeling errors. Moreover, the training dataset size could be enhanced by using more ground truth and sensor data to improve the model fitting process and prevent an imbalance of the dataset. Progressively, additional and relevant input features like other ground truth data, such as infection type and strength, could improve the growth class predictions and be compared with sensor-data-only feature-based classification results. These results could be compared with similar sensor-derived feature inputs and used to validate ground truth transferability to sensor-derived feature classification. Lastly, if necessary, the assessment could be slightly modified to improve the classification results.

5. Conclusions and Future Work

In this study a new geoprocessing workflow was developed and evaluated in combination with ML model predictions (RFC, SVM) for separating growth classes according to the growth class assessment after Porten [43], based on different input (spectral, structural, texture) features derived from highly-precise georeferenced UAV-sensor data. A vine row mask was generated by combining pixel- and object-based image segmentation approaches for vine canopy extraction.
The selected ML classifiers (SVM, RFC) were relatively efficient in vine vigour classification. SVM produced higher estimation accuracies than RFC models. Therefore, integrating image feature types, like spectral, structural, and textural features, offers a promising avenue for general growth class prediction
The primary outcomes of this study are as follows:
(1)
Combining spectral, structural, and texture features indicates the best classification results for both machine learning classifiers (Random Forest Classifier and Support Vector Machine Classifier). Nevertheless, SVM performed better than RFC, due to classifier properties and ground truth data set characteristics.
(2)
The structural input features are the most positively influencing feature type. The canopy structural features achieved higher accuracy and f1-weighted scores, than models using spectral or textural features alone or combined. Canopy structural features most likely provide a more accurate representation of canopy architecture than spectral and texture features.
(3)
Although less influential than structural features, canopy spectral and texture features are also an essential indicator for growth class estimation. They may also offer a means to overcome saturation issues associated with spectral features from nadir camera positions.
(4)
The class-specific accuracies show that most growth classes were correctly predicted, or the mismatch between the predicted and labeled classes was only minor. Therefore, the class-specific accuracy would increase to over 80 or 90% when considering the neighborhood growth classes as matches. Nevertheless, some classes, like growth class 0, 1, or 9, are difficult to separate with sufficient probability due to different error sources and distortions. These include subjective ground truth labeling (inter- and intrapersonal sources of error), uncertainty in distinguishing the growth classes, different approaches of the individual classification methods, complex correlation and varying influence strength of input parameters to the particular ML growth class prediction, and other disruptive influences (e.g. radiometric interference) during data acquisition.
(5)
The comparison of training and test datasets overall accuracy and growth class specific user and producer accuracies of the class-specific prediction hint to overfitting of the different models for both classifiers (SVM, RFC). Moreover, the difference between unbalanced (accuracy) and balanced prediction metrics (f1-weighted score) indicate imbalance of the ground truth dataset, which should be further considered and improved in future studies, for example, by integrating ground truth data from other years, growth stages and other vine fields.
In summary, UAV-derived multispectral image data can provide a good foundation for growth class and grapevine vigour predictions. Nevertheless, there are several limitations to this approach. Based on the complex geometry of vine canopies, 2D multispectral and CHM approaches oversimplify the complex 3D geometrical structure of grapevine canopies. Therefore, considering additional sensor types like LiDAR-sensors combined with multi-, hyper-, and thermal sensor data provide additional geometrical, spectral and thermal information. Combined with enhanced sensor data acquisition and subsequent processing routines, extended sensor data types and sensor data fusion techniques can improve growth class and vine vigour predictions even further. Hence, future work can contribute to enhance growth classification assessments, vigour predictions and canopy volume estimations, for a more efficient vineyard management and viticulture automatization approaches.

Supplementary Materials

The following supporting information can be downloaded at: https://github.com/RD1989NEW/code_publication.git

Author Contributions

Conceptualization, R.P.D., M.A.W., M.P., T.U. and R.R.; Formal analysis, T.U.; Funding acquisition, M.A.W.; Investigation, R.P.D.; Methodology, R.P.D., M.P., T.U. and R.R.; Project administration, M.A.W.; Supervision, M.A.W. and T.U.; Validation, R.R.; Visualization, R.P.D.; Writing—original draft, R.P.D.; Writing—review and editing, R.P.D., M.A.W. and M.P. All authors have read and agreed to the published version of the manuscript.

Funding

The work is conducted within the project “Smarter Weinberg”, funded by the German BMDV (Bundesministerium für Digitales und Verkehr) under the funding scheme of the “5G innovation competition”, funding reference: 45FGU114, in the time frame 11/2021 to 01/2025, https://smarter-weinberg.de/ (accessed on 02 January 2025).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in the study will be available from the corresponding author upon request.

Acknowledgments

The work presented here was partly carried out as part of the Smarter Weinberg project. The Smarter Weinberg joint project (https://smarter-weinberg.de) is funded by the German Federal Ministry for Digital and Transport in the funding program for 5G implementation projects from November 2021 to January 2025.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

ANOVAAnalysis Of Variance
APIApplication Programming Interface
CHMCanopy Height Model
CRSCoordinate Reference System
CSMCrop Surface Model
DSMDigital Surface Model
DTM
EPSG
Digital Terrain Model
European Petroleum Survey Group Geodesy
ETRSEuropean Terrestrial Reference System
GCPGround Control Points
GDALGeospatial Data Abstraction Library
GISGeographic Information System
GLCMGray-Level Co-occurrence Matrix
GNDVIGreen Normalized Vegetation Index
GNSSGlobal Navigation Satellite System
GPRGaussian Process Regressor
GPSGlobal Positioning System
LAILeaf Area Index
LiDARLight Detection and Ranging
MLMachine Learning
NDVINormalized Difference Vegetation Indices
NDREINormalized Difference Red Edge Index
NDWINormalized Difference Water Index
NIRNear Infrared
OA
OBIA
Overall Accuracy
Object-Based Image Analysis
OBISObject-Based Image Segmentation
OSAVIOptimized Soil Adjusted Vegetation Indices
QGISQuantum GIS
PLSRPartial Least Square Regression
RE
RGB
Red-Edge
Red Green Blue
RFCRandom Forest Classifier
RMSERoot Mean Squared Error
SAGASystem for Automatic Geoscientific Analysis
SAPOSSAtelitten POSitionierungsdienst der deutschen Landvermessung
SfMStructure from Motion
SVMSupport Vector Machines
TSAVITransformed Soil Adjusted Vegetation Indice
VIVegetation Indices
UAVUnmanned Aerial Vehicle
UTMUniversal Transverse Mercator

References

  1. Zhang, C.; Kovacs, J.M. The application of small unmanned aerial systems for precision agriculture: A review. Precis. Agric. 2012, 13, 693–712. [Google Scholar] [CrossRef]
  2. Lllorens, J.; Llop, J.; Escola, A. Ultrasonic and LIDAR Sensors for Electronic Canopy Characterization in vineyards, Advances to improve Pesticide Application methods. Sensors 2011, 11, 2177–2194. [Google Scholar] [CrossRef]
  3. Campos, J.; García-Ruíz, F. Assessment of Vineyard Canopy Characteristics from Vigour Maps Obtained Using UAV and Satellite Imagery. Sensors 2021, 21, 2363. [Google Scholar] [CrossRef]
  4. Moreno, H.; Andújar, D. Proximal sensing for geometric characterization of vines: A review of the latest advances. Comput. Electron. Agric. 2023, 210, 107901. [Google Scholar] [CrossRef]
  5. Tardaguila, J.; Stoll, M.; Guitiérrez, S.; Profitt, T.; Diago, M.P. A review: Smart applications and digital technologies in viticulture. Smart Agric. Technol. 2021, 1, 100005. [Google Scholar] [CrossRef]
  6. Ferro, M.V.; Catania, P. Technologies and Innovative Methods for Precision Viticulture: A Comprehensive Review. Horticulture 2023, 9, 399. [Google Scholar] [CrossRef]
  7. Zualkernan, I.; Abuhani, D.A.; Hussain, M.H.; Khan, J.; El Mohandes, M. Machine Learning from Precision Agriculture Using Imagery from Unmanned Aerial Vehicle (UAVs): A Survey. Drones 2023, 7, 382–418. [Google Scholar] [CrossRef]
  8. Matese, A.; Di Gennaro, S.F. Technology in precision viticulture: A state of the art review. Int. J. Wine Res. 2015, 7, 69–81. [Google Scholar] [CrossRef]
  9. Khaliq, A.; Comba, L.; Biglia, A.; Aimonino, D.R.; Chiaberge, M.; Gay, P. Comparison of Satellite and UAV-Based Multispectral Imagery for Vineyard Variability Assessment. Remote Sens. 2019, 11, 436. [Google Scholar] [CrossRef]
  10. Sassu, A.; Gambella, F.; Ghiani, L.; Mercenaro, L.; Caria, M.; Pazzona, A.L. Advances in Unmanned Aerial System Remote Sensing for Precision Viticulture. Sensors 2021, 21, 956. [Google Scholar] [CrossRef]
  11. Galidaki, G.; Panagiotopoulou, L.; Vardoulaki, T. Use of UAV-borne multispectral data and vegetation indices for discriminating and mapping three Indigenous vine varieties of Greece. J. Cent. Eur. Agric. 2021, 22, 762–770. [Google Scholar] [CrossRef]
  12. Bellvert, J.; Girona, J. The Use of Multispectral and Thermal Images as a Tool for Irrigation Scheduling in Vineyards. Options Méditerranéenes. Ser. B Stud. Res. 2012, 67, 131–137. [Google Scholar]
  13. Comba, L.; Gay, P.; Primicerio, J.; Aimonino, D.R. Vineyard Detection from Unmanned Aerial Systems Images. Comput. Electron. Agric. 2015, 114, 78–87. [Google Scholar] [CrossRef]
  14. Pádua, L.; Marques, P.; Hruška, J.; Adão, T.; Bessa, J.; Sousa, A.; Peres, E.; Morais, R.; Sousa, J.J. Vineyard Properties Extraction Combining UAS-based RGB Imagery with Elevation Data. Int. Remote Sens 2018, 39, 5377–5401. [Google Scholar] [CrossRef]
  15. Nolan, A.P.; Park, S.; O’Connell, M.; Fuentes, S.; Ryu, D.; Chung, H. Detection and segmentation of vine rows using high resolution UAS imagery in a commercial vineyard. In Proceedings of the 21 International Congress on Modelling and Simulation, Gold Cost, Australia, 29 November–4 December 2015; Modelling and Simulation Society of Australia and New Zealand (MSSANZ): Canberra, Australia, 2015. [Google Scholar]
  16. Song, Z.; Zhang, Z.; Yang, S.; Ding, D.; Ning, J. Identifying sunflower lodging based on image fusion and deep semantic segmentation with UAV remote sensing imaging. Comput. Electron. Agric. 2020, 179, 105812. [Google Scholar] [CrossRef]
  17. Cinat, P.; Di Gennaro, S.F.; Berton, A.; Matese, A. Comparison of Unsupervised Algorithms for Vineyard Canopy Segmentation from UAV Multispectral Images. Remote Sens. 2019, 11, 1023–1046. [Google Scholar] [CrossRef]
  18. Delenne, C.; Durrieu, S.; Rabatel, G.; Deshayes, M.; Bailly, J.-S.; Lelong, C. Textural approaches for vineyard detection and characterization using very high spatial resolution remote-sensing data. J. Remote Sens. 2008, 29, 341–345. [Google Scholar] [CrossRef]
  19. Ivorra, E.; Sánchez, A.J.; Camarasa, J.G.; Diago, M.P.; Tardaguila, J. Assessment of grape cluster yield components based on 3D descriptors using stereo vision. Food Control 2015, 50, 273–282. [Google Scholar] [CrossRef]
  20. Palleja, T.; Landers, A.J. Real time canopy density validation using ultrasonic envelope signals and point quadrat analysis. Comput. Electron. Agric. 2017, 134, 43–50. [Google Scholar] [CrossRef]
  21. Mesas-Carrascosa, F.J.; Castro, A.I.; Torres-Sánchez, J.; Triviῇo-Tarradas, P.; Jiménez-Brenes, F.M.; Garcia-Ferrer, A.; López-Granados, F. Classification of 3D-Point Clouds Using Color Vegetation Indices for Precision Viticulture and Digitizing Applications. Sensors 2020, 21, 956. [Google Scholar] [CrossRef]
  22. Comba, L.; Biglia, A.; Aimonino, D.R.; Gay, P. Unsupervised detection of vineyards by 3D point-cloud UAV photogrammetry for precision agriculture. Comput. Electron. Agr. 2018, 155, 84–95. [Google Scholar] [CrossRef]
  23. Yang, Y.; Shen, X.; Cao, L. Estimation of the Living Vegetation Volume (LVV) for Individual Urban Street Trees Based on Vehicle-Mounted LiDAR Data. Remote Sens. 2024, 16, 16662–16687. [Google Scholar] [CrossRef]
  24. Barros, T.; Conde, P.; Gonçalves, G.; Premebida, C.; Monteiro, M.; Ferreira, C.; Nunes, U. Multispectral Vineyard Segmentation: A Deep Learning Comparison Study. Comput. Electron. Agric. 2022, 195, 106782. [Google Scholar] [CrossRef]
  25. Poblete, T.; Ortega-Farías, S.; Moreno, M.; Bardeen, M. Artificial Neuronal Network to Predict Vine Water Status Spatial Variability Using Multispectral Information Obtained from an Unmanned Aerial Vehicle (UAV). Sensors 2017, 17, 2488. [Google Scholar] [CrossRef]
  26. Jurado, J.M.; Pádua, L.; Feito, F.R.; Sousa, J.J. Automatic Grapevine Trunk Detection on UAV-Based Point Cloud. Remote Sens. 2020, 12, 3043. [Google Scholar] [CrossRef]
  27. del-Campo-Sanchez, A.; Moreno, M.; Ballesteros, R.; Hernandez-Lopez, D. Geometric Characterization of Vines from 3D Point Clouds Obtained with Laser Scanner Systems. Remote Sens. 2019, 11, 2365. [Google Scholar] [CrossRef]
  28. De Castro, A.; Jiménez-Brenes, F.M.; Torres-Sánchez, J.; Peña, J.M.; Borra-Serrano, I.; López-Granados, F. 3-D Characterization of Vineyards Using a Novel UAV Imagery-Based OBIA Procedure for Precision Viticulture Applications. Remote Sens 2018, 10, 584. [Google Scholar] [CrossRef]
  29. Jiménez-Brenes, F.M.; López-Granados, F.; Torres-Sánchez, J.; Peña, J.M.; Ramírez, P.; Castillejo-González, I.L.; de Castro, A.I. Automatic UAV-based detection of Cynodon dactylon for site specific vineyard management. PLoS ONE 2019, 14, e0218132. [Google Scholar] [CrossRef]
  30. Chedid, E.; Avia, K.; Dumas, V.; Ley, L.; Reibel, N.; Butterlin, G.; Soma, M.; Lopez-Lozano, R.; Baret, F.; Merdinoglu, D.; et al. LiDAR is Effective in Characterization Vine Growth and Detecting Associated Genetic Loci. Plant Phenomics 2023, 5, 16. [Google Scholar] [CrossRef] [PubMed]
  31. Cantürk, M.; Zabawa, L.; Pavlic, D.; Dreier, A.; Klingbeil, L.; Kuhlmann, H. UAV-based individual plant detection and geometric parameter extraction in vineyards. Front. Plant Sci. 2023, 14, 1244384. [Google Scholar] [CrossRef] [PubMed]
  32. Gavrilović, M.; Jovanović, D.; Božović, P.; Benka, P.; Govedarica, M. Vineyard Zoning and Vine Detection Using Machine Learning in Unmanned Aerial Vehicle Imagery. Remote Sens. 2024, 16, 584. [Google Scholar] [CrossRef]
  33. Dhakal, R.; Maimaitijiang, M.; Chang, J.; Caffe, M. Utilizing Spectral, Structural and Textural Features for Oat Above-Ground Biomass Using UAV-based Multispectral Data and Machine Learning. Sensors 2023, 23, 9708–9729. [Google Scholar] [CrossRef] [PubMed]
  34. Wang, X.; Zhao, Q.; Han, F.; Zhang, J.; Jiang, P. Canopy and Height Estimation of Trees in a Shelter Forest Based on Fusion of an Airborne Multispectral Image and Photogrammetric Point Cloud. J. Sens. 2021, 14, 1251–1271. [Google Scholar] [CrossRef]
  35. Matese, A.; Di Gennaro, S.F. Beyond the traditional NDVI index as a key factor to mainstream the use of UAV in precision viticulture. Sci. Rep. 2021, 11, 2711–2723. [Google Scholar] [CrossRef] [PubMed]
  36. Comba, L.; Biglia, A.; Ricauda Aimonino, D.; Tortia, C.; Mania, E.; Guidoni, S.; Gay, P. Leaf Area Index evaluation in vineyards using 3D point clouds from UAV imagery. Sci. Rep. 2020, 11, 881–896. [Google Scholar] [CrossRef]
  37. Aboutalebi, M.; Torres-Rua, A.F.; Kustas, W.P.; Nieto, H.; Coopmans, C.; McKee, M. Assessment of different methods for shadow detection in high resolution optical imagery and evaluation of shadow impact on calculation of NDVI and evapotranspiration. Irrig. Sci. 2019, 37, 407–429. [Google Scholar] [CrossRef]
  38. Giovos, R.; Tassopoulos, D.; Kalivas, D.; Lougkos, N.; Priovolou, A. Remote Sensing Vegetation Indices in Viticulture: A Critical Review. Agriculture 2021, 11, 457–477. [Google Scholar] [CrossRef]
  39. Shahi, T.B.; Xu, C.-Y.; Neupane, A.; Guo, W. Machine Learning methods for precision agriculture with UAV imagery: A review. Electron. Res. Arch. 2022, 30, 4277–4317. [Google Scholar] [CrossRef]
  40. Bellvert, J.; Girona, J.; Zarco-Tejada, P.J. Seasonal evolution of crop water stress index in grapevine varieties determined with high-resolution remote sensing thermal imagery. Irrig. Sci. 2014, 33, 81–93. [Google Scholar] [CrossRef]
  41. Santesteban, L.G.; Di Gennaro, S.F.; Herrero-Langreo, A.; Miranda, C.; Royo, J.B.; Matese, A. High-Resolution UAV based Thermal Imaging to Estimate the Instantaneous and Seasonal Variability of Plant Water Status Within A Vineyard. Agric. Water Mang. 2017, 183, 49–59. [Google Scholar] [CrossRef]
  42. Espinoza, C.Z.; Khot, L.R.; Sankaran, S.; Jacoby, P.W. High Resolution Multispectral and Thermal Remote Sensing-Based Crop Water Stress assessment in Subsurface Irrigated Grapevines. Remote Sens. 2017, 9, 961. [Google Scholar] [CrossRef]
  43. Porten, M. Validierung des Vegetativen und Generativen Wuchszustandes von Reben in variierenden Bewirtschaftungs-und Stressbelastungssystemen. Ph.D. Dissertation, Trier University, Trier, Germany, 2020. Available online: https://ubt.opus.hbz-nrw.de/frontdoor/index/index/docId/1522 (accessed on 3 November 2024).
  44. Zhang, X.; Han, L.; Dong, Y.; Shi, Y.; Huang, W.; Han, L. A deep learning-based approach for automated yellow rust disease detection from high-resolution hyperspectral UAV images. Remote Sens. 2019, 11, 1554. [Google Scholar] [CrossRef]
  45. Nevavuori, P.; Narra, N.; Lipping, T. Crop yield prediction with deep convolutional neural networks. Comput. Electron. Agric. 2019, 163, 32–43. [Google Scholar] [CrossRef]
  46. Negro, C.; Sabella, E.; Nicolì, F.; Pierro, R.; Materazzi, A.; Panattoni, A.; Aprile, A.; Nutricati, E.; Vergine, M.; Miceli, A.; et al. Biochemical Changes in Leaves of Vitis vinifera cv. Sangiovese Infected by Bois Noir Phytoplasma. Pathogens 2020, 9, 262–289. [Google Scholar] [CrossRef] [PubMed]
  47. Fischer, M.; Ashnaei, S.P. Grapevine, esca complex, and environment: The disease triangle. Pythopathologia Mediterr. 2019, 58, 17–37. [Google Scholar] [CrossRef]
  48. Wang, Y.M.; Ostendorf, B.; Pagay, V. Detecting Virus Infections in Red and White Winegrape Canopies Using Proximal Hyperspectral Sensing. Sensors 2023, 23, 2851. [Google Scholar] [CrossRef]
  49. Müller, E.; Lipps, H.-P.; Walg, O. Der Winzer. Weinbau, 4th ed.; Wollgrasweg 41; Eugen Ulmer KG: Stuttgart, Germany, 2019. [Google Scholar]
  50. van Leeuwen, C.; Destrac-Irvine, A.; Dubernet, M.; Duchêne, E.; Gowdy, M.; Marguerit, E.; Pieri, P.; Parker, A.; de Rességuier, L.; Ollat, N. An Update on the Impact of Climate Change in Viticulture and Potential Adaptations. Agronomy 2019, 9, 514–533. [Google Scholar] [CrossRef]
  51. Dougherty, P.H. The Geography of Wine: Regions, Terroir and Techniques; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  52. Gade, D.W. Tradition, territory and terroir in French viniculture. Ann. Assoc. Geogr. 2004, 94, 848–867. [Google Scholar] [CrossRef]
  53. Arnó, J.; Escolà, A.; Vallès, J.M.; Llorens, J.; Sanz, R.; Masip, J. Leaf area index estimation in vineyards using a ground based lidar scanner. Precis. Agric. 2013, 14, 290–306. [Google Scholar] [CrossRef]
  54. Caruso, G.; Tozzini, L.; Rallo, G.; Primicerio, J.; Moriondo, M.; Palai, G. Estimating biophysical and geometrical parameters of grapevine canopies (‘sangiovese’) by an unmanned aerial vehicle (uav) and vis-nir cameras. Vitis 2017, 56, 63–70. [Google Scholar] [CrossRef]
  55. Bramley, R.; Le Moigne, M.; Evain, S.; Ouzman, J.; Florin, L.; Fadaili, E.; Hinze, C.; Cerovic, Z.G. On-the-go sensing of grape berry anthocyanins during commercial harvest: Development and prospects. Aust. J. Grape Wine Res. 2011, 17, 316–326. [Google Scholar] [CrossRef]
  56. Cortell, J.M.; Sivertsen, H.K.; Kennedy, J.A.; Heymann, H. Influence of vine vigour on Pinot noir fruit composition, wine chemical analysis, and wine sensory attributes. Am. J. Enol. Vitic. 2008, 59, 1–10. [Google Scholar] [CrossRef]
  57. Gatti, M.; Squeri, C.; Garavani, A.; Vercesi, A.; Dosso, P.; Diti, I.; Poni, S. Effects of variable rate nitrogen application on cv. Barbera Perform. Veg. Growth Leaf Nutr. Status-Am. J. Enol. Vitic. 2018, 69, 196–209. [Google Scholar]
  58. Pena-Neira, A.; Duenas, M.; Duarte, A.; Hernandez, T.; Estrella, I.; Loyola, E. Effects of ripening stages and of plant vegetative vigour on the phenolic composition of grapes (*Vitis vinifera* L.) cv. Cabernet Sauvignon in the Maipo Valley (Chile). Vitis 2017, 43, 51–57. [Google Scholar]
  59. Bonilla, I.; Martinez de Toda, F.; Martinez-Casasnovas, J.A. Unexpected relationships between vine vigour and grape composition in warm climate conditions. J. Int. Sci. Vigne 2016, 49, 127–136. [Google Scholar]
  60. Bellvert, J.; Zarco-Tejada, P.J.; Girona, J.; Fereres, E. Mapping Crop Water Stress Index in A “Pinot Noir” Vineyard: Comparing Ground Measurements with Thermal Remote Sensing Imagery from Unmanned Aerial Vehicle. Precis. Vitic. 2013, 15, 361–376. [Google Scholar] [CrossRef]
  61. Panten, K.; Bramley, R.G.V. Whole-of-block experimentation for evaluating a change to canopy management intended to enhance wine quality. Aust. J. Grape Wine Res. 2012, 18, 147–157. [Google Scholar] [CrossRef]
  62. Dorin, B.; Reynolds, A.G.; Lee, H.-S.; Carrey, M.; Shemrock, A.; Shabanian, M. Detecting cool-climate Riesling vineyard variation using unmanned aerial vehicles and proximal sensors. Drone Syst. Appl. 2024, 12, 1–18. [Google Scholar] [CrossRef]
  63. Rey-Caramés, C.; Diago, M.P.; Martín, M.P.; Lobo, A.; Tardaguila, J. Using RPAS Multi-Spectral Imagery to Characterize Vigour, Leaf Development, Yield Components and Berry Composition Variability within a Vineyard. Remote Sens. 2015, 7, 14458–14481. [Google Scholar] [CrossRef]
  64. Ballesteros, R.; Intrigliolo, D.S.; Ortega, J.F.; Ramírez-Cuesta, J.M.; Buesa, I.; Moreno, M.A. Vineyard yield estimation by combining remote sensing, computer vision and artificial neural network techniques. Precis. Agric. 2020, 21, 1242–1262. [Google Scholar] [CrossRef]
  65. Romero, M.; Luo, Y.; Su, B.; Fuentes, S. Vineyard water status estimation using multispectral imagery from an UAV platform and machine learning algorithms for irrigation scheduling management. Comput. Electron. Agric. 2018, 147, 109–117. [Google Scholar] [CrossRef]
  66. Soubry, I.; Patias, P.; Tsioukas, V. Monitoring Vineyards with UAV and multisensor for the assessment of water stress and grape maturity. J. Unmanned Veh. Syst. 2017, 5, 37–50. [Google Scholar] [CrossRef]
  67. Wan, L.; Cen, H.; Zhu, J.; Zhang, J.; Zhu, Y.; Sun, D. Grain yield prediction of rice using multitemporal UAV-based RGB and multispectral images and model transfer-a case study of small farmlands in the South of China. Agric. For. Meteorol. 2020, 291, 108096. [Google Scholar] [CrossRef]
  68. Ji, Y.; Chen, Z.; Cheng, Q.; Liu, R.; Li, M.; Yan, X. Estimation of plant height and yield based on UAV imagery in faba bean (Vicia fabia L.). Plant Methods 2022, 18, 26. [Google Scholar] [CrossRef] [PubMed]
  69. Mathenge, M.; Sonneveld, B.G.J.S.; Broerse, J.E.W. Application of GIS in Agriculture in Promoting Evidence-Informed Decision Making for Improving Agriculture Sustainability: A Systematic Review. Sustainability 2022, 14, 9974–9988. [Google Scholar] [CrossRef]
  70. Mathews, A.J. Applying Geospatial Tools and Techniques to Viticulture. Geography Compass 2013, 7, 22–34. [Google Scholar] [CrossRef]
  71. Moral, F.J.; Rebollo, F.J.; Paniagua, L.L.; García-Martín, A. A GIS-based multivariate clustering for characterization and ecoregion mapping from a viticultural perspective. Span. J. Agric. Res. 2016, 14, 10. [Google Scholar] [CrossRef]
  72. Ferrer, M.; Echeverría, G.; Pereyra, G.; Neves, G.G.; Pan, D.; Mirás-Avalos, J.M. Mapping Vineyard vigour using airborne remote sensing: Relations with yield, berry composition and sanitary status under humid climate conditions. Precis. Agric. 2020, 21, 178–197. [Google Scholar] [CrossRef]
  73. Albetis, J.; Jacquin, A.; Goulard, M.; Poilvé, H.; Rousseau, J.; Clenet, H.; Dedieu, G.; Duthoit, S. On the potential of UAV Multispectral Imagery to Detect Flavescence dorée and Grapevine Trunk Diseases. Remote Sens. 2019, 11, 23. [Google Scholar] [CrossRef]
  74. Vanegas, F.; Bratanov, D.; Powell, K.; Weiss, J.; Gonzalez, F. A Novel Methodology for Improving Plant Pest Surveillance in Vineyards and Crop Using UAV-based Hyperspectral and Spatial Data. Sensors 2018, 18, 260. [Google Scholar] [CrossRef] [PubMed]
  75. Helman, D.; Bahat, I.; Netzer, Y.; Ben-Gal, A.; Alchanatis, V.; Peeters, A.; Cohen, Y. Using Time Series of High-Resolution Planet Satellite Images to Monitor Grapevine Stem Water Potential in Commercial Vineyards. Remote Sens. 2018, 10, 1615–1636. [Google Scholar] [CrossRef]
  76. Baret, F.; Guyot, G.; Major, D.J. A vegetation Index which minimizes soil brightness effects on LAI and APAR estimation. In Proceedings of the 12th Canadian Symposium on Remote Sensing Geoscience and Remote Sensing Symposium, Vancouver, BC, Canada, 10–14 July 1989; 1989; pp. 1355–1358. [Google Scholar]
  77. Borgogno-Mondino, E.; Novello, V.; Lessio, A.; de Palma, L. Describing the spatio-temporal variability of vines and soil by satelitte-based spectral indices: A case study in Apulia (South Italy). Int. J. Appl. Earth Obs. Geoinformatics 2018, 68, 42–50. [Google Scholar] [CrossRef]
  78. Heralick, R.M.; Shanmugam, K.; Dinstein, I.H. Textural features for image classification. IEEE Trans. Syst. Man Cybern. Syst. 1973, SMC-3, 610–621. [Google Scholar] [CrossRef]
  79. Stanton, J.M. Galton, Pearson and Peas: A Brief History of Linear Regression for Statistics Instructors. J. Stat. Educ. 2001, 9, 1–13. [Google Scholar] [CrossRef]
  80. Nichol, J.E.; Sarker, M.L.R. Improved biomass estimation using the texture parameters of two high-resolution optical sensors. IEEE Trans. Geosci. Remote Sens. 2010, 49, 930–948. [Google Scholar] [CrossRef]
  81. Sassu, A.; Ghiani, L.; Salvati, L.; Mercenaro, L.; Deidda, A.; Gambella, F. Integrating UAVs and Canopy Height Models in Vineyard Management: A Time-Space Approach. Remote Sens. 2022, 14, 140–143. [Google Scholar] [CrossRef]
  82. Otsu, N.A. A threshold Selection Method from Gray Level Histograms. IEEE Trans. Syst. Man Cybern 1979, 9, 62–66. [Google Scholar] [CrossRef]
  83. Junges, A.H.; Fontana, D.C.; Anzanello, R.; Bremm, C. Normalized Difference Vegetation Index obtained by ground-based remote sensing to characterize vine cycle in Rio Grande do Sul, Brazil. Ciênca Agrotecnologia 2017, 41, 543–553. [Google Scholar] [CrossRef]
  84. Whiteside, T.G.; Boggs, G.S.; Maier, S.W. Comparing object-based and pixel-based classification for mapping savannas. Int. J. Appl. Earth Obs. Geoinf. 2011, 13, 884–893. [Google Scholar] [CrossRef]
  85. Adams, R.; Bischof, L. Seeded Region Growing. IEEE Trans. Pattern Anal. Mach. Intell. 1994, 16, 641–647. [Google Scholar] [CrossRef]
  86. Lemenkova, P. Object Based Image Segmentation Algorithm of SAGA GIS for Detecting Urban Spaces in Yaoundé, Cameroon. Cent. Eur. J. Geogr. Sustain. Dev. 2020, 2, 38–51. [Google Scholar] [CrossRef]
  87. Lemenkova, P. SAGA GIS for information extraction on presence and conditions of vegetation of northern coast of Iceland bbased on nthe Landsat TM. Acta Biol. Marisiensis 2020, 3, 10–21. [Google Scholar] [CrossRef]
  88. Hossain, M.D.; Chen, D. Segmentation for Object-Based Image Analysis (OBIA): A Review of algorithms and challenges from remote sensing perspective. ISPRS J. Photogramm. Remote Sens. 2019, 150, 115–134. [Google Scholar] [CrossRef]
  89. Drǎguţ, L.; Tiede, D.; Levick, S.R. ESP: A tool to estimate scale parameter for multiresolution image segmentation of remotely sensed data. Int. J. Geogr. Inf. Sci. 2010, 24, 859–871. [Google Scholar] [CrossRef]
  90. Singla, S.; Eldawy, A. Raptor Zonal Statitics: Fully Distributed Zonal Statistics of Big Raster+ Vector Data. In Proceedings of the IEEE International Conference on Big Data (Big Data), Atlanta, GA, USA, 10–13 December 2020; IEEE: New York, NY, USA, 2020. [Google Scholar] [CrossRef]
  91. Guo, Y.; Wang, H.; Wu, Z.; Wang, S.; Sun, H.; Senthilnath, J. Modified Red Blue Vegetation Index for Chlorophyll Estimation and Yield Prediction of Maize from Visible Images Captured by UAV. Sensors 2020, 20, 5055–5070. [Google Scholar] [CrossRef] [PubMed]
  92. Naidu, G.; Zuva, T.; Sibanda, E. A Review of Evaluation Metrics in Machine Learning. Artif. Intell. Appl. Netw. Syst. 2023, 724, 15–25. [Google Scholar]
  93. Sitaula, C.; Basnet, A.; Mainali, A.; Shahi, T.B. Deep Learning based methods for sentiment analysis on Nepali COVID-19 related tweets. Comput. Intell. Neurosci. 2021, 2021, 2158184. [Google Scholar] [CrossRef] [PubMed]
  94. Rakshit, H.; Zadeh, P.B. A New Approach to Classify Drones Using a Deep Convolutional Neural Network. Drones 2024, 8, 319. [Google Scholar] [CrossRef]
  95. Miller, C.; Portlock, T.; Nyaga, D.M.; O’sullivan, J.M. A review of model evaluation metrics for machine learning in genetics and genomics. Front. Bioinform. 2024, 4, 1457619. [Google Scholar] [CrossRef] [PubMed]
  96. Ye, S.; Pontius, R.G., Jr.; Rakshit, R. A review of accuracy assessment for object-based image analysis: From per-pixel to per-polygon approaches. ISPRS J. Photogramm. Remote Sens. 2018, 141, 137–147. [Google Scholar] [CrossRef]
  97. Ez-zahouani, B.; Teodoro, A.; El Kharki, O.; Jianhua, L. Remote sensing imagery segmentation in object-based analysis: A review of methods, optimization, and quality evaluation over the past 20 years. Remote Sens. Appl. Soc. Environ. 2023, 32, 101032. [Google Scholar] [CrossRef]
  98. Michez, A.; Lejeune, P.; Bauwens, S.; Herinaina, A.A.L.; Blaise, Y.; Castro-Muἧoz, E. Mapping and monitoring of biomass and grazing in pasture with an unmanned aerial system. Remote Sens. 2019, 11, 473–476. [Google Scholar] [CrossRef]
  99. Shu, M.; Wang, Z.; Guo, W.; Qiao, H.; Fu, Y.; Guo, Y.; Wang, L.; Ma, Y.; Gu, X. Effects of Variety and Growth Stage on UAV Multispectral Estimation of Plant Nitrogen Content of Winter Wheat. Agriculture 2024, 14, 1775. [Google Scholar] [CrossRef]
  100. Gillot, J.M.; Michelin, J.; Hadjard, D.; Houot, S. An accurate method for predicting spatial variability of maize yield from UAV-based plant height estimation: A tool for monitoring agronomic field experiments. Precis. Agric. 2021, 22, 897–921. [Google Scholar] [CrossRef]
  101. Karakizi, C.; Oikonomou, M.; Karantzalos, K. Spectral discrimination and reflectance properties of various vine varieties from satellite, UAV and proximate sensors. The international Archives of the Photogrammetry. Remote Sens. Spat. Inf. Sci. 2015, 40, 31–37. [Google Scholar] [CrossRef]
  102. Maimaitijiang, M.; Sagan, V.; Sidike, P.; Hartling, S.; Esposito, F.; Fritschi, F.B. Soybean yield prediction from UAV using multimodal data fusion and deep learning. Remote Sens. 2020, 237, 111599. [Google Scholar] [CrossRef]
  103. Liu, Y.; Liu, S.; Li, J.; Guo, X.; Wang, S.; Lu, J. Estimating biomass of winter oilseed rape using vegetation indices and texture metrics derived from UAV multispectral images. Comput. Electron. Agric. 2019, 166, 105026. [Google Scholar] [CrossRef]
  104. Liu, Y.; Fan, Y.; Feng, H.; Chen, R.; Bian, M.; Ma, Y.; Yue, J.; Yang, G. Estimating potato above-ground biomass based on spectral and spatial features extracted from high-definition digital camera images. Comput. Electron. Agric. 2024, 213, 108229. [Google Scholar] [CrossRef]
  105. Zheng, H.B.; Cheng, T.; Zhou, M.; Li, D.; Yao, X.; Tian, Y.C.; Cao, W.X.; Zhu, Y. Improved estimation of rice aboveground biomass combining textural and spectral analysis of UAV-Imagery. Precis. Agric. 2019, 20, 611–629. [Google Scholar] [CrossRef]
  106. Yue, J.; Yang, G.; Tian, Q.; Feng, H.; Xu, K.; Zhou, C. Estimate of winter-wheat above-ground biomass based on UAV ultrahigh-ground-resolution image textures and vegetation indices. Photogramm. Remote Sens. 2019, 150, 226–244. [Google Scholar] [CrossRef]
  107. Yue, J.; Yang, G.; Li, C.; Li, Z.; Wang, Y.; Feng, H.; Xu, B. Estimation of Winter Wheat Above-Ground Biomass Using Unmanned Aerial Vehicle-Based Snapshot Hyperspectral Sensor and Crop Height Improved Models. Remote Sens. 2017, 9, 708–726. [Google Scholar] [CrossRef]
  108. Li, B.; Xu, X.; Zhang, L.; Han, J.; Bian, C.; Li, G.; Liu, J.; Jin, L. Above-ground biomass estimation and yield prediction in potato by using UAV-based RGB and hyperspectral imaging. ISPRS J. Photogramm. Remote Sens. 2020, 162, 161–172. [Google Scholar] [CrossRef]
  109. Wang, C.; Nie, S.; Xi, X.; Luo, S.; Sun, X. Estimating the biomass of maize with hyperspectral and LiDAR data. Remote Sens. 2016, 9, 11. [Google Scholar] [CrossRef]
  110. Zha, H.; Miao, Y.; Wang, T.; Li, Y.; Sun, J.Z.W.; Feng, Z.; Kusnierek, K. Improving unmanned aerial vehicle remote sensing-based nitrogen nutrition index prediction with machine learning. Remote Sens. 2020, 12, 215–226. [Google Scholar] [CrossRef]
  111. Grüner, E.; Wachendorf, M.; Astor, T. The potential of UAV-borne spectral and textural information for predicting aboveground biomass and N fixation in legume-grass mixture. PLoS ONE 2020, 15, e0234703. [Google Scholar] [CrossRef] [PubMed]
  112. Yoosefzadeh-Najafabadi, M.; Earl, H.J.; Tulpan, D.; Sulik, J.; Eskandari, M. Application of machine learning algorithms in plant and breeding: Predicting yield from hyperspectral reflectance in soybean. Fron. Plant Sci. 2021, 11, 2169. [Google Scholar] [CrossRef] [PubMed]
  113. Fu, Y.; Yang, G.; Wang, J.; Song, X.; Feng, H. Winter wheat biomass estimation based on spectral indices, band depth analysis and partial least regression using hyperspectral measurements. Comput. Electron. Agric. 2014, 100, 51–59. [Google Scholar] [CrossRef]
  114. Wang, F.; Yang, M.; Ma, L.; Zhang, T.; Qin, W.; Li, W.; Zhang, Y.; Sun, Z.; Wang, Z.; Li, F.; et al. Estimation of above-ground biomass of winter wheat based on consumer-grade multi-spectral UAV. Remote Sens. 2022, 14, 1271. [Google Scholar] [CrossRef]
  115. Duan, B.; Liu, Y.; Gong, Y.; Peng, Y.; Wu, X.; Zhu, R.; Fang, S. Remote estimation of rice LAI based on Fourier spectrum texture from UAV image. Plant Methods 2019, 15, 124–135. [Google Scholar] [CrossRef] [PubMed]
  116. Geipel, J.; Link, J.; Claupein, W. Combined spectral-and spatial modeling of corn yield based on aerial images and crop surface models acquired with an unmanned aircraft system. Remote Sens. 2014, 6, 10335–10355. [Google Scholar] [CrossRef]
  117. Zhou, X.; Zheng, H.; Xu, X.; He, J.; Ge, X.; Yao, X. Predicting grain yield of rice using multi-temporal vegetation indices from UAV-based multispectral and digital imagery. ISPRS J. Photogramm. 2017, 130, 246–255. [Google Scholar] [CrossRef]
  118. Kerkech, M.; Hafiane, A.; Canals, R. Deep learning approach with colorimetric spaces and vegetation indices for vine diseases detection in UAV images. Comput. Electron. Agric. 2018, 155, 237–243. [Google Scholar] [CrossRef]
  119. Prabhakara, K.; Hively, W.D.; McCarthy, G.W. Evaluating the relationship between biomass, percent groundcover and remote sensing indices across six winter cover crop fields in Maryland, United States. Int. J. Appl. Earth Obs. Geoinformatics 2015, 39, 88–102. [Google Scholar] [CrossRef]
  120. Maimaitijang, M.; Ghulam, A.; Sidike, P.; Hartling, S.; Maimaitijang, M.; Peterson, K.; Shavers, E.; Fishman, J.; Peterson, J.; Kadam, S.; et al. Unmanned Aerial System (UAS)-based phenotyping of soybean using multi-sensor data fusion and extreme learning machine. ISPRS J. Photogramm. Remote Sens. 2017, 134, 43–58. [Google Scholar] [CrossRef]
  121. Mutanga, O.; Skidmore, A.K. Narrow band vegetation indices overcome the saturation problem in biomass estimation. Remote Sens. 2004, 25, 3999–4014. [Google Scholar] [CrossRef]
  122. Heege, H.J.; Reusch, S.; Thiessen, E. Prospects and results for optical systems for site-specific on the-o control of nitrogen-top dressing in Germany. Precis. Agric. 2008, 9, 115–131. [Google Scholar] [CrossRef]
  123. Huang, X.; Zhang, L.; Wang, L. Evaluation of morphological texture features for mangrove forest mapping and species discrimination using multispectral IKONOS imagery. IEEE Geosci. Remote Sens. Lett. 2009, 6, 393–397. [Google Scholar] [CrossRef]
  124. Pádua, L.; Matese, A.; Di Gennaro, S.F.; Morais, R.; Peres, E.; Sousa, J.J. Vineyard Classification using OBIA on UAV-based RGB and multispectral data: A case study in different wine regions. Comput. Electron. 2022, 196, 106905. [Google Scholar] [CrossRef]
  125. Willkomm, M.; Bolten, A.; Bareth, G. Non-Destructive Monitoring of Rice by Hyperspectral in-Field Spectrometry and UAV-Based Remote Sensing: Case Study of Field-Grown Rice in North Rhine-Westphalia, Germany. In Proceedings of the 2016 XXIII ISPRS Congress, Prague, Czech Republic, 12–19 July 2016. [Google Scholar]
  126. Acorsi, M.G.; das Dores Abati Miranda, F.; Martello, M.; Smaniotto, D.A.; Sartor, L.R. Estimating Biomass of Black Oat Using UAV-Based RGB Imaging. Agronomy 2019, 9, 344–357. [Google Scholar] [CrossRef]
  127. Bendig, J.; Bolten, A.; Bennertz, S.; Broscheit, J.; Eichfuss, S.; Bareth, G. Estimating biomass of barley using crop surface models (CSMs) drived from UAV-based RGB imaging. Remote Sens. 2014, 6, 10395–10412. [Google Scholar] [CrossRef]
  128. Shu, M.; Mengyuan, S.; Qizhou, D.; Xiaohong, Y.; Baoguo, L.; Yuntao, M. Estimating the maize above-ground biomass by contructing the tridimensional concept model based UAV-based digital and multi-spectral images. Field Crop Res. 2022, 282, 108491. [Google Scholar] [CrossRef]
  129. Bah, M.D.; Hafiane, A.; Canals, R. Deep learning with unsupervised data labeling for weed detection in line crops in UAV images. Remote Sens. 2018, 10, 1690–1711. [Google Scholar] [CrossRef]
  130. Zhang, J.; Chang, T.; Shi, L.; Wang, W. Combining spectral-and texture features of UAV hyperspectral images for leaf nitrogen content monitoring in winter wheat. Int. J. Remote Sens. 2022, 43, 2335–2356. [Google Scholar] [CrossRef]
  131. Schuhmacher, P.; Mislimshoeva, B.; Brenning, A.; Zandler, H.; Brandt, M.; Samimi, C.; Koellner, T. Do Red Edge and Texture Attributes from High-Resolution Satellite Data Improve Wood Volume Estimation in a Semi-Arid Mountainous Region? Remote Sens. 2016, 8, 540–558. [Google Scholar] [CrossRef]
  132. Wengert, M.; Piepho, H.-P.; Astor, T.; Graß, R.; Wijesingha, J. Assesing spatial variability of barley whole crop biomass yield and leaf area index in silvooarable agroforestry systems using UAV-borne remote sensing. Remote Sens. 2021, 13, 2751–2774. [Google Scholar] [CrossRef]
  133. Colombo, R.; Bellingeri, D.; Fasolini, D.; Marino, C.M. Retrieval of leaf area index in different vegetation types using high resolution satellite data. Remote Sens. Environ. 2003, 86, 120–131. [Google Scholar] [CrossRef]
  134. Tilly, N.; Aasen, H.; Bareth, G. Fusion of plant height and vegetation indices for the estimation of barley biomass. Remote Sens. 2015, 7, 11449–11480. [Google Scholar] [CrossRef]
  135. Possoch, M.; Bieker, S.; Hoffmeister, D.; Bolten, A.; Schellberg, J.; Bareth, G. Multi-temporal crop surface models combined with the RGB vegetation index from UAV-based images for forage monitoring in grassland. Int. Arch. Photogramm. Remote Sens. Spat. Inf. Sci. 2016, 41, 991–998. [Google Scholar] [CrossRef]
  136. Kwak, G.-H.; Park, N.-W. Impact of Texture Information on Crop Classification with Machine Learning and UAV Images. Appl. Sci. 2019, 9, 643–659. [Google Scholar] [CrossRef]
  137. Viljanen, N.; Honkavaara, E.; Näsi, R.; Hakala, T.; Niemeläinen, O.; Kaivosoja, J. A novel machine learning method for estimating biomass of grass swards using photogrammetric canopy height model, images and vegetation indices captured by a drone. Agriculture 2018, 5, 70–97. [Google Scholar] [CrossRef]
  138. Näsi, R.; Viljanen, N.; Kaivosoja, J.; Alhonoja, K.; Hakala, T.; Markelin, L.; Honkavaara, E. Estimating biomass and nitrogen amount of barley and grass using UAV and aircraft based spectral and photogrammetric 3D features. Remote Sens. 2018, 10, 1082–1113. [Google Scholar] [CrossRef]
  139. Moeckel, T.; Safari, H.; Reddersen, B.; Fricke, T.; Wachendorf, M. Fusion of ultrasonic and spectral sensor data for improving the estimation of biomass in grasslands with heterogeneous sward structure. Remote Sens. 2017, 9, 98–111. [Google Scholar] [CrossRef]
  140. Lu, N.; Zhou, J.; Han, Z.; Li, D.; Cao, Q.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W.; Cheng, T. Improved estimation of aboveground biomass in wheat from RGB imagery and point cloud data acquired with low-cost unmanned aerial vehicle system. Plant Methods 2019, 15, 17. [Google Scholar] [CrossRef] [PubMed]
  141. Adugna, T.; Xu, W.; Fan, J. Comparison of Random Forest and Support Vector Machine Classifiers for Regional Land Cover Mapping Using Coarse Resolution FY-3C Images. Remote Sens. 2022, 14, 574–595. [Google Scholar] [CrossRef]
  142. Sheykhousma, M.; Mahdianpari, M.; Mohammadimanesh, F.; Ghamisi, P.; Homayouni, S. Support Vector Machine Vs Random Forest for Remote Sensing Image Classification: A Meta-Analysis and Systematic Review. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 6308–6325. [Google Scholar]
  143. Talukdar, S.; Singha, P.; Mahato, S.; Shahfahad; Pal, S.; Liou, Y.-A.; Rahman, A. Land-Use Land-Cover Classification by Machine Learning Classifiers for Satellite Observations-A Review. Remote Sens. 2020, 12, 1135–1158. [Google Scholar] [CrossRef]
  144. Shao, G.; Tang, L.; Zhang, H. Introducing Image Classification Efficacies. IEEE Access 2016, 4, 134809–134816. [Google Scholar] [CrossRef]
  145. Zhen, Z.; Chen, S.; Yin, T.; Chavanon, E.; Lauret, N.; Guilleux, J.; Henke, M.; Qin, W.; Cao, L.; Li, J.; et al. Using the Negative Soil Adjustment Factor of Soil Adjusted Vegetation Index (SAVI) to Resist Saturation Effects and Estimate Leaf Area Index (LAI) in Dense Vegetation Areas. Sensors 2021, 21, 2115–2128. [Google Scholar] [CrossRef] [PubMed]
  146. Keshtkar, H.; Voigt, W.; Alizadeh, E. Land-cover classification and analysis of change using machine-learning classifiers and multi temporal remote sensing imagery. Arab. J. Geosci. 2017, 10, 154–168. [Google Scholar] [CrossRef]
  147. Abdi, A.M. Land cover and land use classification performance of machine learning algorithms in a boreal landscape using Sentinel-2 data. GIScience Remote Sens. 2020, 57, 1–20. [Google Scholar] [CrossRef]
Figure 1. (A) Investigation area in Bernkastel-Kues within the Moselle wine region mapped on a high-precision orthomosaic (CRS (Coordinate Reference System) with EPSG (European Petroleum Survey Group) 25,832 ETRS (European Terrestrial Reference System) 89/UTM (Universal Transverse Mercator) zone 32N). Red points represent each vine position that was localized with differential-GPS (see Section 2.5 for more information) (B) Zoomed out view of the investigation area and overview of local vineyard structure and the Moselle river (mapped on Google Earth Satellite map from QuickMapService plugin in QGIS version 3.22) (C) Intermediate zoom of the investigation area, with orthomosaic on Google Earth Satellite map. It can be seen that the UAV- sensor- based orthomosaic and the Google Satellite map show some offset to each other, due to different absolute geographic accuracies and spatial resolution.
Figure 1. (A) Investigation area in Bernkastel-Kues within the Moselle wine region mapped on a high-precision orthomosaic (CRS (Coordinate Reference System) with EPSG (European Petroleum Survey Group) 25,832 ETRS (European Terrestrial Reference System) 89/UTM (Universal Transverse Mercator) zone 32N). Red points represent each vine position that was localized with differential-GPS (see Section 2.5 for more information) (B) Zoomed out view of the investigation area and overview of local vineyard structure and the Moselle river (mapped on Google Earth Satellite map from QuickMapService plugin in QGIS version 3.22) (C) Intermediate zoom of the investigation area, with orthomosaic on Google Earth Satellite map. It can be seen that the UAV- sensor- based orthomosaic and the Google Satellite map show some offset to each other, due to different absolute geographic accuracies and spatial resolution.
Sensors 25 00431 g001
Figure 2. (A) Zoomed-out view of the canopy-free vine rows of the investigation area. (B) Zoomed-in view of the training system in the investigation area (photos taken in December 2024).
Figure 2. (A) Zoomed-out view of the canopy-free vine rows of the investigation area. (B) Zoomed-in view of the training system in the investigation area (photos taken in December 2024).
Sensors 25 00431 g002
Figure 3. Ground truth template examples with label descriptions for the growth classes after Porten [43]. The specific visual characteristics and correlations to viticultural, oenological, and environmental parameters are described in detail by [43].
Figure 3. Ground truth template examples with label descriptions for the growth classes after Porten [43]. The specific visual characteristics and correlations to viticultural, oenological, and environmental parameters are described in detail by [43].
Sensors 25 00431 g003
Figure 4. Color-coded growth classification after Porten [43] for single grapevines in the investigation area mapped on multispectral orthomosaic. All geodata are projected to CRS with EPSG: 25,832 ETRS89/UTM zone 32N.
Figure 4. Color-coded growth classification after Porten [43] for single grapevines in the investigation area mapped on multispectral orthomosaic. All geodata are projected to CRS with EPSG: 25,832 ETRS89/UTM zone 32N.
Sensors 25 00431 g004
Figure 5. Visualization of the developed and applied geo- and image processing workflow in this study, with QGIS and different geospatial libraries in Phyton. Geoprocessing was the foundation for further statistical analysis, and machine learning model predictions of the growth classes after [43].
Figure 5. Visualization of the developed and applied geo- and image processing workflow in this study, with QGIS and different geospatial libraries in Phyton. Geoprocessing was the foundation for further statistical analysis, and machine learning model predictions of the growth classes after [43].
Sensors 25 00431 g005
Figure 6. Sampling rectangles around the vines’ position for zonal statistics pixel aggregation process, together with growth class categorized grapevine stem positions and vine row extracted OSAVI (OSAVI extracted). All geodata are projected to CRS with EPSG: 25,832 ETRS89/UTM zone 32N.
Figure 6. Sampling rectangles around the vines’ position for zonal statistics pixel aggregation process, together with growth class categorized grapevine stem positions and vine row extracted OSAVI (OSAVI extracted). All geodata are projected to CRS with EPSG: 25,832 ETRS89/UTM zone 32N.
Sensors 25 00431 g006
Figure 7. Growth class grouped CHM Volume boxplots with significance stars (*) between boxplots generated according to the Mann–Whitney-U-test with p-value significance. ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance).
Figure 7. Growth class grouped CHM Volume boxplots with significance stars (*) between boxplots generated according to the Mann–Whitney-U-test with p-value significance. ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance).
Sensors 25 00431 g007
Figure 8. Input feature group (1–7) grouped boxplot OA (overall accuracy) in % for the SVM classifier of the seven different SVM models (see legend color of grouped boxplots), with significance stars (*) generated according to the Mann–Whitney-U-test between statistical significant model (SVM 1–SVM 7) results, where significant accuracy differences, derived from repeated-k-fold-cross-validation occurred, with p-value significance classes: ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Figure 8. Input feature group (1–7) grouped boxplot OA (overall accuracy) in % for the SVM classifier of the seven different SVM models (see legend color of grouped boxplots), with significance stars (*) generated according to the Mann–Whitney-U-test between statistical significant model (SVM 1–SVM 7) results, where significant accuracy differences, derived from repeated-k-fold-cross-validation occurred, with p-value significance classes: ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Sensors 25 00431 g008
Figure 9. Input feature group (1–7) grouped boxplot accuracy in % for the RF classifier of the seven different RF classifier models (see legend color of grouped boxplots), with significance stars (*) generated according to the Mann–Whitney-U-test between statistical significant model (RF 1–RF 7) results, where significant OA (overall accuracy) differences derived from repeated-k-fold-cross-validation occurred, with p-value significance classes: * Signal greater than 0.1 (weak significance).** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Figure 9. Input feature group (1–7) grouped boxplot accuracy in % for the RF classifier of the seven different RF classifier models (see legend color of grouped boxplots), with significance stars (*) generated according to the Mann–Whitney-U-test between statistical significant model (RF 1–RF 7) results, where significant OA (overall accuracy) differences derived from repeated-k-fold-cross-validation occurred, with p-value significance classes: * Signal greater than 0.1 (weak significance).** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Sensors 25 00431 g009
Figure 10. Pairwise statistical comparison of OA (overall accuracy) of the test and train data in % for the SVM classifier of the seven different feature groups input sets (1–7) with significance stars (*) generated according to the Mann–Whitney-U-test between boxplots where significant accuracy differences (OA) derived from repeated-k-fold-cross-validation occurred, with p-value significance classes. * Signal greater than 0.1 (weak significance). *** Signal less than 0.001 (vital significance).
Figure 10. Pairwise statistical comparison of OA (overall accuracy) of the test and train data in % for the SVM classifier of the seven different feature groups input sets (1–7) with significance stars (*) generated according to the Mann–Whitney-U-test between boxplots where significant accuracy differences (OA) derived from repeated-k-fold-cross-validation occurred, with p-value significance classes. * Signal greater than 0.1 (weak significance). *** Signal less than 0.001 (vital significance).
Sensors 25 00431 g010
Figure 11. Pairwise statistical comparison of accuracy in % and f1-weighted score in % of the test data sets for the RF classifier of the seven different input feature groups (see legend color of grouped boxplots) with significance stars (*) generated according to the Mann–Whitney-U-test between accuracy and f1-weighted, with significant differences derived from repeated-k-fold-cross-validation occurred, with p-value significance classes. *** Signal less than 0.001 (vital significance).
Figure 11. Pairwise statistical comparison of accuracy in % and f1-weighted score in % of the test data sets for the RF classifier of the seven different input feature groups (see legend color of grouped boxplots) with significance stars (*) generated according to the Mann–Whitney-U-test between accuracy and f1-weighted, with significant differences derived from repeated-k-fold-cross-validation occurred, with p-value significance classes. *** Signal less than 0.001 (vital significance).
Sensors 25 00431 g011
Figure 12. Pairwise statistical comparison of overall accuracy of train data set in % for the SVM classifier with f1-weighted in score in % of the seven different input feature groups (see legend color of grouped boxplots) with significance stars (*) generated according to the Mann–Whitney-U-test between accuracy and f1-weighted, where significant accuracy differences derived from repeated-k-fold cross- validation occurred, with p-value significance classes. ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Figure 12. Pairwise statistical comparison of overall accuracy of train data set in % for the SVM classifier with f1-weighted in score in % of the seven different input feature groups (see legend color of grouped boxplots) with significance stars (*) generated according to the Mann–Whitney-U-test between accuracy and f1-weighted, where significant accuracy differences derived from repeated-k-fold cross- validation occurred, with p-value significance classes. ** Signal greater than 0.01 (intermediate significance). *** Signal less than 0.001 (vital significance). No stars between boxplots indicate no statistical differences between the model outputs according to Mann-Whitney-U-test.
Sensors 25 00431 g012
Figure 13. Visualization example of the difference between the ground truth growth classes and the model predicted growth classes from the output from SVM 7 model. Red numbers next to grapevine stems (brown points) with values over zero represent an underestimation of the model prediction compared to ground truth data. In contrast, values less than zero would indicate an overestimation of the growth class model prediction, compared to the ground truth data. Zero values indicate a perfect match of the ground truth with the ML model prediction. The red rectangles represent the area of the zonal statistics aggregation, and the red outline the generated vine row mask (see Section 2.6.9), where the spatial aggregation of the features was achieved. Pixels outside the vine row mask were not considered for spatial aggregation. All mapped geodata are projected to CRS with EPSG: 25,832/ ETRS89/UTM zone 32N.
Figure 13. Visualization example of the difference between the ground truth growth classes and the model predicted growth classes from the output from SVM 7 model. Red numbers next to grapevine stems (brown points) with values over zero represent an underestimation of the model prediction compared to ground truth data. In contrast, values less than zero would indicate an overestimation of the growth class model prediction, compared to the ground truth data. Zero values indicate a perfect match of the ground truth with the ML model prediction. The red rectangles represent the area of the zonal statistics aggregation, and the red outline the generated vine row mask (see Section 2.6.9), where the spatial aggregation of the features was achieved. Pixels outside the vine row mask were not considered for spatial aggregation. All mapped geodata are projected to CRS with EPSG: 25,832/ ETRS89/UTM zone 32N.
Sensors 25 00431 g013
Table 1. Growth and infection classification assessment after Porten [43] with growth classes as integer-coded values and also color-coded for individual growth classes as well as disease/infection type with additive (Additive column) code to describe the infection type; in the description column in parenthesis, the range of the infection strength (INS column) from light (0.5) to severe (4.5) is coded, including the intermediate infection classes 1, 1.5, 2, 2.5, 3, 3.5, and 4. The virus infection class, listed later in some tables, is not listed here due to challenges in visually recognizing virus symptoms with appropriate precision. Within this study, only the growth class of each vine, without the infection description, was used as the target classes for later machine learning model predictions.
Table 1. Growth and infection classification assessment after Porten [43] with growth classes as integer-coded values and also color-coded for individual growth classes as well as disease/infection type with additive (Additive column) code to describe the infection type; in the description column in parenthesis, the range of the infection strength (INS column) from light (0.5) to severe (4.5) is coded, including the intermediate infection classes 1, 1.5, 2, 2.5, 3, 3.5, and 4. The virus infection class, listed later in some tables, is not listed here due to challenges in visually recognizing virus symptoms with appropriate precision. Within this study, only the growth class of each vine, without the infection description, was used as the target classes for later machine learning model predictions.
Growth ClassGrowth DescriptionAdditiveDescriptionINS
APO/0Apoplexy/Death/Missing stemEEsca symptoms 0.5-4.5
1Almost deadSBlackwood disease (Bois Noir) 0.5–4.5
3Very weak growthESEsca and Bois Noir 0.5–4.5
5Medium growthPInfection with Peronospora0.5–4.5
7Medium to good growthNNew planted vine/vines 0.5–4.5
9Excellent-to-excessive growthNARe-expulled from the trunk after APO or other damage 0.5–4.5
Color-Code description: 0—Apoplexy/Death/Missing stem—dark red| 1—Almost dead—red| 3—Very weak growth—orange| 5—Medium growth—cyan, | 7—Medium to good growth—green| 9—Excellent-to-excessive growth)dark green.
Table 2. Spectral properties (spectral band, center wavelength in nanometer, and bandwidth at full width at half maximum (FWHM)) of the different spectral bands of the MicaSenseRed-Edge-MXTM. The numbers in brackets after the spectral band name indicate the band number of the generated orthomosaics of this study.
Table 2. Spectral properties (spectral band, center wavelength in nanometer, and bandwidth at full width at half maximum (FWHM)) of the different spectral bands of the MicaSenseRed-Edge-MXTM. The numbers in brackets after the spectral band name indicate the band number of the generated orthomosaics of this study.
Spectral BandCenter Wavelength (nm)Bandwidth FWHM (nm)
Blue (1)47532
Green (2)56027
Red (3)66816
Red Edge (4)71712
Near Infrared (5)84257
Table 3. Calculated VI with relevant literature references for vine row extraction (OSAVI), and as input features for the ML-based prediction of the growth classes after [43].
Table 3. Calculated VI with relevant literature references for vine row extraction (OSAVI), and as input features for the ML-based prediction of the growth classes after [43].
ID Spectral IndicesSource ReferenceAbbreviation
1Normalized Difference Vegetation Index[72]NDVI
2Normalized Difference Red Edge Index[73]NDREI
3Optimized Soil Adjusted Vegetation Index[74]OSAVI
4Green Normalized Difference Vegetation Index[75]GNDVI
5Transformed Soil Adjusted Vegetation Index[76]TSAVI
6Normalized Difference Water Index[77]NDWI
Table 4. Utilized structural features used as input features for the ML- based prediction of the growth classes after [43].
Table 4. Utilized structural features used as input features for the ML- based prediction of the growth classes after [43].
ID F.I.Height and Volume MeasuresName
1Mean HeightCHMmean
2Median Height CHMmedian
3Minimum HeightCHMmin
4Maximum HeightCHMmax
5Standard Deviation HeightCHMstd
6VarianceCHMvar
7Aggregated Pixel VolumeCHM Volume
Table 5. The Gray-Level Co-occurrence Matrix (GLCM) textural features and their definitions, used as input features for ML-based prediction of the growth classes after [43].
Table 5. The Gray-Level Co-occurrence Matrix (GLCM) textural features and their definitions, used as input features for ML-based prediction of the growth classes after [43].
S.N.Texture MeasureFormula
1.Mean (ME) M E = x = 0 N 1 z = 0 N 1 k P ( x , z )
2.Variance (VAR) V A R = x = 0 N 1 z = 0 N 1 x µ 2 P ( x , z )
3.Homogeneity (HOM) H O M = x = 0 N 1 z = 0 N 1 1 1 + x + z 2 P ( x , z )
4.Contrast (CON) C O N = x = 0 N 1 z = 0 N 1 x z 2 P ( x , z )
5.Dissimilarity (DIS) D I S = x = 0 N 1 z = 0 N 1 P x , z | x z |
6.Entropy (ENT) E N T = x = 0 N 1 z = 0 N 1 P x , z l o g ( P x , z )
7.Angular Second Moment (ASM) A S M = x = 0 N 1 z = 0 N 1 ( P x , z ) 2
8.Correlation (COR) C O R = x = 0 N 1 z = 0 N 1 P ( x , z ) ( x M E ) ( z M E ) V A x V A z
N o t e : P x , z = V x , z i = 0 N 1 j = 0 N 1 V ; w h e r e   V A x , z   r e p r e s e n t s the value in the row at the cell x and column z within the moving window. N represents the number of rows or columns in the window.
Table 6. Overview table of the input feature groups (1–7) built from the input features of the different feature types (spectral, structural, texture) and their combinations.
Table 6. Overview table of the input feature groups (1–7) built from the input features of the different feature types (spectral, structural, texture) and their combinations.
ID Input Feature GroupAbbreviation Input Features
1. spectral features fitted to growth class ground truthspn = 6
NDVI + NDVIRE + OSAVI + GNDVI + NDWI + TSAVI
2. structural features fitted to growth class ground truthstrn = 4
CHM(mean) + CHM(max) + CHM(std) + CHM(Volume)
3. texture features fitted to growth class ground truthtexn = 4
Contrast + Correlation + Entropy + Angular Second Moment
4. spectral and structural features fitted to growth class ground truthsp + str/sp-strn = 10
NDVI + NDVIRE + OSAVI + TSAVI + GNDVI + NDWI+ CHM(mean) + CHM(max) + CHM(std) + CHM(Volume)
5. spectral and texture features fitted to growth class ground truthsp + tex/sp-texn = 10
NDVI + NDVIRE + OSAVI + GNDVI + NDWI + TSAVI + Contrast + Correlation + Entropy+ Angular Second Moment
6. structural and texture features fitted to growth class ground truthstr + tex/str-texn = 8
CHM(mean) + CHM(max) + CHM(std) + CHM(Volume) + Contrast + Correlation + Entropy + Angular Second Moment
7. spectral, structural, and texture features fitted to growth class ground truthsp + str + tex/sp-str-texn = 14
NDVI + NDVIRE + OSAVI + GNDVI + NDWI + TSAVI + CHM(mean) + CHM(max) + CHM(std) + CHM(Volume) + Contrast + Correlation + Entropy + Angular Second Moment
Table 7. ML classifiers SVM and RFC, with preprocessing steps and selected parameters, parameter settings and parameter ranges for hyperparameter-tuning achieved through grid-search in sklearn, listed for the different classifiers.
Table 7. ML classifiers SVM and RFC, with preprocessing steps and selected parameters, parameter settings and parameter ranges for hyperparameter-tuning achieved through grid-search in sklearn, listed for the different classifiers.
Machine Learning ClassifiersPre-Processing of Input FeaturesTrain–Test SplitParameter for Hyperparameter Tuning
SVM (Support Vector Machines)- selection of features
- eliminating NoData values
- standard scaling of the input features
repeated-k-fold-cross-validation splits = 5, repeats = 5 (n = 25), random state = 1, jobs = −1 with hyperparameter
tuning (grid search)
kernel: [linear, rbf];
c: 0.001, 0.01, 0.1, 1, 10, 15, 20, 100, 1000]:
gamma: [0.001, 0.01, 0.1, 1, 10]
RFC (Random Forest Classifier)- selection of features
- eliminating NoData values
- standard scaling of the input features
repeated-k-fold-cross-validation splits = 5, repeats = 5 (n = 25), random state = 1, jobs = −1 with hyperparameter
tuning (grid search)
n_estimators: [25, 50, 100, 150, 300, 500],
max_features: [sqrt, log2, none],
max_depth: [3, 6, 9, 15, 20, 30],
max_leaf nodes: [3, 6, 9]
max_samples: [2, 4, 6]
min_samples_leaf: [1, 2, 4]
criterion: [entropy, gini]
Table 8. Results and comparison of model score metrics (accuracymean, f1-weighted mean, accuracystd, f1-weightedstd,) of training and test data fits, received from repeated k-fold cross-validation (n = 25) for the different ML classifiers (SVM, RFC) with grid search-based hyperparameter tuning with repeated-k-fold-cross-validation on the different input feature groups (1–7). For the three best models (SVM4, SVM6, SVM7) according to accuracy of the test data set, the model metrics results are marked with bold green, whereas for the three worst models (RF1, RF3, RF5) according to accuracy of the test data set the metrics are marked in bold red. In the model type column, models trained with only structural features (2) or together with other feature types (4,6,7) are marked in bold gray, to highlight the strong influence of structural features on model classification performance.
Table 8. Results and comparison of model score metrics (accuracymean, f1-weighted mean, accuracystd, f1-weightedstd,) of training and test data fits, received from repeated k-fold cross-validation (n = 25) for the different ML classifiers (SVM, RFC) with grid search-based hyperparameter tuning with repeated-k-fold-cross-validation on the different input feature groups (1–7). For the three best models (SVM4, SVM6, SVM7) according to accuracy of the test data set, the model metrics results are marked with bold green, whereas for the three worst models (RF1, RF3, RF5) according to accuracy of the test data set the metrics are marked in bold red. In the model type column, models trained with only structural features (2) or together with other feature types (4,6,7) are marked in bold gray, to highlight the strong influence of structural features on model classification performance.
Train DataTest Data
Model TypeAccuracyF1-WeightedAccuracy std. F1-Weighted std. AccuracyF1-WeightedAccuracy std. F1-Weighted std.
RF 131.51%22.25%2%4.40%30.41%20.98%2.79%4.93%
RF 241.58%32.66%3%5.29%40.04%31.57%3.33%5.59%
RF 334.10%29.10%2%5.31%33.17%28.32%2.84%6.90%
RF 437.38%34.11%4%5.51%37.38%33.05%3.99%6.27%
RF 535.07%27.20%3%3.24%33.23%26.79%4.23%4.38%
RF 639.96%27.73%3%5.14%38.86%26.65%3.65%4.82%
RF 738.00%31.63%2%4.53%36.77%30.01%4.37%6.30%
SVM 150.48%46.76%1%4.89%40.65%37.48%3.16%3.40%
SVM 250.34%51.23%2%2.41%47.38%46.21%1.89%2.76%
SVM 340.13%52.49%2%2.64%37.98%47.72%2.56%2.90%
SVM 451.75%49.63%2%1.63%47.42%45.86%2.78%3.12%
SVM 550.13%35.68%3%1.67%44.21%33.59%1.92%2.44%
SVM 651.05%47.38%2%3.47%47.57%42.39%2.69%2.97%
SVM 753.66%50.11%2%2.27%48.51%45.50%2.22%3.16%
Table 9. Confusion matrix of the test data set for the best overall accuracy prediction score (SVM 7) with 48.51% test accuracy and 53.7% train accuracy from repeated- k-fold cross-validation with tuned hyperparameter. The color marked cells represent the colors according to the color-coding after [43], including neighbor cells of each growth class (to improve the contrast of the values in these cells these are marked in bold to show the percentage of overlapping neighbor growth classes). Letter P in the horizontal bar of the growth classes (GC) stands for the predicted value view (GP), whereas letter L in the vertical bar stands for the labeled growth class view (GL) for each growth class (0-9). Additionally, the UA (User Accuracy) and PA (Producer Accuracy) were calculated for interpretation of classification results and model evaluation (see following text in discussion). The column, and row headers as well as the values in the PA row and the UA column are marked in bold to receive better contrast for reading, and to show their importance for interpreting the class specific accuracies.
Table 9. Confusion matrix of the test data set for the best overall accuracy prediction score (SVM 7) with 48.51% test accuracy and 53.7% train accuracy from repeated- k-fold cross-validation with tuned hyperparameter. The color marked cells represent the colors according to the color-coding after [43], including neighbor cells of each growth class (to improve the contrast of the values in these cells these are marked in bold to show the percentage of overlapping neighbor growth classes). Letter P in the horizontal bar of the growth classes (GC) stands for the predicted value view (GP), whereas letter L in the vertical bar stands for the labeled growth class view (GL) for each growth class (0-9). Additionally, the UA (User Accuracy) and PA (Producer Accuracy) were calculated for interpretation of classification results and model evaluation (see following text in discussion). The column, and row headers as well as the values in the PA row and the UA column are marked in bold to receive better contrast for reading, and to show their importance for interpreting the class specific accuracies.
GC0 GP1 GP3 GP5 GP7 GP9 GPUA
0 GL29.2834.2527.626.632.210.0029.28
1 GL14.7736.2929.1115.614.220.0036.29
3 GL3.6910.8235.8844.595.010.0035.88
5 GL0.001.0816.8165.7316.160.2265.73
7 GL0.300.302.7239.5857.100.0057.10
9 GL0.000.000.009.0981.829.099.09
PA60.9543.8632.0036.2734.2918.92SVM 7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dillner, R.P.; Wimmer, M.A.; Porten, M.; Udelhoven, T.; Retzlaff, R. Combining a Standardized Growth Class Assessment, UAV Sensor Data, GIS Processing, and Machine Learning Classification to Derive a Correlation with the Vigour and Canopy Volume of Grapevines. Sensors 2025, 25, 431. https://doi.org/10.3390/s25020431

AMA Style

Dillner RP, Wimmer MA, Porten M, Udelhoven T, Retzlaff R. Combining a Standardized Growth Class Assessment, UAV Sensor Data, GIS Processing, and Machine Learning Classification to Derive a Correlation with the Vigour and Canopy Volume of Grapevines. Sensors. 2025; 25(2):431. https://doi.org/10.3390/s25020431

Chicago/Turabian Style

Dillner, Ronald P., Maria A. Wimmer, Matthias Porten, Thomas Udelhoven, and Rebecca Retzlaff. 2025. "Combining a Standardized Growth Class Assessment, UAV Sensor Data, GIS Processing, and Machine Learning Classification to Derive a Correlation with the Vigour and Canopy Volume of Grapevines" Sensors 25, no. 2: 431. https://doi.org/10.3390/s25020431

APA Style

Dillner, R. P., Wimmer, M. A., Porten, M., Udelhoven, T., & Retzlaff, R. (2025). Combining a Standardized Growth Class Assessment, UAV Sensor Data, GIS Processing, and Machine Learning Classification to Derive a Correlation with the Vigour and Canopy Volume of Grapevines. Sensors, 25(2), 431. https://doi.org/10.3390/s25020431

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop