Next Article in Journal
Oilseed Radish: Nitrogen and Sulfur Management Strategies for Seed Yield and Quality—A Case Study in Poland
Previous Article in Journal
Characterizing Spatial and Temporal Variations in N2O Emissions from Dairy Manure Management in China Based on IPCC Methodology
Previous Article in Special Issue
An Accurate Approach for Predicting Soil Quality Based on Machine Learning in Drylands
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Prediction of Live Bulb Weight for Field Vegetables Using Functional Regression Models and Machine Learning Methods

1
Department of Mathematics and Statistics, Chonnam National University, Gwangju 61186, Republic of Korea
2
Department of Statistics, Chonnam National University, Gwangju 61186, Republic of Korea
3
Division of Culture Contents, Chonnam National University, Yeosu 59626, Republic of Korea
*
Authors to whom correspondence should be addressed.
Agriculture 2024, 14(5), 754; https://doi.org/10.3390/agriculture14050754
Submission received: 22 February 2024 / Revised: 6 May 2024 / Accepted: 10 May 2024 / Published: 12 May 2024
(This article belongs to the Special Issue Applications of Data Analysis in Agriculture—2nd Edition)

Abstract

:
(1) Background: This challenge is exacerbated by the aging of the rural population, leading to a scarcity of available manpower. To address this issue, the automation and mechanization of outdoor vegetable cultivation are imperative. Therefore, developing an automated cultivation platform that reduces labor requirements and improves yield by efficiently performing all the cultivation activities related to field vegetables, particularly onions and garlic, is essential. In this study, we propose methods to identify onion and garlic plants with the best growth status and accurately predict their live bulb weight by regularly photographing their growth status using a multispectral camera mounted on a drone. (2) Methods: This study was conducted in four stages. First, two pilot blocks with a total of 16 experimental units, four horizontals, and four verticals were installed for both onions and garlic. Overall, a total of 32 experimental units were prepared for both onion and garlic. Second, multispectral image data were collected using a multispectral camera repeating a total of seven times for each area in 32 experimental units prepared for both onions and garlic. Simultaneously, growth data and live bulb weight at the corresponding points were recorded manually. Third, correlation analysis was conducted to determine the relationship between various vegetation indexes extracted from multispectral images and the manually measured growth data and live bulb weights. Fourth, based on the vegetation indexes extracted from multispectral images and previously collected growth data, a method to predict the live bulb weight of onions and garlic in real time during the cultivation period, using functional regression models and machine learning methods, was examined. (3) Results: The experimental results revealed that the Functional Concurrence Regression (FCR) model exhibited the most robust prediction performance both when using growth factors and when using vegetation indexes. Following closely, with a slight distinction, Gaussian Process Functional Data Analysis (GPFDA), Random Forest Regression (RFR), and AdaBoost demonstrated the next-best predictive power. However, a Support Vector Machine (SVM) and Deep Neural Network (DNN) displayed comparatively poorer predictive power. Notably, when employing growth factors as explanatory variables, all prediction models exhibited a slightly improved performance compared to that when using vegetation indexes. (4) Discussion: This study explores predicting onion and garlic bulb weights in real-time using multispectral imaging and machine learning, filling a gap in research where previous studies primarily focused on utilizing artificial intelligence and machine learning for productivity enhancement, disease management, and crop monitoring. (5) Conclusions: In this study, we developed an automated method to predict the growth trajectory of onion and garlic bulb weights throughout the growing season by utilizing multispectral images, growth factors, and live bulb weight data, revealing that the FCR model demonstrated the most robust predictive performance among six artificial intelligence models tested.

1. Introduction

Agriculture plays a pivotal role in sustaining the human diet economy. The global population is growing tremendously, leading to a substantial increase in the demand for and supply of food. However, conventional farming methods prove insufficient to meet these escalating requirements. Consequently, there is a pressing need for innovative automated agricultural techniques, and smart agriculture, through the application of artificial intelligence (AI), is emerging as a key solution to the challenges that we face today. Smart agriculture, harnessing the power of AI technologies, such as machine learning, deep learning, and statistical analysis, is crucial in addressing the complexities of modern farming. Various aspects of agricultural processes, including crop selection, crop yield prediction, soil suitability classification, and water management, benefit from these advanced technologies. Machine learning (ML) algorithms contribute to crop selection and management, deep learning techniques enable crop selection and crop production forecasting, and time-series analysis aids in forecasting crop demand, product price, and crop yield. Additionally, ML and deep learning algorithms could be integrated to develop smart agricultural technologies for various tasks, including soil assessment and soil suitability classification. Therefore, the adoption of cutting-edge technologies within AI in agricultural fields holds the potential to assist farmers in producing high-quality crops. The development of the agricultural sector will also contribute considerably to rural development. Within the agricultural sector, diverse technologies, including disease detection, diagnosis, and soil-specific fertilizer recommendations, can be leveraged to improve agricultural practices, such as crop production, real-time monitoring, and harvesting.
Here, we will consider the pivotal role that AI and ML technologies play in revolutionizing smart agriculture in previous research published so far. First, we will examine review papers that describe how artificial intelligence technology is applied to the growth status of field crops, the effects of nitrogen and irrigation, and yield prediction. Eli-Chukwu [1] outlines AI’s applications in soil, crop, weed, and disease management, stressing both their potentials and constraints alongside expert system integration for enhanced productivity. Ayed and Hanana [2] showcase AI and ML’s efficacy across various agricultural supply chain sectors, indicating a clear upward trend in their adoption to enhance the food industries. Noteworthy contributions by Jagli et al. [3] introduce AI applications such as sensor, robot, and drone utilization for irrigation, weeding, and spraying, collectively aiding in resource conservation, soil fertility preservation, and improved crop productivity. Hossen et al. [4] advocate for AI-driven automation in soil and disease management, crop monitoring, and weed control, highlighting their potential to alleviate agricultural challenges and reduce labor-intensive tasks. Akkem et al. [5] review ML and deep learning’s relevance in agriculture, emphasizing their role in soil fertility assessment and crop selection, including time-series analysis and prediction. Oliveira and Silva [6] document the prevalent use of AI technologies, including ML, convolutional neural networks, and IoT, in agriculture, while also outlining the future directions and challenges. Sharma et al. [7] present a systematic review of ML applications in agriculture in their review paper. They focused on a prediction of soil parameters such as organic carbon and moisture content, crop yield prediction, and disease and weed detection in crops and species detection. They also examine ML using computer vision to classify diverse sets of crop images to monitor crop quality and yield assessment.
Second, we will review papers that describe the methods for predicting the raw weight or yield of onions or garlic using statistical models or machine learning techniques. Jeong et al. [8] prepared experimental plots where garlic and onions were grown according to various planting times and fertilizer rates. Then, they collected RGB and multispectral UAV images of onions and garlic at a spatial resolution of less than 1 cm to establish the correlation between the UAV images of garlic and onions and various biophysical parameters. In addition, they estimated the fresh weight of garlic and onions using two spectral indices, vegetation fraction and 3D-based height estimation obtained using an RGB camera, and studied the nutritional status of crops based on NDVI calculated using multispectral images. Lee et al. [9] assessed the climate change from 1973 to 2017 in Gyeongsangnam-do, South Korea, and analyzed the relationship between weather conditions and onion bulb yield from the 1991/1992 to 2016/2017 growing seasons. Through linear regression analysis, they confirmed that temperature and precipitation were positively correlated with onion bulb yield. Przygocka-Cyna et al. [10] verified the seasonal trend of onion biomass and how its yield varies depending on the dynamics of nitrogen (N) and sulfur (S) uptake based on three years of field surveys. These experimental conditions consisted of three levels of nitrogen and three levels of sulfur absorption, and the total dry weight (TDW), total N uptake (TNU), and total S uptake (TSU) of onions were measured at 10-day intervals. Through experiments, they confirmed that TDW and TNU can be well explained by using an exponential linear model and TSU by using a quadratic growth model during the season. Desta et al. [11] conducted an experiment to evaluate the effects of three harvest stages (60, 80, and 100% top fall), two curing levels (non-cured and cured bulbs), and three storage methods (floor, shelf, and net bag) on the storability of the garlic variety ‘Tseday’ during 2014-2015. They conducted the experiment in a randomized complete block design repeated three times at both sites. The results of their study showed that harvesting at a peak drop rate of 80%, and curing and storing bulbs in racks or mesh bags led to good yields and post-harvest quality and effective storage potential of garlic bulbs under ambient storage conditions. Kim et al. [12] quantify onion and garlic growth parameters using UAV-derived data, developing predictive models for bulb weight estimation. Mwinuka et al. [13] evaluate thermal and multispectral imaging for assessing African eggplant irrigation performance, highlighting the sensitivity of the crop water stress index derived from mobile phone-based thermal images. Salari et al. [14] conducted a study at the Agricultural Research Farm of Kabul University to study the effects of land management practices and sowing dates on onion growth and yield. Different agronomic traits including the number of leaves per plant, leaf length, leaf area per plant, leaf area index, normalized difference vegetation index (NDVI), maturity period, marketable yield, and total yield were studied in these trials. They found that the planting date had a significant effect on onion growth and yield, but land management practices did not have a significant effect on onion growth and maturity. Kim et al. [15] performed a garlic bulb weight estimation analysis reflecting the characteristics of Korean garlic grown in an open field using growth survey data. Additionally, they built a step-by-step model to predict garlic bulb weight based on the fact that factors affecting garlic bulb weight can vary depending on the growth stage. During the analysis process, LASSO regression analysis was used for variable selection and coefficient estimation of garlic bulb weight. Kim and Soon [16] learned the Neural Prophet (NP) lagged time-series model using onion and garlic data from the Korea Rural Economic Institute. And they predicted the average fresh bulb weight of onion and garlic using the learned NP model. The prediction results showed that the average absolute error was within 5%.
In this study, we investigated how vegetation index and growth data derived from multispectral imaging affect the real-time live bulb weight prediction of onions and garlic grown in a field during the cultivation period. To implement this problem, we describe the methods for manually collecting growth data and extracting various vegetation indices from multispectral images. We also used concurrency regression models and various machine learning methods to understand how the weight of raw bulbs increases during the growing season of onions and garlic.
The structure of the paper is as follows. Section 2 describes the procedure that was used to collect multispectral images of onions and garlic grown in each experimental plot using a camera mounted on a drone, and collected data by directly measuring the corresponding growth data and live bulb weight. The methods for extracting various vegetation indexes from multispectral images, which are used to predict the fresh bulb weight of onions and garlic, are presented. Various statistical models and ML methods that can predict the live bulb weight based on the extracted vegetation index and growth data are described. Section 3 describes the experimental procedure for predicting the live bulb weight using the R-package and compares the performance of various prediction methods. Section 4 discuss the current problem and future works. Section 5 summarizes the research results derived so far.

2. Materials and Methods

2.1. Study Area and Data Collection

First, to obtain the experimental data, onions and garlic were cultivated at the Chive Vegetable Research Institute in Muan-gun, Jeollanam-do. The cultivation period spanned from October 2022 to June 2023, and the final harvest was conducted on 2 June 2023. The experimental field was divided into two blocks, and each block was further divided into a total of 16 experimental plots. Figure 1 illustrates the overall structure of the experimental plot for onions and garlic.
Second, four growth parameters (plant height, leaf length, leaf width, and the number of leaves), as well as the live bulb weight of onions and garlic were measured at regular intervals. A total of 64 measurements were taken by randomly selecting four points in each of the 16 experimental units, and the same measurement was repeated seven times from 14 March to 30 May. The date on which each measurement was performed, and the measured values of each growth parameter and the live bulb weight at a particular location are given in Table 1.
Third, multispectral images of onions and garlic growing in the experimental area were repeatedly captured seven times using a drone, from 15 March to 31 May. The multispectral images comprised five channels of blue, green, red, red edge, and near-infrared (NIR). Figure 2 presents a sample multi-spectral image and its five distinct channel images.

2.2. Extraction of Vegetation Indexes

First, multispectral images were taken using a camera mounted on a drone to monitor the growth status of the onion and garlic crops in the experimental area during the growth period. The multispectral image acquired using the RedEdge-M multispectral camera (Micasense, Seattle, WA, USA) comprised five spectral bands: Blue (475 ± 20 nm), Green (560 ± 20 nm), Red (668 ± 10 nm), Red edge (717 ± 10 nm), and NIR (840 ± 40 nm). The parameters related to drone flight were set with a flight altitude of 20 m, a longitudinal overlap of 80%, and a lateral overlap of 60% for image capture. Multispectral images of the onions and garlic growing in the experimental area were collected through repeated filming six times from 15 March to 31 May. Figure 3 presents a sample multispectral image of the onion and garlic experimental plots.
Second, the methods for extracting vegetation indexes influencing the live bulb weight from the multispectral images of onions and garlic obtained during the cultivation period were examined [17,18,19,20,21]. For this task, 20 different vegetation indexes estimated from multispectral image bands (R, G, B, Red edge, and NIR) were used. The vegetation indexes were extracted using the numpy and skimage packages in a Python 3.8 environment. Each of the five multispectral images taken for each treatment plot was loaded using the skimage package, and the vegetation index was calculated by applying the relevant formula for each vegetation index using the numpy package. Table 2 specifies the names of the extracted vegetation indexes, their abbreviations, calculation formulas, and related references.

2.3. Prediction of the Live Bulb Weight

A correlation analysis method was designed to determine how the different vegetation indexes and growth data were related to the live bulb weight. Statistical regression models and ML methods capable of accurately predicting the live bulb weight were also examined. Particularly, we considered two functional simultaneous regression model-based methods, namely the Spline Smoothing-based method and Gaussian Process Regression-based method, and four ML methods, including Support Vector Regression, Random Forest Regression, AdaBoost, and Deep Neural Network, were examined.

2.3.1. Correlation Analysis

Generally, a statistical method that analyzes the correlation between two real variables, such as vegetation index and the live bulb weight, is called correlation analysis. Here, a measure of the correlation between the two variables, X and Y , was represented as the correlation coefficient defined as follows:
ρ = C o v ( X , Y ) S D X × S D ( Y ) ,
where C o v ( X , Y ) denotes the covariance of X and Y , and S D X and S D ( Y ) denote the standard deviation of X and Y , respectively. Furthermore, when n observed values for two variables ( X , Y ) are denoted as x 1 , y 1 ,   ,   ( x n , y n ) , the population correlation coefficient ρ is estimated as the sample correlation coefficient defined as follows.
r = i = 1 n ( x i x ¯ ) ( y i y ¯ ) ( x i x ¯ ) 2 ( y i y ¯ ) 2 .
Next, the problem of testing whether a relationship exists between the two variables was considered. First, the statistical hypothesis for the correlation coefficient is given as follows.
H 0 : ρ = 0 (No correlation exists) versus H 1 :   ρ 0 (Correlation exists).
Second, the test statistic and rejection area at significance level α were defined as follows.
T = n 2 1 r 2 r     t α 2 ( n 2 ) ,
where t α 2 ( n 2 ) represents the percentile of the t-distribution with (n − 2) degrees of freedom.

2.3.2. Nonparametric Functional Concurrent Regression Model

Generally, function-to-function regression refers to a situation where both independent and dependent variables in a regression model are of a functional nature. Functional concurrent regression is a specific type of function-to-function regression that relates the response function at a specific point to the covariate value at that point and the point itself [24,37,38,39,40,41,42,43,44,45,46,47]. Standard functional concurrent models are linear (a linear combination of the covariates is used), and are often criticized for their linearity assumption and lack of flexibility. To address this issue, a nonparametric functional concurrent regression that models the response function at a specific point using a multivariate non-parametric function of both the point and the covariate value at that point was considered. Such models offer heightened flexibility and predictive accuracy, especially when the underlying relationship is nonlinear.
Here, the nonparametric consistency regression model is expressed as a mathematical formula. The observed data for i th subjects was assumed to be { Y i t ,   X i 1 t ,   ,   X i Q t }  ( i = 1 , , n ) , where Y i t represents the functional response for t = t 1 , ,   t m and X i 1 t ,   ,   X i Q ( t ) for t = t 1 , ,   t m are the Q corresponding functional covariates. The general form of an additive nonlinear functional concurrent model with Q covariates is defined as follows:
Y i t = μ Y t + q = 1 Q F q X i q t , t + ϵ i t , i = 1 , , n
where μ t is an unknown intercept function, and F q K , q = 1 , , Q are unknown bivariate smooth functions.
Here, the methods for estimating the unknown intercept function   μ t and unknown bivariate smooth function F q X i q t , t given in model (1) were examined. Generally, there are three methods for estimating these two functions: a Spline Smoothing-based method, a Gaussian Process Regression-based method, and a Kernel Smoothing-based method.
First, an estimation method based on Spline Smoothing was considered. In the modeling process, { B μ , d ( t ) } d = 1 K μ was considered to be a set of univariate B-spline basis functions defined on [0, 1], where K μ represents the basis dimensions. Based on these basis functions, μ Y t = d = 1 K μ B μ , d ( t ) θ μ , d = B μ T ( t ) Θ μ , where B μ ( t ) is the K μ -dimensional vector of B μ , d ( t ) ’s, and Θ μ is the vector of unknown parameters θ μ , d ’s. Furthermore, F q X i q t , t was modeled as a bivariate basis expansion using a tensor product of univariate B-spline basis functions. For q = 1 , , Q , the B-spline basis functions { B X q , k ( x ) } k = 1 K x q and { B T q , l ( t ) } l = 1 K t q were employed for x and t , respectively, and were used to model F q X i q t , t . Thus, F q X i q t , t was represented as follows:
F q X i q t , t = k = 1 K x q l = 1 K t q θ q , k , l B X q , k x B T q , l t = Z i , q T ( t ) Θ q ,
where Z i , q T ( t ) is the K x q K t q -dimensional vector of B i , X q , k x B T q , l t , and Θ q is the vector of unknown parameters, θ i , k , l ’s. Based on the above expansions, model (1) was expressed as follows:
Y i t = B μ T t Θ μ + q = 1 Q Z i , q T ( t ) Θ q + ϵ i ( t )
In this representation, a large number of basis functions were expected to result in a better but rougher fit, while a small number of basis functions were expected to result in an overly smooth estimate. Inconsistent with the literature, a large number of basis functions were employed to fully capture the complexity of the functions, and the coefficients were penalized to ensure the smoothness of the resulting fit. Consequently, a criterion that incorporates penalty functions for parameter vectors was considered.
Here, Θ μ and Θ q ,   q = 1 , , Q were estimated by minimizing a penalized criterion as follows:
L Θ = i = 1 n Y i t ( B μ T t Θ μ + q = 1 Q Z i , q T t Θ q ) 2 + Θ μ T P μ Θ μ + q = 1 Q Θ q T P q Θ q
where P μ and P q ,   q = 1 , , Q are the penalty matrices for the smoothness of μ Y t and F q X i q t , t ,   q = 1 , , Q , respectively, and contain penalty parameters that regularize the trade-off between the goodness of fit and the smoothness of fit. The minimization of L Θ was straightforward. A closed-form expression of the estimator is as follows:
Θ ^ = i = 1 n Z i T Z i + P 1 i = 1 n Z i T Y i
where Y i = [ Y i t 1 , , Y i t m   ] T and Z i = [ B μ | Z i , 1 | | Z i , Q ] . The penalty parameters can be determined based on Generalized Cross-Validation (GCV) or restricted maximum likelihood (REML).
Finally, the prediction of response trajectory when a new covariate and its evaluation points are given was considered. Considering that a new unknown response Y n e w ( t ) is obtained based on new observations [ X 1 , n e w t ,   , X Q , n e w t ] , then, this new response can be predicted using the following equation:
Y ^ n e w t = d = 1 K μ B μ , d ( t ) θ ^ μ , d + q = 1 Q k = 1 K x , q l = 1 K t , q B X , q , k ( X q , n e w t ) B T , q , l ( t ) θ ^ q , k , l
where θ ^ μ , d and θ ^ q , k , l are estimated based on the above formula.
Second, an estimation method based on Gaussian Process Regression was explored. A Gaussian process is a set of random variables, where any finite subset follows a multivariate normal distribution. Such a process via G P { m · ,   C · , · } , where m · and C · , · are the mean and covariance functions, respectively. A realization X ( · ) from such a process is a random function with E X x = m ( x ) and C o v X x , X x = C ( x , x ) , where x and x are points in the domain of the process. Furthermore, for any finite set of points x 1 , , x n , the vector ( X ( x 1 ) , , X ( x n ) ) T followed a multivariate normal distribution.
Given a dataset consisting of n data points y i ,   x i ,   i = 1 , , n , where for each i , x i is a Q-dimensional vector of inputs, and y i is the output, a Gaussian Process Regression model is defined as follows:
Y i = f x i + ϵ i ,
where ϵ i ~ N ( 0 , σ 2 ) is an error term. The unknown function f x i is a nonlinear function of x i . The prior for this function was assumed to correspond to a Gaussian process, i.e., for each i , f x i followed a multivariate normal distribution with zero mean, and there existed a covariance function C x i , x j = C o v f x i , f x j . An example of such a covariance function is as follows:
C x i , x j = C x i , x j ; θ = v 0 e x p 1 2 q = 1 Q w q ( x i q x j q ) 2 + a 0 + a 1 q = 1 Q x i q x j q
where θ = ( w 1 , , w Q , v 0 , a 0 ,   a 1 ) denotes the set of unknown parameters. Therefore, Y = ( y 1 , , y n ) followed a normal distribution with zero mean and a covariance matrix as follows:
Ψ θ = C θ + σ v 2 I
where I is an identity matrix, C θ is an ( n × n ) matrix with elements as given in (11), and Ψ θ is an ( n × n ) matrix.
Shi et al. [45] developed a Gaussian Process Regressions framework for nonparametric concurrent models as follows:
Y i t = Z i T β t + F i X i 1 t ,   , X i Q t + ϵ i ( t ) ,
where β t is a vector of unknown coefficient functions, and F i { X i 1 t ,   , X i Q t } is the Gaussian process with zero mean and covariance kernel function C x i , x j   x i = x i ( t ) . The model components were estimated using a two-stage approach. In the first stage, each observed response curve was smoothed using a B-spline function: Y i t = Φ T t A i , where Φ t = [ Φ 1 t , , Φ Q ( t ) ] T are B-spline basis functions, and A i is a Q × 1 vector of coefficients. These coefficients were estimated by minimizing ( Y i t Φ T t A i ) 2 d t with respect to A i for each i = 1 , , n . The unknown coefficient function β ( t ) was also modeled using a B-spline basis function: β t = Φ T t B , where B is a matrix of unknown coefficients. The matrix B was estimated as ( Z T Z ) 1 Z T A , where Z = [ Z 1 , , Z n ] T and similar calculations were performed for A . In the second stage, the residuals from the first stage were computed as follows:
F i ~ X i 1 t ,   , X i Q t = Y i t Z i T β ^ t .
Then, F i ~ X i 1 t ,   , X i Q t was modeled through a Gaussian process with the covariance function C ( x , x :   θ ) . The covariance components involved in C ( · ) were estimated either through Maximum Likelihood Estimation (MLE) or using a Bayesian approach.
Next, the generation of a prediction y at a new test point ( t , x ) with x = x ( t ) was considered. Using Equation (11), the predicted value was given by y ^ = Z i T β ^ t + F ^ i { x } , where F ^ i { x } is predicted by its conditional mean E F i x D ) through the Gaussian process with the estimated covariance function C ( x , x :   θ ^ ) .

2.3.3. Machine Learning Methods

The first method, Support Vector Regression (SVR), was initially proposed by Drucker et al. [48]. It is a variant of support vector machines (SVMs), a popular algorithm for classification tasks. SVR extends an SVM’s capabilities to solve regression problems by optimizing an epsilon-insensitive loss function. SVR is an algorithm that aims to discover the optimal fit for a given dataset. It operates by constructing a hyperplane in high-dimensional feature space to maximize the margin around training data points while minimizing the error on unseen data points. Unlike other regression models that minimize the error across all data points, SVR focuses on minimizing the error within a specific margin.
The second method, Random Forest Regression (RFR), is a supervised learning algorithm that uses an ensemble learning method for regression [49] as shown in Figure 4. The ensemble learning method is a technique that combines predictions from multiple ML algorithms to generate predictions that are more accurate than those generated by a single model.
The RFR model is renowned for its robustness and accuracy, excelling across various problem domains, especially those involving features with non-linear relationships. However, it comes with notable drawbacks, including a lack of interpretability, susceptibility to overfitting, and the need to choose the number of trees for inclusion in the model.
The third method, Adaptive Boost (AdaBoost), is an algorithm that sets the initial model as a weak model. Subsequently, it uses weights at each step to sequentially fit a new model that compensates for the weaknesses of the previous model. The final model is created by linearly combining these sequentially fitted models. At every step, AdaBoost increases the weight of data with large errors or misclassification in the previous learning data, while reducing the weight of data with low errors or correct classification. By iteratively adjusting the weights and extracting new training data, AdaBoost emphasizes previously mispredicted data, resulting in a model that predicts them more accurately. Originally introduced for binary classification by Freund and Schapire [50], AdaBoost has been adapted for regression problems. Its success in delivering accurate ensembles and its resistance to overfitting led Breiman to call AdaBoost the “best off-the-shelf classifier in the world” (NIPS Workshop 1996) [51].
The fourth method is based on an Artificial Neural Network (ANN) or a simple traditional neural network, which aims to solve trivial tasks using a straightforward network outline. An ANN is loosely inspired by biological neural networks. It is a collection of layers to perform a specific task. Each layer consists of a collection of nodes that operate together. Typically comprising an input layer, one to two hidden layers, and an output layer, ANNs are suited for solving basic mathematical and computer problems, including fundamental gate structures with their respective truth tables. However, they face challenges when applied to complex tasks like image processing, computer vision, and natural language processing. In contrast, Deep Neural Networks (DNNs) feature a more intricate hidden layer structure, encompassing various layers, such as a convolutional layer, max-pooling layer, dense layer, and other unique layers [52]. These additional layers enhance the model’s ability to understand complex problems and provide optimal solutions. A DNN has more layers (more depth) than an ANN, and each layer adds complexity to the model and enables the model to process the inputs concisely to provide the ideal solution as the output. DNNs have garnered extremely high traction due to their high efficiency in handling a wide variety of deep-learning projects.

3. Experimental Results

3.1. Graphical Interpretation and Correlation Analysis

First, the relationships between the 20 vegetation indexes and the four growth parameters (plant height, leaf length, leaf width, and the number of leaves) were investigated using graph analysis and correlation analysis. Figure 5 presents scatter plots indicating the relationship between four growth characteristics and the NDVI, a representative vegetation index extracted from the multispectral images of onions and garlic. Figure 5 indicates that plant height, leaf width, and the number of leaves exhibited a highly positive correlation with the NDVI, while the leaf length exhibited a comparatively limited correlation with the NDVI, depending on the duration of cultivation.
Next, correlation analysis was conducted to determine the relationship between the 20 vegetation indexes and the four growth parameters. As shown in Table 3, three of the growth parameters, excluding the leaf length, exhibited a very high correlation with the 20 vegetation indexes. Notably, the leaf length exhibited a low correlation with the 20 vegetation indexes. The vegetation indexes that exhibited the highest correlation with the growth parameters were the NDRE, RDVI, PSND, and NDVI, which are commonly used indexes for determining the crop growth status. These findings are consistent with the observations derived from the scatter plots.
Second, graph analysis and correlation analysis were conducted to determine the relationships between the studied vegetation indexes and live bulb weight during the growing season of onions and garlic. Figure 6 visually illustrates the relationship between various vegetation indexes and the live bulb weights of onions and garlic. Figure 6 clearly indicates an overall negative correlation between the live bulb weight and vegetation indexes during the growing period. This observation is consistent with the fact that, as the stems and leaves of onions and garlic mature, they gradually dry out, leading to a decrease in the extracted vegetation indexes.
In addition, we calculated the correlation coefficient to numerically confirm that there is a negative correlation between the live bulb weight and vegetation index, which we identified from the above scatter plot. From Table 4 given below, we can see that most of the vegetation indices have a strong negative correlation with the live bulb weight.
Third, graph analysis and correlation analysis were conducted to determine the relationship between the four growth parameters and the live bulb weights of onions and garlic during the cultivation period. Figure 7 visually illustrates the relationships between the four growth characteristics of onions and garlic and the fresh bulb weight. The results indicate that as onions and garlic grow, the values of the three growth factors, excluding stem length, exhibit a negative correlation with the weight of the fresh bulb. On the other hand, only the leaf length exhibits a positive correlation with the bulb weight.
Additionally, the correlation coefficient was calculated to numerically substantiate the positive correlation observed in the scatter plots depicting the relationships between the live bulb weight and vegetation indexes. The correlation coefficients presented in Table 5 indicate a negative correlation between the live bulb weight and three of the growth factors, excluding the leaf length. On the other hand, among the growth factors studied, only the leaf length exhibited a positive correlation with the live bulb weight. This is consistent with the observations derived from the scatter plots.

3.2. Prediction of the Live Bulb Weight

First, 20 vegetation indexes were used as the input variables and two types of statistical regression models and four types of ML models were considered for predicting the raw weight of onions and garlic. To visually assess the prediction power of these models, two-dimensional scatter plots comparing the actual observed and predicted values were generated. Figure 8 illustrates the performance of the six prediction models. The results clearly indicate that FCR and RFR have the best prediction ability, followed by AdaBoost and SVR, while Gaussian Process Functional Data Analysis (GPFDA) and DNN have the lowest prediction ability.
Second, to predict the fresh bulb weight of onions and garlic, four growth factors (plant height, leaf width, leaf length, and number of leaves) were used as the input variables, and two types of statistical regression models and four types of ML models were considered. Two-dimensional scatter plots comparing the actual observed and predicted values were generated to visually present the prediction power of the models. Figure 9 clearly depicts that, in the case of onions, FCR and GPFDA have the best prediction ability, followed by RFR and AdaBoost, while SVM and DNN have the lowest prediction ability. On the other hand, in the case of garlic, FCR has the best prediction ability, followed by RFR and AdaBoost, while GPFDA, SVM, and DNN have a comparatively poorer prediction ability.
Third, the following four measures were considered to quantitatively evaluate the prediction accuracy of the two types of statistical regression models and four types of machine learning models used to predict the fresh bulb weight of onions and garlic in this paper. These are the coefficient of determination ( R 2 ), root mean square error (RMSE), normalized root mean square error (nRMSE), and mean absolute percentage error (MAPE), respectively.
R 2 = 1 i = 1 n ( y i y ^ i ) 2 i = 1 n ( y i y ¯ ) 2 , R M S E = i = 1 n ( y i y ^ i ) 2 n p 1 , n R M S E = 100 × R M S E y ¯ , M A P E = 100 n × i = 1 n | y i y ^ i y i |
where S S E is the sum of square error, S S T is the total sum of square, and n is the number of observations.
Fourth, the onion and garlic fresh bulb weight prediction performance of the six prediction models was quantitatively analyzed using four indicators. Hyperparameters for each machine learning algorithm are listed in Table 5. For the training and testing, we randomly shuffled the entire dataset and utilized 80% as training data, 20% as test data
Table 6 shows the four measurement values corresponding to the six prediction models used for predicting the live bulb weight of onions and garlic when 20 vegetation indexes were used as input variables. These indexes were obtained from images captured over multiple dates and calculated from those images. For each date of image capture, 20 vegetation indices were calculated, resulting in 20 vegetation values measured on the same date. The results clearly indicate that the FCR model has the best prediction power, the RFR and AdaBoost models have the second-best prediction power, and the SVR and DNN models have the third-best prediction power, while the GPFDA model has the poorest prediction power.
Table 7 shows the four measurement values corresponding to the six prediction models used for predicting the live bulb weight of onions and garlic when four growth factors were used as the input variables. The results indicate that the FCR model has the best prediction power, the RFR and AdaBoost models have the second-best prediction power, and the SVR and GPFDA models have the third-best prediction power, while the DNN model has the poorest prediction power.
Together, the results of the two types of analyses revealed that using the four growth factors rather than the 20 vegetation indexes as the input variables resulted in improved live bulb weight prediction for both onions and garlic. Moreover, the prediction power of the six prediction models maintained a consistent ranking, irrespective of whether growth factors or vegetation indexes were used as the input variables.

4. Discussion

Previous studies have primarily utilized artificial intelligence and machine learning technologies for productivity enhancement, disease management, and crop monitoring [1,2,3,4,5,6,7]. Additionally, attempts have been made to use UAVs and spectral images to acquire various biological parameters [17,18,19,20,21]. However, not much research has been conducted so far on predicting the weight of onion and garlic bulbs in real time.
This study was conducted in four stages. First, two pilot blocks each were selected for both onions and garlic, with each block comprising a total of 16 experimental units, four horizontals and four verticals. Thus, a total of 32 experimental units were prepared for both onions and garlic. Second, image data were collected using a multispectral camera, which was repeated a total of seven times for each area in the 32 experimental units prepared for both onions and garlic. Simultaneously, growth data and the live bulb weight at the corresponding points were recorded manually. Third, correlation analysis was conducted to determine the relationship between various vegetation indexes extracted from multispectral images and the manually measured growth data and live bulb weights. Fourth, based on the vegetation index extracted from multispectral images and previously collected growth data, we proposed a method to predict the live bulb weight of onions and garlic in real time during the cultivation period using a functional concurrence regression model and ML techniques.
The academic value and applicability of our research results described in this paper are as follows. First, our findings can inform agricultural practices aimed at optimizing the cultivation of onions and garlic. By understanding the factors influencing the weight of fresh bulbs during the growth period, farmers can adjust their cultivation methods, such as irrigation, fertilization, and pest control, to enhance the yield and quality. Second, our study contributes to the advancement of machine learning applications in agriculture. The development of predictive models for onion and garlic bulb weight estimation opens avenues for the creation of decision support systems for farmers. These systems can provide real-time recommendations based on environmental conditions and crop characteristics, aiding in more precise and efficient farming practices. Third, our research results can be applied to various fields beyond agriculture. Third, our research results can be applied to various fields beyond agriculture. The methodologies and techniques employed, such as image analysis and machine learning algorithms, can be adapted for the monitoring and managing of various other crops and plant species. This broader application potential underscores the significance of our research in advancing technology-driven solutions for sustainable agriculture and food security.

5. Conclusions

In this study, we introduced an automated method to predict the growth trajectory of onion and garlic bulb weights during the growing season. To establish this method, data, including multispectral images, growth factors, and the live bulb weight, were collected at consistent intervals from planting to harvest. Subsequently, six artificial intelligence models were employed to predict the live bulb weight based on 20 vegetation indexes derived from multispectral images, pre-collected growth factors, and live bulb weight data. The experimental results revealed that the FCR model showed the most robust prediction performance both when using four growth factors and when using 20 vegetation indexes. Following closely, with a slight distinction, GPFDA, RFR, and AdaBoost exhibited the next-best predictive power. However, SVM and DNN displayed comparatively poorer predictive power. Notably, when using the four growth factors as explanatory variables, all the predictive models exhibited improved prediction performance compared to that when using the 20 vegetation indexes.

Author Contributions

Study conception and design: M.H.N., W.C. and I.N.; data collection: D.K. and M.H.N.; software, D.K.; analysis and interpretation of results: I.N., W.C. and D.K.; manuscript draft preparation: W.C., D.K. and I.N. All authors have read and agreed to the published version of the manuscript.

Funding

This study was supported by the research funding from the Rural Development Administration (RDA) (RS-2022-RD010424).

Institutional Review Board Statement

Ethical review and approval were waived for this study due to the data collection being conducted through observational methods without direct interaction with human subjects. Therefore, no separate ethical review or approval process was deemed necessary.

Data Availability Statement

The datasets generated or analyzed during the study are available from the corresponding author upon reasonable request; however, some data are unavailable due to commercial restrictions.

Conflicts of Interest

The authors declare no conflicts of interest to report regarding the present study.

References

  1. Eli-Chukwu, N.C. Applications of Artificial Intelligence in Agriculture: A Review. Eng. Technol. Appl. Sci. Res. 2019, 9, 4377–4383. [Google Scholar] [CrossRef]
  2. Ayed, R.B.; Hanana, M. Artificial Intelligence to Improve the Food and Agriculture Sector. J. Food Qual. 2021, 2021, 5584754. [Google Scholar]
  3. Jagli, D.; Purohit, S.; Agale, T.; Kahar, D. Smart Farming Using Artificial Intelligence. In Proceedings of the ACM-2022: Algorithms Computing and Mathematics Conference, Chennai, India, 29–30 August 2022. [Google Scholar]
  4. Hossen, M.I.; Fahad, N.; Sarkar, M.R.; Rabbi, M.R. Artificial Intelligence in Agriculture: A Systematic Literature Review. Turk. J. Comput. Math. Educ. 2023, 14, 137–146. [Google Scholar]
  5. Akkem, Y.; Biswas, S.K.; Varanasi, A. Smart farming using artificial intelligence: A review. Eng. Appl. Artif. Intell. 2023, 120, 105899. [Google Scholar] [CrossRef]
  6. Oliveira, R.C.; Souza e Silva, R.D. Artificial Intelligence in Agriculture: Benefits, Challenges, and Trends. Appl. Sci. 2023, 13, 7405. [Google Scholar] [CrossRef]
  7. Sharma, A.; Jain, A.; Gupta, P.; Chowdary, V. Machine Learning Applications for Precision Agriculture: A Comprehensive Review. IEEE Access 2021, 9, 4843–4873. [Google Scholar] [CrossRef]
  8. Jeong, S.-J.; Kim, D.-W.; Yun, H.; Cho, W.-J.; Kwon, Y.-S.; Kim, H.-J. Monitoring the growth status variability in Onion (Allium cepa) and Garlic (Allium sativum) with RGB and multi-spectral UAV remote sensing imagery. In Proceedings of the 7th Asian-Australasian Conference on Precision Agriculture, Hamilton, New Zealand, 16–18 October 2017. [Google Scholar]
  9. Lee, J.; Min, B.; Yoon, S.; Lee, M.; Kim, H.; Hong, K. A multiple-regression model of bulb onion yield in response to meteorological conditions in Gyeongsangnam province, Republic of Korea. Acta Hortic. 2019, 1251, 81–90. [Google Scholar] [CrossRef]
  10. Przygocka-Cyna, K.; Barłóg, P.; Grzebisz, W.; Spizewski, T. Onion (Allium cepa L.) Yield and Growth Dynamics Response to In-Season Patterns of Nitrogen and Sulfur Uptake. Agronomy 2020, 10, 1146. [Google Scholar] [CrossRef]
  11. Desta, B.; Woldetsadik, K.; Ali, E.N. Effect of Harvesting Time, Curing and Storage Methods on Storability of Garlic Bulbs. Open Biotechnol. J. 2021, 15, 36–45. [Google Scholar] [CrossRef]
  12. Kim, D.-W.; Jeong, S.-J.; Lee, W.S.; Yun, H.; Chung, Y.S.; Kwon, Y.-S.; Kim, H.-J. Growth monitoring of field-grown onion and garlic by CIE L*a*b* color space and region-based crop segmentation of UAV RGB images. Precis. Agric. 2023, 24, 1982–2001. [Google Scholar] [CrossRef]
  13. Mwinuka, P.R.; Mbilinyi, B.P.; Mbungu, W.B.; Mourice, S.K.; Mahoo, H.F.; Schmitter, P. The feasibility of hand-held thermal and UAV-based multispectral imaging for canopy water status assessment and yield prediction of irrigated African eggplant (Solanum aethopicum L). Agric. Water Manag. 2021, 245, 106584. [Google Scholar] [CrossRef]
  14. Salari1, H.; Antil, R.S.; Saharawat, Y.S. Responses of onion growth and yield to different planting dates and land management practices. Agron. Res. 2021, 19, 914–1928. [Google Scholar]
  15. Kim, J.; Lee, H.; Suh, T. Analysis of Predictions of Garlic Bulb Weight using LASSO Regression Mode. Hortic. Sci. Technol. 2023, 41, 437–447. [Google Scholar]
  16. Kim, W.; Soon, B.M. Advancing Agricultural Predictions: A Deep Learning Approach to Estimating Bulb Weight Using Neural Prophet Model. Agronomy 2023, 13, 1362. [Google Scholar] [CrossRef]
  17. Xue, J.; Su, B. Significant Remote Sensing Vegetation Indices: A Review of Developments and Applications. Hindawi J. Sens. 2017, 2017, 1353691. [Google Scholar] [CrossRef]
  18. Gong, Y.; Duan, B.; Fang, S.; Zhu, R.; Wu, X.; Ma, Y.; Peng, Y. Remote estimation of rapeseed yield with unmanned aerial vehicle (UAV) imaging and spectral mixture analysis. Plant Methods 2018, 14, 70. [Google Scholar] [CrossRef] [PubMed]
  19. García-Martínez, H.; Flores-Magdaleno, H.; Ascencio-Hernández, R.; Khalil-Gardezi, A.; Tijerina-Chávez, L.; Mancilla-Villa, O.R.; Vázquez-Peña, M.A. Corn Grain Yield Estimation from Vegetation Indices, Canopy Cover, Plant Density, and a Neural Network Using Multispectral and RGB Images Acquired with Unmanned Aerial Vehicles. Agriculture 2020, 10, 277. [Google Scholar] [CrossRef]
  20. da Silva, E.E.; Baio, F.H.R.; Teodoro, L.P.R.; da Silva Junior, C.A.; Borges, R.S.; Teodoro, P.E. UAV-multispectral and vegetation indices in soybean grain yield prediction based on in situ observation. Remote Sens. Appl. Soc. Environ. 2020, 18, 100318. [Google Scholar] [CrossRef]
  21. Wan, L.; Cen, H.; Zhu, J.; Zhang, J.; Zhu, Y.; Sun, D.; Du, X.; Zhai, L.; Weng, H.; Li, Y.; et al. Grain yield prediction of rice using multi-temporal UAV-based RGB and multispectral images and model transfer—A case study of small farmlands in the South of China. Agric. For. Meteorol. 2020, 291, 108096. [Google Scholar] [CrossRef]
  22. Gitelson, A.A.; Vina, A.; Ciganda, V.; Rundquist, D.C.; Arkebauer, T.J. Remote estimation of canopy chlorophyll content in crops. Geophys. Res. Lett. 2005, 32. [Google Scholar] [CrossRef]
  23. Vincini, M.; Frazzi, E.; D’Alessio, P. A broad-band leaf chlorophyll vegetation index at the canopy scale. Precis. Agric. 2008, 9, 303–319. [Google Scholar] [CrossRef]
  24. Jiang, C.-R.; Wang, J.-L. Functional Single Index Models for Longitudinal Data. Ann. Stat. 2011, 39, 362–388. [Google Scholar] [CrossRef]
  25. Hunt, E.R.; Daughtry, C.S. What good are unmanned aircraft systems for agricultural remote sensing and precision agriculture? Int. J. Remote Sens. 2018, 39, 5345–5376. [Google Scholar] [CrossRef]
  26. Sripada, R.P.; Heiniger, R.W.; White, J.G.; Weisz, R. Aerial Color Infrared Photography for Determining Early In-Season Nitrogen Requirements in Corn. Agron. J. 2006, 98, 968–977. [Google Scholar] [CrossRef]
  27. Burgos-Artizzu, X.P.; Ribeiro, A.; Guijarro, M.; Pajares, G. Real-time image processing for crop/weed discrimination in maize fields. Comput. Electron. Agric. 2011, 75, 337–346. [Google Scholar] [CrossRef]
  28. Bendig, J.; Yu, K.; Aasen, H.; Bolten, A.; Bennertz, S.; Broscheit, J.; Gnyp, M.L.; Bareth, G. Combining UAV-based plant height from crop surface models, visible, and near infrared vegetation indices for biomass monitoring in barley. Int. J. Appl. Earth Obs. Geoinf. 2015, 39, 79–87. [Google Scholar] [CrossRef]
  29. Wang, C.; Gao, Y.Y.; Pop, I.M.; Vool, U.; Axline, C.; Brecht, T.; Heeres, R.W.; Frunzio, L.; Devoret, M.H.; Catelani, G.; et al. Measurement and control of quasiparticle dynamics in a superconducting qubit. Nat. Commun. 2014, 5, 5836. [Google Scholar] [CrossRef] [PubMed]
  30. Hamuda, E.; Glavin, M.; Jones, E. A survey of image processing techniques for plant extraction and segmentation in the field. Comput. Electron. Agric. 2016, 125, 184–199. [Google Scholar] [CrossRef]
  31. Rondeaux, G.; Steven, M.; Baret, F. Optimization of soil-adjusted vegetation indices. Remote Sens. Environ. 1996, 55, 95–107. [Google Scholar] [CrossRef]
  32. Black, P.; William, D. Assessment and classroom learning. Assess. Educ. 1998, 5, 7–74. [Google Scholar] [CrossRef]
  33. Roujean, J.-L.; Breon, F.-M. Estimating PAR absorbed by vegetation from bidirectional reflectance measurements. Remote Sens. Environ. 1995, 51, 375–384. [Google Scholar] [CrossRef]
  34. Zhong, L.; Zhang, Y.; Duan, C.A.; Deng, J.; Pan, J.; Xu, N.L. Causal contributions of parietal cortex to perceptual decision-making during stimulus categorization. Nat. Neurosci. 2019, 22, 963–973. [Google Scholar] [CrossRef] [PubMed]
  35. Raper, T.B.; Varco, J.J. Canopy-scale wavelength and vegetative index sensitivities to cotton growth parameters and nitrogen status. Precis. Agric. 2015, 16, 62–76. [Google Scholar] [CrossRef]
  36. Hunt, T.E.; Clark-Carter, D.; Sheffield, D. The Development and Part Validation of a U.K. Scale for Mathematics Anxiety. J. Psychoeduc. Assess. 2011, 29, 455–466. [Google Scholar] [CrossRef]
  37. Morris, J.S. Functional Regression. Annu. Rev. Stat. Its Appl. 2015, 2, 321–359. [Google Scholar] [CrossRef]
  38. Maity, A. Nonparametric functional concurrent regression models. Adv. Rev. 2019, 9, e1394. [Google Scholar] [CrossRef]
  39. Leroux, A.; Xiao, L.; Crainiceanu, C.; Checkley, W. Dynamic prediction in functional concurrent regression with an application to child growth. Stat. Med. 2018, 37, 1376–1388. [Google Scholar] [CrossRef]
  40. Janet, S.; Kim, J.S.; Arnab Maity, A.; Ana-Maria Staicu, A.-M. Additive nonlinear functional concurrent model. Stat. Its Interface 2018, 11, 669–685. [Google Scholar]
  41. Bhattacharjee, S.; Müller, H.-G. Concurrent Object Regression. Electron. J. Stat. 2022, 16, 4031–4089. [Google Scholar] [CrossRef]
  42. Pan, R.; Wang, Z.; Wu, Y. Detection of Interaction Effects in a Nonparametric Concurrent Regression Model. Entropy 2023, 25, 1327. [Google Scholar] [CrossRef]
  43. Ghosal, R.; Maity, A. Variable selection in nonparametric functional concurrent regression. Can. J. Stat. 2022, 50, 142–161. [Google Scholar] [CrossRef]
  44. Shi, J.Q.; Wang, B.; Murry-Smith, R.; Titterington, D.M. Gaussian Process Functional Regression Modelling for Batch Data. Biometrics 2007, 63, 714–723. [Google Scholar] [CrossRef] [PubMed]
  45. Shi, J.Q.; Murry-Smith, R.; Titterington, D.M. Hierarchical Gaussian process mixtures for regression. Stat. Comput. 2005, 15, 31–41. [Google Scholar] [CrossRef]
  46. Sun, Y.; Fang, X. Robust estimators of functional single index models for longitudinal data. In Communications in Statistics-Theory and Methods; Taylor & Francis: Abingdon, UK, 2023; pp. 1–15. [Google Scholar]
  47. Konzon, E.; Cheng, Y.; Shi, J.Q. Gaussian Process for Functional Data Analysis: The GPFDA Package for R. arXiv 2021, arXiv:2102.00249, 1–24. [Google Scholar]
  48. Drucker, H.; Burges, C.J.C.; Kaufman, L.; Smola, A.; Vapnik, V. Support vector regression machines. In Advances in Neural Information Processing Systems 9 (NIPS’96), Proceedings of the 9th International Conference on Neural Information Processing Systems, Denver CO, USA, 3–5 December 1996; MIT Press: Cambridge, MA, USA, 1996; pp. 155–161. [Google Scholar]
  49. Borup, D.; Christensen, B.J.; Mühlbach, N.S.; Nielsen, M.S. Targeting predictors in random forest regression. Int. J. Forecast. 2023, 39, 841–868. [Google Scholar] [CrossRef]
  50. Freund, Y.; Schapire, R.E. A decision-theoretic generalization of on-line learning and an application to boosting. J. Comput. Syst. Sci. 1997, 55, 119–139. [Google Scholar] [CrossRef]
  51. Ferrario, A.; Hämmerli, R. On Boosting: Theory and Applications. ETH Zurich Research Collection 2019; pp. 1–40. Available online: https://ssrn.com/abstract=3402687 (accessed on 9 May 2024).
  52. Iqbal, H.; Sarker, I.H. Deep Learning: A Comprehensive Overview on Techniques, Taxonomy, Applications and Research Directions. SN Comput. Sci. 2021, 2, 420. [Google Scholar]
Figure 1. The structure of the experimental plots for onions and garlic.
Figure 1. The structure of the experimental plots for onions and garlic.
Agriculture 14 00754 g001
Figure 2. A sample multi-spectral image and its five distinct channel images.
Figure 2. A sample multi-spectral image and its five distinct channel images.
Agriculture 14 00754 g002
Figure 3. Sample multispectral images of the onion and garlic experimental plots.
Figure 3. Sample multispectral images of the onion and garlic experimental plots.
Agriculture 14 00754 g003
Figure 4. Flow of ensemble learning for regression in Random Forest.
Figure 4. Flow of ensemble learning for regression in Random Forest.
Agriculture 14 00754 g004
Figure 5. Scatter plots depicting the relationships between four growth characteristics and NDVI.
Figure 5. Scatter plots depicting the relationships between four growth characteristics and NDVI.
Agriculture 14 00754 g005
Figure 6. Scatter plots depicting the relationships between live bulb weight and four vegetation indexes.
Figure 6. Scatter plots depicting the relationships between live bulb weight and four vegetation indexes.
Agriculture 14 00754 g006
Figure 7. Scatter plots depicting the relationships between live bulb weight and four growth parameters.
Figure 7. Scatter plots depicting the relationships between live bulb weight and four growth parameters.
Agriculture 14 00754 g007
Figure 8. Scatter plots depicting the relationships between the observed and predicted values of live bulb weight based on 20 vegetation indexes.
Figure 8. Scatter plots depicting the relationships between the observed and predicted values of live bulb weight based on 20 vegetation indexes.
Agriculture 14 00754 g008
Figure 9. Scatter plots depicting the relationships between the observed and predicted values of live bulb weight based on four growth factors.
Figure 9. Scatter plots depicting the relationships between the observed and predicted values of live bulb weight based on four growth factors.
Agriculture 14 00754 g009
Table 1. Measurement dates, growth parameter values, and the live bulb weight of onions and garlic during the cultivation period.
Table 1. Measurement dates, growth parameter values, and the live bulb weight of onions and garlic during the cultivation period.
CropCrop ParametersDate of Observation
3/144/34/195/95/235/30
GarlicPlant height38.2561.6386.2584.2571.0062.75
Leaf length72127.2536.253637.75
Leaf width10.5514.22516.714.8314.612.68
No. of leaves 6.257.585.7554
Live bulb weight0.0380.5698.53328.66744.66753.000
OnionPlant height30.0848.1368.7570.0058.7546.33
Leaf length3.855917.87518.2519.67
Leaf width9.815.3820.919.039.6310.67
No. of leaves5.87.59.57.85.54.7
Live bulb weight0.1381.97629.633101169.33173.33
Table 2. Details of the multispectral vegetation indexes evaluated in the present study.
Table 2. Details of the multispectral vegetation indexes evaluated in the present study.
IndicesAcronymEquationReference
Chlorophyll index redCIred R n R r 1 Gitelson et al. (2005) [22]
Chlorophyll vegetation indexCVI R n R r R g 2 Vincini et al. (2008) [23]
Enhanced vegetation indexEVI 2.5 ( R n R r ) R n + 6 R r 7.5 R b + 1 Gitelson et al. (2005) [22]
Two-band enhanced vegetation indexEVI2 2.5 ( R n R r ) R n + 2.4 R r + 1 Jiang et al. (2008) [24]
Green chlorophyll indexGCI R n R g 1 Gitelson et al. (2005) [22]
Green normalized difference vegetation indexGNDVI R n R g R n + R g Hunt & Daughtry (2018) [25]
Green ratio vegetation indexGRVI R n R g Sripada et al. (2006) [26]
Modified excess greenMEXG 1.262 R g 0.884 R r 0.311 R b Burgos-Artizzu et al. (2011) [27]
Modified normalized green–red differenceMNGRD R g 2 R r 2 R g 2 + R r 2 Bendig et al. (2015) [28]
Normalized difference red edgeNDRE R n R r e R n + R r e Wang et al. (2014) [29]
Normalized difference vegetation indexNDVI R n R r R n + R r Gitelson et al. (2005) [22]
Normalized green–red differenceNGRD R g R r R g + R r Hamuda et al. (2016) [30]
Optimized soil adjusted vegetation indexOSAVI R n R r R n + R r + 0.16 Roundeaux et al. (1996) [31]
Pigment-specific normalized vegetation indexPSND R n R b R n + R b Blackbum (1998) [32]
Renormalized difference vegetation indexRDVI R n R r ( R n R r ) 0.5 Roujean & Breon (1995) [33]
Red-edge chlorophyll indexRECI R n R r e 1 Gitelson et al. (2005) [22]
Ratio vegetation indexRVI R n R r Gitelson et al. (2005) [22]
Soil adjusted vegetation indexSAVI 1.5 ( R n R r ) R n + R r + 0.5 Zhong et al. (2019) [34]
Simplified canopy chlorophyll content indexSCCCI N D R E N D V I Raper & Varco (2015) [35]
Triangular greenness indexTGI R g 0.39 R r 0.61 R b Hunt et al. (2011) [36]
Table 3. Correlation coefficients for the relationships between four growth parameters and 20 vegetation indexes (Bold number is highest score in the column).
Table 3. Correlation coefficients for the relationships between four growth parameters and 20 vegetation indexes (Bold number is highest score in the column).
IndexGarlicOnion
Plant
Height
Leaf
Length
Leaf
Width
No. of LeavesPlant
Height
Leaf
Length
Leaf
Width
No. of Leaves
CIred0.7540.1790.8270.7130.611−0.0050.6630.584
CVI0.7680.3290.7070.4720.635−0.0960.7450.750
EVI0.6930.0510.8370.8300.8030.1500.8300.777
EVI20.7310.0970.8420.8040.8200.1790.8300.788
GCI0.7940.2200.8410.6960.666−0.0110.7320.673
GNDVI0.7950.2070.8430.7170.7810.1110.8060.787
GRVI0.7940.2200.8410.6960.666−0.0110.7320.673
MEXG0.6610.1360.7640.7130.8270.6500.5910.519
MNGRD0.6940.1200.8120.7550.7820.2490.7310.645
NDRE0.7630.1660.8680.7710.7330.0170.8080.780
NDVI0.7770.1870.8480.7420.8210.2110.8000.765
NGRD0.6940.1210.8110.7520.7690.2350.7220.634
OSAVI0.7530.1310.8480.7850.8270.2010.8210.782
PSND0.8590.3830.8150.5670.8700.3260.7820.778
RDVI0.7040.0720.8250.8090.8450.2410.8330.795
RECI0.7570.1620.8640.7630.671−0.0450.7700.714
RVI0.7540.1790.8270.7130.611−0.0050.6630.584
SAVI0.7330.1010.8420.8020.8280.1960.8310.790
SCCCI0.5520.0640.7150.6930.474−0.2970.6770.731
TGI0.6320.1860.6700.5890.6700.8100.3370.297
Table 4. Correlation coefficients for the relationships between live bulb weight and 20 vegetation indexes.
Table 4. Correlation coefficients for the relationships between live bulb weight and 20 vegetation indexes.
CropIndex r Index r Index r Index r
GarlicEVI−0.88213SAVI−0.86864RECI−0.82706GNDVI−0.79407
MNGRD−0.87429RDVI−0.86652CIred−0.81859TGI−0.73566
NGRD−0.87265OSAVI−0.85837RVI−0.81859PSND−0.73122
EVI2−0.87170NDVI−0.83005GCI−0.79714SCCCI−0.63974
MEXG−0.87026NDRE−0.82949GRVI−0.79714CVI−0.60588
OnionSCCCI−0.81404EVI−0.78047NDVI−0.76211CIred−0.64783
NDRE−0.80220OSAVI−0.77935PSND−0.74064RVI−0.64783
CVI−0.78715EVI2−0.77855GRVI−0.72242NGRD−0.63909
GNDVI−0.78523RDVI−0.77170GCI−0.72242MEXG−0.17486
SAVI−0.78135RECI−0.76769MNGRD−0.64898TGI0.45755
Table 5. Hyperparameters of each machine learning algorithm.
Table 5. Hyperparameters of each machine learning algorithm.
DNNSVRRFRAdaBoost
Hyper
Parameters
ValueHyper
Parameters
ValueHyper
Parameters
ValueHyper
Parameters
Value
optimizeradamC10n_estimators100n_estimators100
learning rate0.07epsilon0.2criterionsquared_errorlosslinear
lossmsekernellinearmax_depth8
epochs100
batch_size8
Table 6. Predictions of live bulb weight using 20 vegetation indexes.
Table 6. Predictions of live bulb weight using 20 vegetation indexes.
CropModelR2RMSEnRMSEMAPE
GarlicFCR0.9623.23315.10813.283
GPFDA0.79510.38348.85639.654
SVR0.9165.75025.18219.688
RFR0.9404.86821.61014.757
AdaBoost0.9524.39019.48712.927
DNN0.9384.98922.14516.217
OnionFCR0.97311.41312.4959.428
GPFDA0.85252.76449.31939.490
SVR0.77438.49837.00627.215
RFR0.95321.16120.34112.310
AdaBoost0.96019.60018.84011.858
DNN0.90229.37228.23319.438
Table 7. Predictions of live bulb weight using growth factor items.
Table 7. Predictions of live bulb weight using growth factor items.
CropModelR2RMSEnRMSEMAPE
GarlicFCR0.9932.0249.4597.744
GPFDA0.8966.83832.17620.253
SVR0.9026.15027.30218.918
RFR0.9365.23723.24715.350
AdaBoost0.9305.35823.78415.219
DNN0.8507.47533.18224.404
OnionFCR0.9957.7828.5207.157
GPFDA0.98015.11614.1294.111
SVR0.91327.28126.22319.043
RFR0.95520.07419.29511.764
AdaBoost0.94921.84220.99514.548
DNN0.90726.81025.77116.809
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, D.; Cho, W.; Na, I.; Na, M.H. Prediction of Live Bulb Weight for Field Vegetables Using Functional Regression Models and Machine Learning Methods. Agriculture 2024, 14, 754. https://doi.org/10.3390/agriculture14050754

AMA Style

Kim D, Cho W, Na I, Na MH. Prediction of Live Bulb Weight for Field Vegetables Using Functional Regression Models and Machine Learning Methods. Agriculture. 2024; 14(5):754. https://doi.org/10.3390/agriculture14050754

Chicago/Turabian Style

Kim, Dahyun, Wanhyun Cho, Inseop Na, and Myung Hwan Na. 2024. "Prediction of Live Bulb Weight for Field Vegetables Using Functional Regression Models and Machine Learning Methods" Agriculture 14, no. 5: 754. https://doi.org/10.3390/agriculture14050754

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop