Next Article in Journal
Testing and Evaluation of Wi-Fi RTT Ranging Technology for Personal Mobility Applications
Previous Article in Journal
Spatially Offset Raman Spectroscopy Combined with Attention-Based LSTM for Freshness Evaluation of Shrimp
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Age Classification of Rice Seeds in Japan Using Gradient-Boosting and ANFIS Algorithms

1
School of Systems Engineering, Kochi University of Technology, Kochi 782-8502, Japan
2
Faculty of Agriculture, Kochi University, Kochi 780-8072, Japan
3
School of Information and Communications Technology, Hanoi University of Science and Technology, No. 1, Dai Co Viet Road, Hanoi 100000, Vietnam
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(5), 2828; https://doi.org/10.3390/s23052828
Submission received: 31 December 2022 / Revised: 28 February 2023 / Accepted: 3 March 2023 / Published: 5 March 2023
(This article belongs to the Section Smart Agriculture)

Abstract

:
The rapidly changing climate affects an extensive spectrum of human-centered environments. The food industry is one of the affected industries due to rapid climate change. Rice is a staple food and an important cultural key point for Japanese people. As Japan is a country in which natural disasters continuously occur, using aged seeds for cultivation has become a regular practice. It is a well-known truth that seed quality and age highly impact germination rate and successful cultivation. However, a considerable research gap exists in the identification of seeds according to age. Hence, this study aims to implement a machine-learning model to identify Japanese rice seeds according to their age. Since agewise datasets are unavailable in the literature, this research implements a novel rice seed dataset with six rice varieties and three age variations. The rice seed dataset was created using a combination of RGB images. Image features were extracted using six feature descriptors. The proposed algorithm used in this study is called Cascaded-ANFIS. A novel structure for this algorithm is proposed in this work, combining several gradient-boosting algorithms such as XGBoost, CatBoost, and LightGBM. The classification was conducted in two steps. First, the seed variety was identified. Then, the age was predicted. As a result, seven classification models were implemented. The performance of the proposed algorithm was evaluated against 13 state-of-the-art algorithms. Overall, the proposed algorithm has a higher accuracy, precision, recall, and F1-score than the others. For the classification of variety, the proposed algorithm scored 0.7697, 0.7949, 0.7707, and 0.7862, respectively. The results of this study confirm that the proposed algorithm can be employed in the successful age classification of seeds.

1. Introduction

Japan is a country with a population of 127 million people. For over 2000 years, rice has been a significant staple in Japan. Around 300 of the 40,000 or so distinct types of rice produced worldwide can be found in Japan [1]. A tendency toward sustainable rice production through technical advancements has emerged, particularly in light of the rising scarcity of resources such as water and land. Researchers are driven to find new solutions to the declining or stagnant yields brought on by poor grain quality and rising production costs due to a significant reliance on agricultural inputs.
Nevertheless, despite these limitations, rice output needs to dramatically increase in the following generation to meet the global food demand, especially for the poor. To ensure the food supply and social, economic, and water sustainability of Asia, a region in which rice is deeply culturally ingrained, it is crucial to produce more rice with a restricted or controlled flow of resources [2]. According to previous research, the quality of rice seeds used primarily for rice cultivation is mainly determined by the age of the rice seeds after harvest. Therefore, this research aims to create a method that can more accurately determine or validate the word-of-mouth age of rice seeds, which might be one factor in assessing the overall quality of rice seeds.
Due to its climate and topography, Japan is particularly susceptible to natural disasters, including earthquakes, storms, and flooding. Unfavorable weather conditions may result in crop failure or harvest loss. Hence, using old rice seeds for cultivation is a common practice. By more accurately determining the age of rice seeds, care necessary for treatment might be taken.
In [3], the germination rate behavior of Portulaca oleracea L. seeds was examined in relation to their place of origin, seed maturation period, and seed age. Three geographical areas representing three different climates were used to gather seeds; then, they were cultured under diverse light and temperature conditions. Compared to seeds from the Canadian site, the results showed that seeds from the United Arab Emirates location exhibited less dormancy and germinated more quickly under a wider variety of incubation conditions. At the Canadian location, seed age substantially impacted germination rate but not for seeds from the Egyptian or United Arab Emirates sites. Overall, the study emphasizes the significance of considering environmental factors when determining how seeds adapt to dormancy and germination rate. In [4], the authors looked at how alfalfa’s tolerance to salt during germination rate changed with natural and artificial aging. The findings demonstrated that different aged seed lots had considerably varying salt tolerances and that seed age enhanced the amount of solute leakage that occurred after ingestion.
Moreover, ref. [5] examined the relationship between seed age and seedling vigor and competitiveness in populations of Bromus tectorum, an annual grass found in meadows and sagebrush steppes. When grown in competition, old seeds from the meadow steppe population displayed germination rate delays, which lowered plant growth and biomass, whereas aged seeds from the sagebrush steppe population did not. According to the study, physiological expenses related to seed age may impact aboveground competitive interactions and the relative fitness of older cohorts in the soil seed bank.
Although the relationship between seed age and germination rate is well established in the literature, to our knowledge, no research has been conducted on age-based rice seed categorization. The relevant works on the correlation between germination rate and rice seed age and the classification of seeds are introduced in the following paragraphs. However, these are not restricted to rice.
According to the above-stated literature, there are several reasons why a machine would be necessary to develop a real-time application for classifying rice seed varieties and identifying their harvested age:
  • Speed:
    Machines can process and analyze large amounts of data much faster than humans, making it possible to classify and identify rice seed varieties in real-time.
  • Accuracy:
    Machines can analyze data with a high level of accuracy, reducing the chances of errors in classification and identification.
  • Consistency:
    Machines can perform tasks consistently and accurately, ensuring that the classification and identification of rice seed varieties are consistent across different batches.
  • Efficiency:
    Machines can work continuously without breaking, allowing for more efficient and cost-effective data processing.
  • Cost:
    Using machines can reduce labor costs, as they can perform tasks that would otherwise require human labor. This can make the development of a real-time application more cost-effective.
Therefore, this study aims to implement a machine learning model that can be employed to develop a real-time application to classify rice seed varieties and identify their harvested age. The following points are introduced as the research outcomes of this study.
  • Implementation of rice seeds dataset based on varieties and harvested age:
    Developing a dataset from scratch when working with AI and machine learning can be challenging, especially if you do not have access to a large and diverse dataset. This is because the performance of a machine learning model is heavily dependent on the quality and quantity of the data on which it is trained. When a dataset is undersized or generated using an artificial method, this can lead to an unsatisfactory performance when using a machine learning algorithm. One solution is to consider whether it is necessary to use a machine learning model. Sometimes, a more straightforward approach, such as a rule-based system or a decision tree, can be used to solve the problem. Here, a dataset was developed from scratch due to the lack of age-based seed image data. This study introduces a rice seed dataset with six different seed varieties and three age categories for the classification task. To our knowledge, this dataset is the only one labeled based on the harvested age of the seeds.
  • Investigation of Red, Green, and Blue (RGB) features for accurate classification: Various features were comprehensively evaluated due to the use of RGB images. The feature descriptors used in this study can be introduced as follows.
    • Color structure;
    • Edge histogram;
    • Region shape;
    • Gray-level co-occurrence matrix (GLCM);
    • Mean value of RGB spaces;
    • Column layout.
  • Implementation of a novel machine-learning model for the cost-effective and efficient identification of the rice seeds’ variety and age: the proposed algorithm is a combination of the gradient-boosting algorithm and cascaded adaptive network-based fuzzy inference system (Cascaded-ANFIS) algorithms. The performances of the proposed algorithm were compared with several feature-based machine learning algorithms.

2. Related Works

In-depth research on the correlation between germination rate and rice seed age was conducted by Jones et al. in 1926 [6]. For the majority of rice types, seed age was suggested to be inversely correlated with the germination rate. Eight different types of rice were employed in the study, and age intervals of six years were investigated. According to a survey, the association between germination rate and wheat seed age was the same as in the prior study. According to them, the germination rate declines at a rate of 0.243% h−1 as people age [7]. Canola seeds were studied by Yun et al. to determine how much seed aging affects the germination rate [8]. They also showed that older seeds have a lower germination rate than fresh ones.
Additionally, Ibrahim et al. in 2013 [9] and Tabatabaei in 2014 [10] reported results from two different studies on seed germination rate with aging and reached the same conclusions. These signs play a significant role in the age-based categorization of seeds. Wu and Tsai presented a leaf image noise reduction. They were able to achieve 92.13% accuracy using background removal and ROI extraction approaches as the innovative forms of implementation [11].
For various tasks assessing food quality, machine vision systems have been developed [12,13,14,15]. Research has concentrated on fusing image analysis with machine learning techniques to create new automatic inspection and certification approaches. The quality control or cultivar categorization tasks primarily examined in [16] for rice seeds (polished) are pertinent to the work presented here. Y. Ogawa provides a thorough overview of computer vision methods, measurements of physical properties, chemical content, and distributions of rice grains for seed quality management in [16].
Using an automated machine vision system to classify rice seeds often involves many steps. Collecting picture data and feature extraction are among the most crucial examples. Morphological, color, and textural qualities, individually or in combination, are frequently used in appearance-based techniques. Lai et al. [17] proposed using image analysis to detect the physical parameters of and manually categorize wheat grains, a proposal that dates back to 1986. Two-dimensional image analysis is used by Sakai et al. [18] to manually categorize four different species of polished rice grains by extracting their dimensions and form parameters. It is common practice to extract the form descriptors from the seed samples and train classifiers such as random forests (RF) [19], neural networks (NN) [20], or a Cubic B-Splines shape model [21].
In contrast to characteristics that are more frequently employed in the literature, such as the chaff tip and depth of concavities of rice kernels, Huang et al. [22] conducted an extensive examination of shape descriptors. While just three kinds of rice were evaluated, their study demonstrates encouraging results in the separation of superficially similar species. With a standard derivation of 7.0%, Kuo et al. [23] investigated 30 types of rice seeds using sparse representation classification with an accuracy of 89.1%. The bulk of the literature employs a small number of species, which the writers briefly acknowledged. However, they did not show how this may affect the reader’s capacity to discriminate between different species. They used a systematic approach, concentrating on the grains’ specific regions of interest.
HSI approaches have recently been used in food and agricultural engineering. Wang et al. [24], using VIS/NIR spectral data, distinguished three different rice cultivars. The authors combined several features from the obtained HSI pictures, including the degree of chalkiness, form features, and spectral properties. Principle component analysis (PCA) was utilized to decrease the dimensionality of the spectral data. The principal components were then used to train an artificial neural network, yielding a classification accuracy of 94.45%. The authors of [25] found a practical way to monitor the nitrogen status in rice using a combination of the least squares support vector machine (LS-SVM) regression algorithm and VIS/NIR spectroscopy within a range of 325–1075 nm. Four rice seed varieties were recently identified using an HSI technique in [26]. The authors [26] achieved up to 100% accuracy in their findings using a random forest (RF) classifier and the whole spectral range of their system, 1039–1612 nm. It is uncertain how the inter/intra-class changed over the four cultivars in [26] because they were crossbred with different species.
The studies in [27,28] investigated several feature combination techniques to find the best feature combination. When combining spectral, texture characteristics, and morphology, the best accuracy (91.67%) was attained in [28]. Using a dataset containing six different rice seed types and a mix of spectral and spatial information, the authors of [27] reported a classification accuracy of 84%.
The previous studies on this subject do not provide sufficient information on the age classification of seeds. Therefore, this study is novel to the best of the authors’ knowledge. There are several methods of seed variety classification based on hyperspectral images. However, a hyperspectral dataset does not provide a convenient environment to implement a system that farmers and other interested parties can use. Hence, the main focus of this study is to implement a comparatively effective and efficient rice seed classification system based on harvested age.

3. Methodology

3.1. Dataset Construction and Preprocessing

The rice seed dataset was newly constructed. This study was conducted with six rice varieties of different harvest ages. Initially, the rice seeds were collected with the help of Prof Akira Miyazaki from the Agriculture Faculty at Kochi University, Japan. Figure 1 shows that the dataset was constructed using a conveyer belt setup. The ultimate objective of this study is to implement a mobile application that can be used in real-time with a smartphone camera. Therefore, as shown in the figure, a smartphone camera was used for data acquisition.
The smartphone that we used was Xiaomi 11T Pro. The macro-lens was used in the camera configuration to obtain feature-rich images. The specifications of the macro-lens are as follows: a 5 Megapixel 1/5′ sensor with an f/2.4 aperture lens. A controller was developed to synchronize the conveyor belt movements and camera shuttle speed. The images were obtained at five-second intervals. Each time, the conveyor belt stopped and obtained a macro-image of the rice seed. The images were saved in JPEG format.
The dataset was preprocessed in three steps: (a) the images were captured using the smartphone camera; (b) the backgrounds were removed from the raw dataset; (c) seeds were cropped and isolated. The background was removed using the rembg library in python, and segmentation was carried out using the OpenCV4 platform. The outputs of each process are shown in Figure 2.
Table 1 shows the completed dataset’s descriptive information. Six rice varieties were used to develop the dataset, namely, Akitakomachi, Fusaotome, Hatsuboshi, Koshihikari, Okiniiri, and Yang DAO-8 (Figure 3). These rice varieties originated in Japan, except Yang DAO-8. Yang DAO-8 originated in China. The harvest ages of these rice samples were 2012, 2016, and 2020. However, Okiniiri seeds comprised only samples from 2012 and 2016, while Yang DAO-8 contained samples from 2012 and 2020. Hence, the total number of classes could be calculated as 16. The completed dataset was uploaded and is publicly available in the Kaggle data repository under Japanese rice seeds agewise classification.
The dataset was divided at a ratio of 7:3 for training and testing, and the same datasets were employed for all algorithms used in this study.

3.2. Feature Extraction

The main component in classifier implementation is feature extraction. As a result, six feature descriptors were used to extract various features from the rice seed dataset. Each of these feature extraction techniques is briefly described in this section. The Color Structure descriptor is the first technique. Although it is based on histogram equalization, it aims to differentiate localized color differences for each color and provides a comprehensive explanation [29]. The Region Shape is the following feature description. Due to the inherent challenges in depicting forms, the shape features are less developed than their color and textural equivalents [30]. It is not feasible to precisely segment an image into meaningful regions using low-level features because of the variety of ways in which a 3D object can be projected into 2D shapes, the complexity of each object’s shape, and the presence of shadows, occlusions, non-uniform illumination, and varying surface reflectivity. As a result, the third feature extraction technique employed the column layout feature descriptor.
The edge histogram (EDH) descriptor was used to show how local edges are distributed throughout images [31]. As a result, this study’s fourth feature extraction technique was the EDH descriptor. The histogram was used to describe edges, a crucial aspect for visualizing picture data. The EDH-described characteristics of a picture cannot be replicated by the uniform color histogram and texture features [32,33]. The gray level co-occurrence matrix is the fifth characteristic descriptor (GLCM). Given a picture made up of pixels, each with a certain intensity, it calculates how frequently particular pairings of gray levels co-occur in an image or part of an image (a specific gray level). The change in intensity at the pixel of interest is measured using the GLCM contents in texture feature computations [34]. Moreover, the sixth feature descriptor was the mean values of red, green, and blue channels.

3.3. Machine Learning Algorithm Development

3.3.1. Gradient-Boosting Algorithms

Gradient-boosting algorithms are used in most literature studies due to their easy implementation, low computational cost and efficiency. This study proposes a novel machine-learning method for rice seed classification, and gradient-boosting algorithms are one of its key components. Gradient-boosting algorithms are based on the suspicion that the overall prediction error is minimized when previous models are combined with the best possible forthcoming model. Setting the expected outcomes for this following model is crucial to minimizing errors. Each case’s target outcome will differ depending on how changing a case’s forecast impacts the overall prediction error. This technique is referred to as “gradient boosting” because target outcomes are defined for each case depending on the gradient of the inaccuracy of the prediction. Each new model advances in a way that minimizes prediction error in the potential predictions for each training instance [35].
  • XGBoost: eXtreme Gradient Boosting.
    Chen et al. [36] invented the XGBoost algorithm. Gradient-boosting machines are used in a novel and extensible way that has been found to increase the computational efficiency of boosted tree algorithms. They were developed specifically to boost model performance and computational effectiveness.
    In an ensemble strategy known as “boosting”, adding more models fixes the errors introduced by previous models. The gradient-boosting approach involves the creation of new models that predict the residuals of previous models, which are then integrated to obtain the conclusive prediction. The model addition process is repeated until there is no observable improvement. A gradient descent method reduces the loss when adding new models.
    A total of 17 of the 29 machine learning (ML) projects posted on Kaggle were successfully completed by XGBoost by 2015. Speed was significantly boosted by using many CPU cores, reducing the look-up times of individual trees created with XGBoost. This method was constructed in R and Python using the SciKit-Learn [37] package and uses unique regularization approaches.
  • CatBoost: Categorical Boosting.
    Diverse characteristics, noisy data, and complex connections can be dealt with using a powerful machine-learning technique called gradient boosting. In 2017, CatBoost, a machine learning method based on gradient-boosting decision trees (GBDT), was introduced by Yandex engineers [38]. CatBoost has the following advantages over other GBDT algorithms:
    • The algorithm effectively handles category features. Using traditional GBDT methods, categorical traits can be substituted by suitable average label values.
    • In CatBoost, many category traits are blended. CatBoost uses a greedy approach to integrate all categorical features from the dataset with all categorical traits and combinations in the current tree.
    • CatBoost can be used to alleviate gradient bias. Each iteration of GBDT generates a weak learner, and each learner is taught using the gradient of the preceding learner. The total findings from each learner’s categorization comprise the output [39].
  • LightGBM: Light Gradient Boosting.
    The LightGBM [40] algorithm from Microsoft is an open-source GBDT. The histogram-based approach provides the foundation for the parallel voting decision tree technique, which speeds up training, uses less memory, and integrates complex network connections to maximize parallel learning [41,42]. In each iteration, the local voting decision is made, selecting the top k characteristics and the global voting choice to receive the top attributes. The training data are distributed among many computers. LightGBM uses the leaf-wise method to determine which leaf has the most significant splitter gain.

3.3.2. Adaptive Network-based Fuzzy Inference System (ANFIS)

A multi-layer adaptive network-based fuzzy inference system called ANFIS was suggested by Jang [43]. When learning and fine-tuning fuzzy inference system (FIS) parameters using a hybrid learning mode, an ANFIS consists of five layers that implement various node functions. The following parameters are updated, and the errors are transferred to the backward pass using the least squared error estimation approach in the forward pass with fixed premise parameters. While fixing the subsequent parameters, the backward pass changes the premise parameters using the gradient descent method. The assumption and associated parameters for membership functions (MF) and fussy inference system (FIS) will be revealed by repeatedly performing forward and backward passes. In automation control [44] and other domains, ANIFS is frequently used.

3.4. Cascaded-ANFIS

Cascaded-ANFIS is an extension of the ANFIS algorithm. The Cascaded-ANFIS was introduced in 2021 and showed several benefits compared to the traditional ANFIS algorithm. ANFIS has two significant limitations, such as the curse of dimensionality and higher computational power consumption. The Cascaded-ANFIS removes these limitations using a simple ANFIS algorithm in different configurations.
Figure 4 illustrates the Cascaded-ANFIS algorithm’s creation.
As shown in the figure, there are two primary components: (1) pair selection and (2) train model. Generally, the Cascaded-ANFIS algorithm selects the best pairs and trains them individually using a two-input–one-output ANFIS model. Then, the output of each node is transferred to the next level as the input.
The pair selection module uses the sequential feature selection (SFS) procedure. The novel aspect of the method is identifying the best fit for each input variable using a two-input–one-output ANFIS model. A nested loop cycles through all potential pair combinations to achieve this. The two-input ANFIS model is then used with these as inputs. Then, the root means square error (RMSE) is calculated, recorded, and compared to the previous RMSE. Finding the lowest RMSE value at the end of the second loop will reveal the matched pair. The training phase can start after the pairings are chosen.
In the training instance, a two-input ANFIS model is also used. The input can be directly sent to the ANFIS module, which can produce current outputs and RMSE for each data pair, as the input variables are paired with the best match from the previous module. There is also a target error in place at this point. As a result, the goal error and RMSE are contrasted. If the desired error is achieved, the process may be terminated. If not, the algorithm moves on to the second iteration.
Rathnayake et al. [45,46] provide further information and technical specifics regarding Cascaded-ANFIS. The novel Cascaded-ANFIS method is capable of handling computational complexity with ease. The distinctive methods created by Cascaded-ANFIS can also manage noisy datasets.

3.5. Proposed Approach

One of the crucial benefits of the Cascaded-ANFIS algorithm is that it can be restructured according to the problem statement. The present study is based on image data, and features are extracted from the images. This method provides an extensive number of feature dimensions. Therefore, this study proposes combining selected gradient-boosting algorithms (XGBoost, CatBoost, and LightGBM) in a Cascaded-ANFIS structure. The proposed approach is illustrated in Figure 5.
As shown in the figure, the image data extracted features employing six feature descriptors. The total number of features that were used in this study is 159. Then, the extracted features were fed to the gradient-boosting algorithms to predict the output. These outputs were then used to train the Cascaded-ANFIS algorithm. The figure shows that the pairs were pre-assigned as XGBoost-CatBoost and CatBoost-LightGBM. This combination was selected based on the testing results of all pairs. The proposed novel structure of the Cascaded-ANFIS algorithm has two levels.
The parameter tuning of each algorithm was performed using the GridSearchCV method of the sci-kit-learn library. The tuned parameters and their values are presented in Table 2 below. This study was conducted purely in a CPU-based environment. The experimental platform information is presented in Table 3.

4. Experimental Results

4.1. Evaluation Criteria

Using a confusion matrix, the proposed model’s performance was examined. Classification matrices were computed and illustrated in Equations (1)–(4) to comprehend the confusion matrix.
A c c u r a c y A v g = i = 1 l t p i + t p i t p i + f n i + f p i + t n i l
P r e c i s i o n μ = i = 1 l t p i i = 1 l ( t p i + f p i )
R e c a l l μ = i = 1 l t p i i = 1 l ( t p i + f n i )
F 1 -S c o r e μ = ( β 2 + 1 ) P r e c i s i o n μ R e c a l l μ β 2 P r e c i s i o n μ + R e c a l l μ
True Positive, False Positive, True Negative and False Negative are denoted as t p i , f p i , t n i , and f n i , respectively. Moreover, l and μ indicate the number of classes and micro-average. When the problem is multiclass, each of these factors offers essential information about the effectiveness of the classification [47].
Moreover, the experiment was conducted in several steps. The planned identification process has two steps: (1) identifying the rice variety; (2) identifying the rice seed age. Hence, this study contains seven classification tasks (i.e., classification between the six rice varieties and age classification models of six rice varieties).
Nevertheless, the performances of the proposed algorithm were comprehensively evaluated with 13 other algorithms. These are as follows.
  • Nearest neighbors;
  • Linear support vector machines (Linear SVM);
  • Radial basis function kernel SVM (RBF SVM);
  • Gaussian process;
  • Decision tree;
  • Random forest;
  • Neural net;
  • Adaptive boosting (AdaBoost);
  • Naive Bayes;
  • Quadratic discriminant analysis (QDA);
  • XGBoost;
  • CatBoost;
  • LightGBM.

4.2. Rice Variety Classification

Table 4 shows the performance in terms of rice variety classification. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.7697, 0.7949, 0.7707, and 0.7862.

4.3. Age Classification of Each Variety

Once the classification of the rice variety is completed, the age of the rice has to be identified. Therefore, several classifiers were trained to evaluate the classification results.

4.3.1. Akitakomachi

Table 5 shows the Akitakomachi rice seed age classification. The results are promising. The proposed model performed better than other algorithms. However, the XGBoost algorithm performance was significantly similar to that of the proposed model. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.7551, 0.7579, 0.7552, and 0.7556.

4.3.2. Fusaotome

Table 6 shows the Fusaotome rice seed age classification. The results were the same as those of the Akitakomachi. The proposed model performed better than other algorithms. However, the CatBoost algorithm performance was significantly similar to that of the proposed model. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.8612, 0.8720, 0.8618, and 0.8616.

4.3.3. Hatsuboshi

Table 7 shows the Hatsuboshi rice seed age classification. The classification results indicated that the neural nets achieved better results than other algorithms. The proposed model shows the second-best performance. However, the results were significantly similar to those of the neural nets. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.7989, 0.8193, 0.7976, and 0.7912.

4.3.4. Koshihikari

Table 8 shows the Koshihikari rice seed age classification. The results are promising. The proposed model performed better than other algorithms. However, the XGBoost algorithm performance was significantly similar to that of the proposed model. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.8815, 0.8806, 0.8810, and 0.8693.

4.3.5. Okiniiri

Table 9 shows the Okiniiri rice seed age classification. The proposed model performance did not achieve the best results. The LightGBM model showed the best results in terms of Okiniiri rice seed age classification. However, the proposed algorithm’s performance was significantly similar to that of the LightGBM model. The table shows that the proposed model achieved a maximum accuracy, precision, recall, and F1-score of 0.9512, 0.9521, 0.9514, and 0.9512, while LightGBM poses 0.9583, 0.9585, 0.9583, and 0.9583, respectively.

4.3.6. Yang DAO-8

Table 10 shows the Yang DAO-8 rice seed age classification. The proposed algorithm achieved the best results in terms accuracy and recall, and F1-score, i.e., 0.7639, 0.7665, and 0.7479. The precision of the proposed algorithm is second to the LightGBM, which is 0.8196, while the proposed algorithm achieved a score of 0.8062. The difference between the precision values is 0.01. Hence, the proposed algorithm can also be selected as the best approach for this classification task.

5. Discussion

Most of the experiment results showed that the proposed algorithm outperformed the other algorithms used in this study. In terms of rice variety classification, the proposed algorithm showed an accuracy of 0.7697, while the second-best algorithm XGBoost achieved an accuracy of 0.7633. The LightGBM and CatBoost algorithm achieved 0.7366 and 0.6288 accuracies, respectively. The rice varieties used in this study are very similar in shape, except for the Yang Dao-8. As shown in Figure 3, Akitakomachi, Fusaotome, Hatuboshi, Koshihikari, and Okiniiri rice seeds are similar in shape. However, the Yang Dao-8 rice seed is longer than the other seeds. This structure similarity impacted the rice seeds’ classification due to the use of Edge Histogram Features.
The results of the age classification of each variety mainly depend on the texture changes due to the aging of the seeds. The Akitakomachi, Fusaotome, Koshihikari, and Yang Dao-8 seed age-based classification was dominated by the proposed model, which achieved the best accuracies. However, the neural nets gave the best results for the Hatsuboshi rice seed age classification, with an accuracy of 0.8167, while the proposed model showed a 0.7989 accuracy. The difference between the accuracies of neural nets and the proposed model is 0.0178, which is significantly small. The best Okiniiri rice age classification accuracy was achieved by the LightGBM model, while the proposed model showed the second-best results, with accuracies of 0.9583 and 0.9512, respectively. The difference between these two models is 0.0071.
Overall, it can be stated that the proposed algorithm showed the best results in all experiments. There are several reasons for the proposed algorithm’s obtaining the best results. Considering each experiment and gradient-boosting model result, it can be seen that each experiment has a different best form: XGBoost, CatBoost and LightGBM. For example, the XGBoost algorithm performed well in the Koshihikari age classification, while in the Okiniiri classification, the LightGBM outperformed XGBoost. As shown in Figure 5, the proposed algorithm is a combination of these three gradient-boosting algorithm results. In other words, the Cascaded-ANFIS section of the proposed algorithm does not depend on the features of the image dataset but on the results of the gradient-boosting models. Therefore, the fuzzy-based ANFIS algorithm calculates the reasoning by mixing the three gradient-boosting models to enhance the performance. This can be clearly seen in each experiment as the algorithm outperformed the gradient-boosting model accuracies. The two-input–one-output ANFIS models generate precise membership functions to deal with the outputs of the gradient-boosting algorithms and enhance the overall accuracy.

6. Conclusions

Identifying seeds by their age is a challenging task. To the best of our knowledge, there are zero studies in the literature on this subject. Therefore, this study aimed to design and develop a machine learning algorithm to classify seeds by age. Due to the lack of data availability, the main objective of this work was to construct a novel dataset using six varieties of Japanese rice. Each rice variety included three harvesting ages, except for Okiniiri and Yang DAO-8. The harvested ages of these rice varieties were 2012, 2016, and 2020.
The rice seed dataset was implemented by constructing a conveyor belt system to automatically acquire the seed images. A smartphone camera with a macro-lens captured many surface features and the system’s real-time flexibility. The dataset was divided into a training set and a testing set at a ratio of 7:3, and the same datasets were used to model the classifiers. Six feature-extractors were used to find the critical points of the RGB image dataset. The features were selected based on the success rate and popularity of previous research.
The proposed algorithm is a combination of four unique machine-learning algorithms. XGBoost, CatBoost, and LightGBM gradient-boosting algorithms were the base classifiers of the proposed algorithm, and the secondary output was tuned using the Cascaded-ANFIS algorithm. The Cascaded-ANFIS can change the structure depending on the problem statement. This study has a high feature dimension, and using the Cascaded-ANFIS as the base classifier could rapidly increase the time consumption and computational power. Therefore, gradient-boosting algorithms were initially appointed to predict the classification outputs, while the Cascaded-ANFIS evaluated the base results.
The experiment was conducted using two steps: classification based on the rice variety and identifying the age of the seed. Since six rice varieties were available, seven classifiers were trained accordingly. Each classification task was evaluated using the confusion matrix parameters: accuracy, precision, recall, and F1-score. Moreover, the performances of the proposed algorithm were comprehensively assessed by training 13 other machine learning algorithms. The results indicate that the proposed algorithm is more capable of identifying the seed variety and age. Although other algorithms obtained better results for some occurrences, the differences in the results between other algorithms and the proposed algorithm were insignificant.
According to the results of this study, the proposed algorithm can identify the variety of the seed and age with higher efficiency and effectiveness. The algorithms were trained using only the CPU power. Therefore, this study can be considered as a solution to replace black box algorithms that require higher complexity and increased power consumption. How to implement the machine learning model as a server and introduce it as a mobile phone application can comprise future study objectives. Improving the dataset density could also be a future research goal.

Author Contributions

Conceptualization, N.R. and A.M.; methodology, N.R.; software, N.R.; validation, T.L.D., Y.H. and A.M.; formal analysis, A.M.; investigation, N.R.; resources, A.M.; data curation, A.M.; writing—original draft preparation, N.R.; writing—review and editing, T.L.D. and Y.H.; visualization, N.R.; supervision, Y.H.; project administration, A.M. and Y.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset that was used in this study is publicly available in the Kaggle data repository. (https://www.kaggle.com/datasets/namalrathnayake1990/japanese-rice-seeds-agewise-classification (accessed on 15 January 2023)).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AdaBoostadaptive boosting
CPUcentral processing unit
CatBoostcategorical boosting
Cascaded-ANFIScascaded adaptive network-based fuzzy inference system
DOAJdirectory of open access journals
FISfuzzy inference system
GBDTgradient-boosting decision trees
GLCMgray level co-occurrence matrix
HSIhyper-spectral images
XGBoosteXtreme gradient boosting
LS-SVMleast squares support vector machines
LightGBMlight gradient boosting
MFmembership function
MDPImultidisciplinary digital publishing institute
MLmachine learning
NNneural networks
PCAprinciple component analysis
RGBred, green, and blue
ROIregion of interest
RFrandom forests
RBFradial basis function
QDAquadratic discriminant analysis

References

  1. Yamaji, S.; Itô, S. The political economy of rice in Japan. In Japanese and American Agriculture; Routledge: Lobdon, UK, 2019; pp. 349–365. [Google Scholar]
  2. Chapagain, T.; Yamaji, E. The effects of irrigation method, age of seedling and spacing on crop performance, productivity and water-wise rice production in Japan. Paddy Water Environ. 2010, 8, 81–90. [Google Scholar] [CrossRef]
  3. El-Keblawy, A.; Al-Ansari, F. Effects of site of origin, time of seed maturation, and seed age on germination behavior of Portulaca oleracea from the Old and New Worlds. Can. J. Bot. 2000, 78, 279–287. [Google Scholar]
  4. Smith, S.; Dobrenz, A. Seed age and salt tolerance at germination in alfalfa 1. Crop. Sci. 1987, 27, 1053–1056. [Google Scholar] [CrossRef]
  5. Rice, K.J.; Dyer, A.R. Seed aging, delayed germination and reduced competitive ability in Bromus tectorum. Plant Ecol. 2001, 155, 237–243. [Google Scholar] [CrossRef]
  6. Jones, J.W. Germination of rice seed as affected by temperature, fungicides, and age. Agron. J. 1926, 18, 576–592. [Google Scholar] [CrossRef] [Green Version]
  7. Soltani, E.; Galeshi, S.; Kamkar, B.; Akramghaderi, F. The Effect of Seed Aging on the Seedling Growth as Affected by. Res. J. Environ. Sci. 2009, 3, 184–192. [Google Scholar] [CrossRef]
  8. Ren, Y.; Zhu, J.; Hussain, N.; Ma, S.; Ye, G.; Zhang, D.; Hua, S. Seedling age and quality upon transplanting affect seed yield of canola (Brassica napus L.). Can. J. Plant Sci. 2014, 94, 1461–1469. [Google Scholar] [CrossRef] [Green Version]
  9. Ibrahim, H.; Oladiran, J.; Mohammed, H. Effects of seed dormancy level and storage container on seed longevity and seedling vigour of jute mallow (Corchorus olitorius). Afr. J. Agric. Res. 2013, 8, 1370–1374. [Google Scholar] [CrossRef] [Green Version]
  10. Tabatabaei, S. The changes of germination characteristics and enzyme activity of barley seeds under accelerated aging. Cercet. Agron. Mold 2015, 48, 62–67. [Google Scholar] [CrossRef] [Green Version]
  11. Wu, Y.J.; Tsai, C.M.; Shih, F. Improving leaf classification rate via background removal and ROI extraction. J. Image Graph. 2016, 4, 93–98. [Google Scholar] [CrossRef] [Green Version]
  12. Jackman, P.; Sun, D.W.; Allen, P. Recent advances in the use of computer vision technology in the quality assessment of fresh meats. Trends Food Sci. Technol. 2011, 22, 185–197. [Google Scholar] [CrossRef]
  13. Polak, A.; Coutts, F.K.; Murray, P.; Marshall, S. Use of hyperspectral imaging for cake moisture and hardness prediction. IET Image Process. 2019, 13, 1152–1160. [Google Scholar] [CrossRef] [Green Version]
  14. Qiao, T.; Ren, J.; Craigie, C.; Zabalza, J.; Maltin, C.; Marshall, S. Quantitative prediction of beef quality using visible and NIR spectroscopy with large data samples under industry conditions. J. Appl. Spectrosc. 2015, 82, 137–144. [Google Scholar] [CrossRef] [Green Version]
  15. Kelman, T.; Ren, J.; Marshall, S. Effective classification of Chinese tea samples in hyperspectral imaging. Artif. Intell. Res. 2013, 2. [Google Scholar] [CrossRef] [Green Version]
  16. Fabiyi, S.D.; Vu, H.; Tachtatzis, C.; Murray, P.; Harle, D.; Dao, T.K.; Andonovic, I.; Ren, J.; Marshall, S. Varietal classification of rice seeds using RGB and hyperspectral images. IEEE Access 2020, 8, 22493–22505. [Google Scholar] [CrossRef]
  17. Lai, F.; Zayas, I.; Pomeranz, Y. Application of pattern recognition techniques in the analysis of cereal grains. Cereal Chem. 1986, 63, 168–172. [Google Scholar]
  18. Sakai, N.; Yonekawa, S.; Matsuzaki, A.; Morishima, H. Two-dimensional image analysis of the shape of rice and its application to separating varieties. J. Food Eng. 1996, 27, 397–407. [Google Scholar] [CrossRef]
  19. Hong, P.T.T.; Hai, T.T.T.; Hoang, V.T.; Hai, V.; Nguyen, T.T. Comparative study on vision based rice seed varieties identification. In Proceedings of the 2015 Seventh International Conference on Knowledge and Systems Engineering (KSE), Ho Chi Minh City, Vietnam, 8–10 October 2015; pp. 377–382. [Google Scholar]
  20. Liu, Z.y.; Cheng, F.; Ying, Y.b.; Rao, X.Q. Identification of rice seed varieties using neural network. J. Zhejiang-Univ.-Sci. 2005, 6, 1095–1100. [Google Scholar] [CrossRef] [Green Version]
  21. Peralta, C.N.M.; Pabico, J.P.; Mariano, V.Y. Modeling shapes using uniform cubic b-splines for rice seed image analysis. In Proceedings of the 2016 IEEE Sixth International Conference on Communications and Electronics (ICCE), Ha-Long, Vietnam, 27–29 July 2016; pp. 326–331. [Google Scholar]
  22. Huang, K.Y.; Chien, M.C. A novel method of identifying paddy seed varieties. Sensors 2017, 17, 809. [Google Scholar] [CrossRef]
  23. Kuo, T.Y.; Chung, C.L.; Chen, S.Y.; Lin, H.A.; Kuo, Y.F. Identifying rice grains using image analysis and sparse-representation-based classification. Comput. Electron. Agric. 2016, 127, 716–725. [Google Scholar] [CrossRef]
  24. Wang, L.; Liu, D.; Pu, H.; Sun, D.W.; Gao, W.; Xiong, Z. Use of hyperspectral imaging to discriminate the variety and quality of rice. Food Anal. Methods 2015, 8, 515–523. [Google Scholar] [CrossRef]
  25. Shao, Y.; Zhao, C.; Bao, Y.; He, Y. Quantification of nitrogen status in rice by least squares support vector machines and reflectance spectroscopy. Food Bioprocess Technol. 2012, 5, 100–107. [Google Scholar] [CrossRef]
  26. Kong, W.; Zhang, C.; Liu, F.; Nie, P.; He, Y. Rice seed cultivar identification using near-infrared hyperspectral imaging and multivariate data analysis. Sensors 2013, 13, 8916–8927. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Vu, H.; Tachtatzis, C.; Murray, P.; Harle, D.; Dao, T.K.; Le, T.L.; Andonovic, I.; Marshall, S. Spatial and spectral features utilization on a hyperspectral imaging system for rice seed varietal purity inspection. In Proceedings of the 2016 IEEE RIVF International Conference on Computing & Communication Technologies, Research, Innovation, and Vision for the Future (RIVF), Hanoi, Vietnam, 7–9 November 2016; pp. 169–174. [Google Scholar]
  28. Sun, J.; Lu, X.; Mao, H.; Jin, X.; Wu, X. A method for rapid identification of rice origin by hyperspectral imaging technology. J. Food Process. Eng. 2017, 40, e12297. [Google Scholar] [CrossRef]
  29. Kim, S.H.; Park, R.H. An efficient algorithm for video sequence matching using the modified Hausdorff distance and the directed divergence. IEEE Trans. Circuits Syst. Video Technol. 2002, 12, 592–596. [Google Scholar]
  30. Rathnayake, N.; Rathnayake, U.; Dang, T.L.; Hoshino, Y. An Efficient Automatic Fruit-360 Image Identification and Recognition Using a Novel Modified Cascaded-ANFIS Algorithm. Sensors 2022, 22, 4401. [Google Scholar] [CrossRef]
  31. Park, D.K.; Jeon, Y.S.; Won, C.S. Efficient use of local edge histogram descriptor. In Proceedings of the 2000 ACM workshops on Multimedia, Los Angeles, CA, USA, 30 October–3 November 2000; pp. 51–54. [Google Scholar]
  32. Somnugpong, S.; Khiewwan, K. Content-based image retrieval using a combination of color correlograms and edge direction histogram. In Proceedings of the 2016 13th International Joint Conference on Computer Science and Software Engineering (JCSSE), Khon Kaen, Thailand, 13–15 July 2016; pp. 1–5. [Google Scholar]
  33. Lin, C.H.; Liu, C.W.; Chen, H.Y. Image retrieval and classification using adaptive local binary patterns based on texture features. IET Image Process. 2012, 6, 822–830. [Google Scholar] [CrossRef] [Green Version]
  34. Hall-Beyer, M. GLCM texture: A tutorial. Natl. Counc. Geogr. Inf. Anal. Remote Sens. Core Curric 2000, 3, 1–12. [Google Scholar]
  35. Bentéjac, C.; Csörgo, A.; Martínez-Muñoz, G. A comparative analysis of gradient boosting algorithms. Artif. Intell. Rev. 2021, 54, 1937–1967. [Google Scholar] [CrossRef]
  36. Chen, T.; Guestrin, C. Xgboost: A scalable tree boosting system. In Proceedings of the 22nd Acm Sigkdd International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  37. Pedregosa, F.; Varoquaux, G.; Gramfort, A.; Michel, V.; Thirion, B.; Grisel, O.; Blondel, M.; Prettenhofer, P.; Weiss, R.; Dubourg, V.; et al. Scikit-learn: Machine learning in Python. J. Mach. Learn. Res. 2011, 12, 2825–2830. [Google Scholar]
  38. Prokhorenkova, L.; Gusev, G.; Vorobev, A.; Dorogush, A.V.; Gulin, A. CatBoost: Unbiased boosting with categorical features. Adv. Neural Inf. Process. Syst. 2018, 31. [Google Scholar]
  39. Friedman, J.H. Stochastic gradient boosting. Comput. Stat. Data Anal. 2002, 38, 367–378. [Google Scholar] [CrossRef]
  40. Meng, Q.; Ke, G.; Wang, T.; Chen, W.; Ye, Q.; Ma, Z.M.; Liu, T.Y. A communication-efficient parallel algorithm for decision tree. Adv. Neural Inf. Process. Syst. 2016, 29. [Google Scholar]
  41. Jin, R.; Agrawal, G. Communication and memory efficient parallel decision tree construction. In Proceedings of the 2003 SIAM International Conference on Data Mining, San Francisco, CA, USA, 1–3 May 2003; pp. 119–129. [Google Scholar]
  42. Ranka, S.; Singh, V. CLOUDS: A decision tree classifier for large datasets. In Proceedings of the 4th Knowledge Discovery And Data Mining Conference, New York, NY, USA, 27–21 August 1998. [Google Scholar]
  43. Jang, J.S. ANFIS: Adaptive-network-based fuzzy inference system. IEEE Trans. Syst. Man Cybern. 1993, 23, 665–685. [Google Scholar] [CrossRef]
  44. Melin, P.; Castillo, O. Intelligent control of a stepping motor drive using an adaptive neuro-fuzzy inference system. Inf. Sci. 2005, 170, 133–151. [Google Scholar] [CrossRef]
  45. Rathnayake, N.; Dang, T.L.; Hoshino, Y. A novel optimization algorithm: Cascaded adaptive neuro-fuzzy inference system. Int. J. Fuzzy Syst. 2021, 23, 1955–1971. [Google Scholar] [CrossRef]
  46. Rathnayake, N.; Rathnayake, U.; Dang, T.L.; Hoshino, Y. A Cascaded Adaptive Network-Based Fuzzy Inference System for Hydropower Forecasting. Sensors 2022, 22, 2905. [Google Scholar] [CrossRef]
  47. Sokolova, M.; Lapalme, G. A systematic analysis of performance measures for classification tasks. Inf. Process. Manag. 2009, 45, 427–437. [Google Scholar] [CrossRef]
Figure 1. Rice seeds dataset construction setup.
Figure 1. Rice seeds dataset construction setup.
Sensors 23 02828 g001
Figure 2. Image preprocessing steps. (a) The images were captured using a smartphone camera. (b) The backgrounds were removed from the raw dataset. (c) Seeds were cropped and isolated.
Figure 2. Image preprocessing steps. (a) The images were captured using a smartphone camera. (b) The backgrounds were removed from the raw dataset. (c) Seeds were cropped and isolated.
Sensors 23 02828 g002
Figure 3. Rice Varieties (harvested in 2012). (a) Akitakomachi. (b) Fusaotome (c) Hatsuboshi. (d) Koshihikari. (e) Okiniiri. (f) Yang Dao-8.
Figure 3. Rice Varieties (harvested in 2012). (a) Akitakomachi. (b) Fusaotome (c) Hatsuboshi. (d) Koshihikari. (e) Okiniiri. (f) Yang Dao-8.
Sensors 23 02828 g003
Figure 4. Construction of the Cascaded-ANFIS algorithm.
Figure 4. Construction of the Cascaded-ANFIS algorithm.
Sensors 23 02828 g004
Figure 5. Construction of the novel proposed algorithm.
Figure 5. Construction of the novel proposed algorithm.
Sensors 23 02828 g005
Table 1. Rice seed dataset sample count.
Table 1. Rice seed dataset sample count.
Rice Variety201220162020
Akitakomachi306279366
Fusaotome325416482
Hatsuboshi376495499
Koshihikari340348509
Okiniiri345292N/A
Yang DAO-8343N/A261
Table 2. Machine learning algorithm parameters tuned using the GridSearchCV.
Table 2. Machine learning algorithm parameters tuned using the GridSearchCV.
AlgorithmParameterValue
XGBOOSTobjectivemulti:softmax
colsample_bytree0.5
learning_rate0.3
max_depth6
Alpha10
n_estimators750
subsample0.7
CatBoostIterations90
learning_rate0.04
eval_metricMultiClass
sampling_frequencyPerTree
penalties_coefficient1
max_leaves64
permutation_count4
Depth4
LightGBMnum_leaves31
objectivebinary
learning_rate0.1
boosting_typegbdt
Cascaded-ANFISMembership_FunctionGaussian
Number_of_MFs3
Number_of_Inputs2
Iterations100
Table 3. Operating system specifications.
Table 3. Operating system specifications.
ProcessorIntel(R) Core(TM) i9-10900K
CPU Frequency3.70GHz
RAM64.0 GB (63.9 GB usable)
Operating SystemWindows 10 Education
Version 22H2
System Type64-bit, x64-based processor
Programming LanguagePython 3
Programming EnvironmentAnaconda
Table 4. Rice seed variety classification results.
Table 4. Rice seed variety classification results.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.50110.57890.51020.5316
Linear SVM0.71330.75420.72180.7338
RBF SVM0.13330.02220.16670.0392
Gaussian Process0.13780.18900.17220.0501
Decision Tree0.49220.55790.50740.4871
Random Forest0.33560.47970.33240.3303
Neural Net0.74780.76860.75460.7601
AdaBoost0.53440.57930.55460.5622
Naive Bayes0.41220.50240.44350.3383
QDA0.40220.50990.3990.3854
XGBoost0.76330.79180.76850.7769
CatBoost0.62880.68180.63510.6501
LightGBM0.73660.77160.74120.7517
Proposed model0.76970.79490.77070.7862
Table 5. Age classification of Akitakomachi rice seeds.
Table 5. Age classification of Akitakomachi rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.55560.55950.55560.5569
Linear SVM0.66670.66760.66670.6651
RBF SVM0.33330.11110.33330.1667
Gaussian Process0.35000.44700.35000.2012
Decision Tree0.65000.65880.65000.6498
Random Forest0.55000.55950.55000.5490
Neural Net0.66670.66820.66670.6656
AdaBoost0.63330.62970.63330.6283
Naive Bayes0.54440.59180.54440.5234
QDA0.44440.63890.44440.3605
XGBoost0.75000.75070.75000.7483
CatBoost0.67220.67280.67220.6686
LightGBM0.71110.71240.71110.7087
Proposed model0.75510.75790.75520.7556
Table 6. Age classification of Fusaotome rice seeds.
Table 6. Age classification of Fusaotome rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.72780.72320.72780.7224
Linear SVM0.82780.82750.82780.8266
RBF SVM0.33330.11110.33330.1667
Gaussian Process0.34440.77900.34440.1899
Decision Tree0.65560.65100.65560.6505
Random Forest0.62220.66170.62220.6157
Neural Net0.82220.82120.82220.8209
AdaBoost0.79440.79600.79440.7930
Naive Bayes0.55560.71080.55560.4528
QDA0.50560.71930.50560.4715
XGBoost0.86110.86220.86110.8608
CatBoost0.86110.86430.86110.8615
LightGBM0.85560.85630.85560.8556
Proposed model0.86120.87200.86180.8616
Table 7. Age classification of Hatsuboshi rice seeds.
Table 7. Age classification of Hatsuboshi rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.61670.61810.61670.5947
Linear SVM0.80560.82130.80560.7957
RBF SVM0.33330.11110.33330.1667
Gaussian Process0.81110.82320.81110.8031
Decision Tree0.70560.70840.70560.6929
Random Forest0.57220.66760.57220.5174
Neural Net0.81670.82560.81670.8090
AdaBoost0.75560.76360.75560.7577
Naive Bayes0.60000.61780.60000.5562
QDA0.46110.67690.46110.3608
XGBoost0.79440.81720.79440.7838
CatBoost0.76110.79060.76110.7458
LightGBM0.77780.79020.77780.7650
Proposed model0.79890.81930.79760.7912
Table 8. Age classification of Koshihikari rice seeds.
Table 8. Age classification of Koshihikari rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.60000.59600.60000.5882
Linear SVM0.81670.81310.81670.8118
RBF SVM0.33330.11110.33330.1667
Gaussian Process0.82780.83800.82780.8227
Decision Tree0.68330.67590.68330.6772
Random Forest0.61670.66810.61670.5673
Neural Net0.85560.85900.85560.8525
AdaBoost0.72780.73320.72780.7302
Naive Bayes0.46110.51580.46110.3832
QDA0.55560.38120.55560.4479
XGBoost0.87220.87290.87220.8692
CatBoost0.72780.74510.72780.6904
LightGBM0.85560.85510.85560.8511
Proposed model0.88150.88060.88100.8693
Table 9. Age classification of Okiniiri rice seeds.
Table 9. Age classification of Okiniiri rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.89170.90530.89170.8907
Linear SVM0.95000.95000.95000.9500
RBF SVM0.50000.25000.50000.3333
Gaussian Process0.89170.90530.89170.8907
Decision Tree0.90000.90000.90000.9000
Random Forest0.90000.90720.90000.8996
Neural Net0.95000.95200.95000.9499
AdaBoost0.95830.95850.95830.9583
Naive Bayes0.89170.90530.89170.8907
QDA0.50830.75210.50830.3516
XGBoost0.95000.95050.95000.9500
CatBoost0.94170.94280.94170.9416
LightGBM0.95830.95850.95830.9583
Proposed model0.95120.95210.95140.9512
Table 10. Age classification of Yang DAO-8 rice seeds.
Table 10. Age classification of Yang DAO-8 rice seeds.
AlgorithmAccuracyPrecision μ Recall μ F1-Score μ
Nearest Neighbors0.61670.66300.61670.5873
Linear SVM0.72500.77230.72500.7125
RBF SVM0.50000.25000.50000.3333
Gaussian Process0.65000.69180.65000.6298
Decision Tree0.63330.67780.63330.6089
Random Forest0.54170.65020.54170.4406
Neural Net0.75000.78130.75000.7429
AdaBoost0.70830.77180.70830.6902
Naive Bayes0.51670.54040.51670.4334
QDA0.50000.25000.50000.3333
XGBoost0.75000.79760.75000.7396
CatBoost0.67500.78240.67500.6409
LightGBM0.75000.81960.75000.7356
Proposed model0.76390.80620.76650.7479
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Rathnayake, N.; Miyazaki, A.; Dang, T.L.; Hoshino, Y. Age Classification of Rice Seeds in Japan Using Gradient-Boosting and ANFIS Algorithms. Sensors 2023, 23, 2828. https://doi.org/10.3390/s23052828

AMA Style

Rathnayake N, Miyazaki A, Dang TL, Hoshino Y. Age Classification of Rice Seeds in Japan Using Gradient-Boosting and ANFIS Algorithms. Sensors. 2023; 23(5):2828. https://doi.org/10.3390/s23052828

Chicago/Turabian Style

Rathnayake, Namal, Akira Miyazaki, Tuan Linh Dang, and Yukinobu Hoshino. 2023. "Age Classification of Rice Seeds in Japan Using Gradient-Boosting and ANFIS Algorithms" Sensors 23, no. 5: 2828. https://doi.org/10.3390/s23052828

APA Style

Rathnayake, N., Miyazaki, A., Dang, T. L., & Hoshino, Y. (2023). Age Classification of Rice Seeds in Japan Using Gradient-Boosting and ANFIS Algorithms. Sensors, 23(5), 2828. https://doi.org/10.3390/s23052828

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop