Next Article in Journal
Seasonal Changes in Amylose and Starch Compositions in ‘Ambrosia’ Apples Associated with Rootstocks and Orchard Climatic Conditions
Previous Article in Journal
Exploring the Impact of Metal-Based Nanofertilizers: A Case Study on Sunflower Pollen Morphology and Yield in Field Conditions
Previous Article in Special Issue
Collaborative Wheat Lodging Segmentation Semi-Supervised Learning Model Based on RSE-BiSeNet Using UAV Imagery
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved MobileNetV3 Mushroom Quality Classification Model Using Images with Complex Backgrounds

College of Engineering and Technology, Jilin Agricultural University, Changchun 130118, China
*
Author to whom correspondence should be addressed.
Agronomy 2023, 13(12), 2924; https://doi.org/10.3390/agronomy13122924
Submission received: 27 October 2023 / Revised: 24 November 2023 / Accepted: 25 November 2023 / Published: 28 November 2023
(This article belongs to the Special Issue Applications of Deep Learning in Smart Agriculture—Volume II)

Abstract

:
Shiitake mushrooms are an important edible fungus, and their nutrient content is related to their quality. With the acceleration of urbanization, there has been a serious loss of population and shortage of labor in rural areas. The problem of harvesting agricultural products after maturity is becoming more and more prominent. In recent years, deep learning techniques have performed well in classification tasks using image data. These techniques can replace the manual labor needed to classify the quality of shiitake mushrooms quickly and accurately. Therefore, in this paper, a MobileNetV3_large deep convolutional network is improved, and a mushroom quality classification model using images with complex backgrounds is proposed. First, captured image data of shiitake mushrooms are divided into three categories based on the appearance characteristics related to shiitake quality. By constructing a hybrid data set, the model’s focus on shiitake mushrooms in images with complex backgrounds is improved. And the constructed data set is expanded using data enhancement methods to improve the generalization ability of the model. The total number of images after expansion is 10,991. Among them, the number of primary mushroom images is 3758, the number of secondary mushroom images is 3678, and the number of tertiary mushroom images is 3555. Subsequently, the SE module in MobileNetV3_large network is improved and processed to enhance the model recognition accuracy while reducing the network size. Finally, PolyFocalLoss and migration learning strategies are introduced to train the model and accelerate model convergence. In this paper, the recognition performance of the improved MobileNetV3_large model is evaluated by using the confusion matrix evaluation tool. It is also compared with other deep convolutional network models such as VGG16, GoogLeNet, ResNet50, MobileNet, ShuffleNet, and EfficientNet using the same experimental conditions. The results show that the improved MobileNetV3_large network has a recognition accuracy of 99.91%, a model size of 11.9 M, and a recognition error rate of 0.09% by the above methods. Compared to the original model, the recognition accuracy of the improved model is increased by 18.81% and the size is reduced by 26.54%. The improved MobileNetV3_large network model in this paper has better comprehensive performance, and it can provide a reference for the development of quality recognition and classification technologies for shiitake mushrooms cultivated in greenhouse environments.

1. Introduction

Shiitake mushrooms are the second leading edible fungus in the world and are loved by people because of their unique aroma, rich nutritional value, and delicious taste [1]. According to statistics, the production of shiitake mushrooms in China in 2021 reached 12.9572 million t, accounting for 31.34% of the total amount of cultivated edible mushrooms, and it was the main edible fungus produced in China in 2021 [2]. Nevertheless, there are still problems to overcome regarding the extensive production of these mushrooms. Since shiitake mushrooms have different nutrient content due to differing quality, they need to be effectively classified. Traditional manual sorting [3] is time-consuming and laborious, and the benefits of this strategy are minimal. With the aging of our country’s population and the serious loss of the rural population, relying on traditional manual sorting is difficult to sustain long term. In addition, shiitake mushrooms of differing quality have different sorting criteria in terms of shape and color, and it is difficult to screen out fresh shiitake mushrooms of excellent quality by relying on traditional machinery and equipment. Therefore, it is necessary to propose a classification method for effectively identifying the quality of shiitake mushrooms.
In recent years, with the rapid development of deep learning technology, Yonis Gulzar et al. conducted a classification study of sunflower diseases using a public data set containing a total of 1892 images of healthy and infected sunflower leaves and flowers. They investigated the classification of sunflower diseases using five deep learning techniques. It was shown that models such as MobileNetV3 and EfficientNetB3 provide high performance while requiring relatively few training cycles [4]. Yonis Gulzar et al. proposed a deep learning method based on Mobilenet_2 as an improved model utilizing transfer learning for training. In all, 40 different fruits were classified and recognized. The results showed that the proposed model achieved the highest accuracy in recognizing different types of fruits. The proposed model can be deployed in mobile applications for practical use [5]. Poonam Dhiman et al. systematically reviewed papers on prediction, detection, and classification of citrus fruit diseases using machine learning, deep learning, and statistical techniques. The best techniques that were found to be superior to others in their respective categories were as follows: SVM in ML methods, ANN in neural network networks, CNN in deep learning methods, and linear discriminant analysis (LDA) in statistical techniques [6]. Normaisharah Mamat et al. focused their research in the field of agriculture on the use of automated image annotation techniques for oil palm fruit maturity classification. They developed a model using deep learning combined with a transfer learning technique to train the model based on a data set. The results showed that the annotation technique successfully and accurately labeled a large number of images [7]. Liu et al. used the improved YOLOX deep learning method to detect the surface texture of shiitake mushrooms, and the model size was reduced by more than half [8]. Zhang et al. combined deep learning methods with spectral analysis to propose an end-to-end spectral qualitative analysis model based on convolutional neural networks that was used to effectively identify 20 grape varieties [9]. Wang et al. proposed an improved potato soil and stone detection method combined with YOLOv4, which realized the rapid detection and effective removal of soil and stones from debris-containing potatoes after harvest [10] Andre Dantes de Medeiros et al. combined X-ray images and deep learning network models to achieve 91%, 95%, and 82% recognition accuracy in terms of seed internal tissue integrity, germination rate, and vitality, respectively [11]. At the same time, the recognition model was gradually developed in the direction of lightweight. Zhang et al. developed a stud posture detection system to achieve rapid detection and high-precision positioning of studs [12]. Zou et al. proposed a deep imitation enhanced learning (DIRL) framework to effectively solve the problem of low efficiency of deep learning when exploring large continuous motion spaces in the field of autonomous driving [13].
Deep learning has been widely applied and developed in the field of visual recognition, but it is less applied in mushroom quality recognition and classification. Accordingly, this paper proposes a mushroom quality recognition classification model based on an improved MobileNetV3_large network [14] to address the above problems. Due to the complex environment of greenhouse mushroom cultivation, complex backgrounds in captured images cause serious interference in the recognition of mushroom features by the model. For this reason, this paper proposes a new method for constructing a data set, which effectively alleviates this problem. The recognition performance of the model directly affects the harvesting efficiency of shiitake mushrooms. In this paper, we improve the recognition performance of the model through simple changes while reducing the network size. In order to save the time cost and arithmetic power for training the model, this paper utilizes an improved training strategy instead of the original model training method. The above method is expected to represent a lightweight, simple, and efficient classification model for mushroom quality recognition. The model can be deployed to mobile devices, such as cell phones, computers, artificial intelligence robots, etc., in practical scenarios. It can provide valuable technical support for the application and development of shiitake mushroom harvesting strategies in the field of precision smart agriculture.
The following points summarize the contributions of this paper:
  • A new method for constructing data sets to enhance the model’s focus on subjects in complex background images is proposed.
  • A mushroom quality classification strategy based on the MobilenetV3_large network model is introduced to classify mushrooms of different quality levels.
  • An improved MobilenetV3_large network model for mushroom quality classification is proposed based on different training strategies to improve the recognition accuracy of the model while reducing the time cost and arithmetic power spent on training the model, such as data enhancement techniques, migration learning techniques, and replacing the loss function with a better one.
  • The recognition performance of the model after replacing the SE attention mechanism in the MobilenetV3_large network model is compared with CBAM, CA, scSE, and improved SE.
  • The recognition performances for mushroom quality classification of eight other popular deep learning models are compared with that of the improved MobilenetV3_large network model.

2. Data Collection and Processing

2.1. Data Source and Division

The image data for this test were collected from Wulipu Village, Shancheng Town, Meihekou, Jilin, China, on 30 June 2023. The acquisition device was a Huawei Nova 3i smartphone (Produced by Huawei manufacturer in Shenzhen, China, the software version number is 2.0.0.263 (C01E261R1P3)), which generates images of 3456 × 4608 pixels, and the picture format was jpg. The shooting method was horizontal from the side, and a total of 275 original images of shiitake mushrooms were collected. Local shiitake mushroom quality grading standards and harvesting principles combined with the guidance suggestions from the planting experts at this location were employed. Finally, the image data for shiitake mushrooms were categorized into three classes, including Grade I, Grade II, and Grade III. Because of the obvious differences in the appearance characteristics of different grades of shiitake mushrooms, the grades of shiitake mushrooms were classified according to the degree of openness of the mushroom gills, the curvature of the outline of the cap, and whether defects occurred in the surface shape. See Table 1 for details.
In order to avoid model interference caused by multiple shiitake mushrooms in the image, the sample picture was appropriately cropped in this study, and only one shiitake mushroom body was retained in each image. Example images are shown in Figure 1.
In order to improve the recognition accuracy and generalization ability of the model and to enhance the robustness of the model, the background data set, a mixed data set, and an enhanced data set for removing shiitake mushrooms were also constructed in this experiment, and each data set was divided into a training set, verification set, and test set according to the 8:1:1 ratio. The number of images in each data set is shown in Table 2.

2.2. Data Processing

2.2.1. Background Segmentation

Since the background of shiitake mushroom images is complex, in order to reduce the interference of complex background information on model training and improve the recognition accuracy of the model, the background of the shiitake mushroom images was divided and removed in this experiment. Example mushroom images before and after removing the background are shown in Figure 2.

2.2.2. Building a Hybrid Data Set

The shiitake mushrooms in the images with complex backgrounds and the images with the background removed represent the same type of shiitake mushroom when observed by the human eye, which stems from the subjective attention of humans on the shiitake mushrooms in the images. For computers, the shiitake mushrooms in the images with complex background and those with the background removed belong to two different types of images. Thus, complex background information may be recognized by the computer as a feature of the image, which in turn affects the training of the model.
In order to minimize the influence of complex background information on the model and increase the model’s focus on mushroom ontology, this experiment constructs a hybrid data set to improve the accuracy of recognition by the model. In this paper, the data set composed of mushroom images with complex backgrounds and those with the background removed is defined as the hybrid data set. In this way, after the model recognizes two images with the same subject, it may pay more attention to the shiitake mushroom features in the image and reduce the interference of complex background information. Some of the images in the hybrid data set are shown in Figure 3.

2.2.3. Data Enhancement

In order to further improve the recognition performance of the model, enhance the generalization ability of the model, and reduce the phenomenon of overfitting, in this experiment, a data enhancement strategy [15] was used to expand the mushroom image data in the mixed data set. Random rotation, random translation, random flipping, and random brightness transformation were used to expand the original data to 20 times the original, totaling 10,991 shiitake mushroom images. Among them, random rotation involved randomly rotating the image data in the range of 0~360°; random translation involved randomly translating the image data in the horizontal and vertical directions; random flipping involved randomly flipping the image data in the horizontal and vertical directions; and random brightness transformation simulated the color characteristics of shiitake mushrooms under different brightness levels by randomly adjusting the RGB value of the input picture between 50 and 100. Examples of the expanded shiitake mushroom images are shown in Figure 4.

3. Construction of a Quality Classification Model for Shiitake Mushrooms

3.1. MobileNetV3-Large Network Model

The MobileNetV3 network improves on the MobileNetV2 network [16] by using a depth-separable convolutional structure and an inverted residual structure to ensure the model is lightweight; additionally, the SE module [17] and h-swish activation function are introduced, and the time-consuming layer structure in the bottleneck is redesigned, further improving the recognition accuracy of the model.
In this paper, the MobileNetV3-Large model is used, which consists of several components, including four 2D convolutional layers, two bottleneck layers (112 × 112), two bottleneck layers (56 × 56), three bottleneck layers (28 × 28), six bottleneck layers (14 × 14), two bottleneck layers (7 × 7), and one pooling layer (7 × 7). The activation functions used are Swish and ReLU. Also, the SE attention mechanism is embedded in the bneck structure in this network to improve the recognition performance and reduce the space occupied by the network. The SE mechanism assigns different weights to different channels of the feature map by compressing the excitation operation. Finally, a 2D convolutional layer (1 × 1) is used to obtain the feature vector to achieve the final classification output.
Table 3 shows the specific structure of the model. Input denotes the size of the input image. Operator denotes each step of the model’s operation on the feature map. exp size denotes the number of output channels of the extended layer. #out denotes the number of output channels of the feature map after passing through each layer of the network. SE denotes whether this attention mechanism is added to each layer of the network. NL denotes the activation function used by each layer of the network. s denotes the step size of the network operation in each layer.

3.2. Improving the MobileNetV3-Large Network Model

The recognition model in this paper is mainly based on the MobileNetV3_large network architecture, in which some of the network layer structures in the SE module of the bneck structure are replaced and the parameters are tuned. The recognition performance of the improved MobileNetV3_large network model is slightly improved, the occupied space is slightly reduced, and the number of network layers remains unchanged after the improvement. Figure 5 shows the schematic diagram of the network structure after the model improvement.

3.2.1. Improved SE Module

In order to strengthen the relationship information between channels, in 2018, Hu et al. [17] first proposed the SE module, which is flexible in operation and simple in structure and can significantly improve the network performance by adding a small number of network parameters. The SE module is mainly composed of two parts, Squeeze and Excitation. First, the Squeeze operation is performed on the input feature map using global average pooling to compress the global spatial information into channel descriptors; then, the importance of each channel in the obtained aggregated information is predicted using the fully connected layer of Excitation to obtain the importance among different channels; finally, the channel weights of the output of Excitation are weighted onto the previous features channel by channel by multiplication channel-by-channel weighting to the previous features to complete the recalibration of the original features in the channel dimension.
In the MobileNetV3-Large network model, the authors embedded 8 SE modules in 15 bneck structures, so the influence of the SE mechanism on the network recognition performance and network size is significant, and optimizing the structure and parameters of the SE mechanism is especially important for improving the network performance. Therefore, this paper improves the SE modules in the bneck structure of the MobileNetV3-Large network model.
As the fully connected layer mainly compresses the features of the image, the convolutional layer mainly extracts the features from the image. After the image has gone through multiple convolutional layers, the features in the output feature map become more advanced. And after the feature map has gone through multiple fully connected layers, the features are compressed more centrally, and the weight that each feature has becomes more important. As the background of the mushroom image in this paper is more complex, the diverse backgrounds may also be recognized by the computer as multiple features of the image, which is obviously unfavorable for the recognition of the model. It is feasible to distinguish the grade of mushroom quality by shallow appearance features, and observing more advanced abstract features may be unnecessary. Therefore, the recognition model in this paper needs to filter out as many complex and diverse background features as possible so that the model can focus more on the more concentrated and important ontological features of shiitake mushrooms in the image. Then, replacing the convolutional layer with a fully connected layer may be a good choice. Therefore, in this paper, the convolutional layer in the SE module is replaced with a fully connected layer to enhance the output channel weights.
After replacing the convolutional layer with a fully connected layer, the network increases the number of computational parameters by a small amount in order to further improve the performance of the model and reduce the space occupied by the model in practical applications. In this paper, we minimize the size of the network by adjusting the compression multiplier of the channels in the SE module from 4× to 16× while ensuring model recognition accuracy. Compared with the original model, the improved model is smaller in size.
The SE module after the improvement allows for better recognition accuracy of the model, while making the model smaller. The structure of SE before and after improvement is shown in Figure 6.

3.2.2. PolyLoss

In order to be more flexible and applicable for the classification tasks of this experiment, the PolyLoss loss function is introduced. Inspired by Taylor’s expansion to approximate the function, Leng Zhaoqi and others proposed PolyLoss, designing and treating the loss function as a linear combination of polynomial functions [18]. PolyLoss can adjust the weight of the polynomial loss through a simple strategy and a line of code to reduce the difficulty of adjusting the hyperparameters. At the same time, the cross-entropy loss function (CrossEntropyLoss) and focus loss function (FocalLoss) commonly used in the classification task of training deep neural networks can also be used as special cases of PolyLoss, and the application is more flexible.
The PolyLoss loss function decomposes the commonly used classification loss functions (such as cross-entropy loss and FocalLoss loss) into a series of weighted polynomials through Taylor expansion; the decomposition form is:
j = 1 α j 1 P t j
where (1) α j ∈R+ represents the weight of the polynomial loss;
(2) P t represents the probability of the target label prediction.
When α j   =   1 j , PolyLoss is equivalent to cross-entropy loss. Simply adjusting the polynomial coefficient α j according to the unused tasks and data sets can achieve better results than the cross-entropy loss function and the focus loss function. The specific formula is as follows:
L Poly = α 1 1 P t + α 2 1 P t 2 + α N 1 P t N + = j = 1 α j 1 P t j
In most cases, adjusting the first polynomial coefficient has a significant gain on the model, and the formula is as follows:
L Poly 1 = 1 + ε 1 1 P t + 1 2 1 P t 2 + = log P t + ε 1 1 P t
After testing, when ε 1 = 1, the model training effect is optimal, so this test sets the value of the superparameter ε 1 of the introduced PolyLoss loss function to 1.

3.3. Migration Learning

In order to enhance the generalization ability of the model, improve the recognition accuracy of the model, and reduce the time cost, data volume, and computing power required for training neural networks, this study uses a combination of a migration learning strategy and the improved MobileNetV3-Large network model to construct a mushroom quality classification network model. The migration learning strategy can effectively alleviate the overfitting problem caused by the small amount of sample data in the deep network model and can speed up the convergence of the model and improve the efficiency of the training model [19].

4. Test Results and Evaluation

4.1. Test Environment

The training of the network model in this experiment is based on the pytorch deep learning framework. The test hardware environment is as follows: the CPU uses Intel (R) Core (TM) i7-12700F clocked at 2.10 GHz, the memory is 32 GB, the GPU uses NVIDIA GeForce RTX4070, and the video memory capacity is 12 GB. Using the Windows10 operating system, the python version is 3.10.
In this experiment, the hyperparameters of the network are balanced and tuned. The total number of training rounds (epochs) is set to 60; the number of training batches per round (batch_size) is set to 64; and the learning rate affects the update speed of the network weights. When the learning rate is set to 1 × 10−4, the training effect of the model is the best. By default, the Adam optimization algorithm [20] is used. The algorithm iteratively updates the network weights based on the training data, which is simple to implement, computationally efficient, and requires less memory.

4.2. Evaluation Indicators

In order to objectively evaluate the effectiveness of the test method, four evaluation indicators [21], namely, the recognition accuracy rate, accuracy rate, recall rate, and F1 score, were used to evaluate the improved MobileNetV3-Large network model proposed in this article.

4.3. The Impact of Data Processing on the Accuracy of the Model

Figure 7 shows the accuracy training process of the original MobileNetV3-large network model on different data sets. It can be seen that training on the original image data set has the worst effect on improving the recognition accuracy of the model. After removing the background data set training, the recognition performance of the model improves slightly. After training from the mixed data set, the accuracy rate of the model is significantly faster than that of the original data set and the background image data set. Using data enhancement strategies to expand the mixed data set, the training process of the model is smoother, and the recognition performance is significantly improved.
Table 4 shows the test results of training the model using different data sets. It can be seen that the model recognition effect trained from the data set constructed from the original picture is poor, and the test accuracy rate is 81.1%. Using a data set that removes the background of the image to train the model, the recognition accuracy of the model is improved, and the test accuracy rate is 86.8%. Using the mixed data set to train the model, the recognition accuracy rate of the model is further improved to 90.7%. Using the data enhancement strategy to expand the mixed data set, the recognition accuracy rate in the test set reaches 97.2%, and the model recognition effect obtained after training is the best.
The number of pictures in the data set constructed from the original picture and the image with the background removed is small, resulting in fewer pictures for testing after the model is trained. In order to objectively compare the effects of different data sets on the training model, this paper uniformly uses the test set divided by the mixed data set to perform model tests on the original picture data set, the removed background data set, and the model trained on the mixed data set. It can be seen that the accuracy rate of the model for the background removed data set and training on the mixed data set is 89.1%, and the recognition accuracy rate of the model trained with the background data set after the unified test is 3.91% lower than that of the model trained with the mixed data set, indicating that the use of the mixed data set to train the model has a better effect. After the mixed data set was enhanced, the training accuracy rate and the test accuracy rate were significantly improved, and the test accuracy rate reached 97.18%, indicating that the use of the mixed data set plus the data enhancement strategy has a better effect on model training.

4.4. Comparative Test of Loss Function

Figure 8 shows the test results obtained by training the model using different loss functions. It can be seen that the convergence effect when using PolyCrossEntropyLoss to train the model is the worst, the loss value is higher during the training process, and the convergence speed is slower. The loss value of the training model using the cross-entropy loss function is reduced, but the convergence speed is slower. The loss value of the FocalLoss training model is further reduced, the convergence speed is faster, and the training effect is better. Using PolyFocalLoss to train the model has the lowest loss value, faster convergence, and the best training effect.
Table 5 shows the results of training the original MobileNetV3-large network under different loss functions. It can be seen that the cross-entropy loss function of the original model is used to train the model. The recognition accuracy rate of the model is 97.18%, the accuracy rate is 97.17%, the recall rate is 97.2%, and the F1 score is 97.18. After using PolyCrossEntropyLoss to train the model, the model recognition accuracy rate increases by 0.63%, the accuracy rate and recall rate are both 97.83%, and the F1 score is 97.83. Using the focus loss function (FocalLoss) to train the model, the recognition performance of the model is further improved, the test accuracy rate reaches 98.18%, and the accuracy rate, recall rate, and F1 score are all improved. After PolyFocalLoss is used to train the model, the recognition accuracy rate of the model is the highest, reaching 98.45%, the accuracy rate is 98.48%, the recall rate is 98.45%, and the F1 score is 98.46.
It can be seen that the loss value of the model during training under PolyCrossEntropyLoss is higher, but the recognition accuracy rate of the model after training under PolyCrossEntropyLoss is higher than that of the model trained under CrossEntropyLoss, indicating that PolyCrossEntropyLoss is more suitable for the distribution of the data set constructed in this experiment, which can better return the error and correct the weight in the process of back propagation. Using PolyFocalLoss to train the model has the best effect, increasing the recognition accuracy of the model by 1.27%, the accuracy rate and recall rate by 1.31% and 1.35%, respectively, and the F1 score by 1.28. It can be shown that improving the loss function is beneficial for improving the recognition performance of the model.

4.5. Comparative Test of the Attention Mechanism

Figure 9 shows the comparison test of the SE module and migration learning. It can be seen that the improved SE module improves the accuracy of the model training process more obviously. At the same time, when the model is trained without migration learning, the accuracy curve and loss curve of the training process oscillate obviously and converge slowly. When the model is trained with migration learning, the accuracy curve and loss curve during training are smooth and stable, and the convergence speed is faster.
In this paper, the attention mechanisms of CBAM [22], CA [23], and scSE [24] are also used to replace the original SE module, and the recognition performance of the model is compared through experiments. Table 6 lists the test results of different attention modules and migration learning strategies on the performance of model recognition. As can be seen from the ①② experiment, the use of the migration learning strategy helps to improve the recognition accuracy of the model, and the model recognition accuracy rate increases by 1.27%. As can be seen from the ②③④⑤⑥ test, the improved SE module has the most obvious effect on the performance of the model. The recognition accuracy of the model on the test set reaches 99.91%, which is 1.46% higher than before the improvement, and the model size is reduced by 4.3 M. By improving the attention SE mechanism in the original MobileNetV3-large network and introducing a migration learning strategy, the recognition accuracy rate of the model is 2.73% higher than before the improvement, while the model size is reduced by 26.54%, indicating that the improvement method of this experiment is optimal.

4.6. Improved Model Evaluation

Figure 10 shows the confusion matrix of the original MobileNetV3-large network and the improved MobileNetV3-large network for the quality classification of shiitake mushrooms. The number of samples misclassified by the original network is 31, and the number of samples misclassified by the improved network is 1. Due to the different forms of third-class shiitake mushrooms, which include some second-class damaged shiitake mushrooms, when the damage is not obvious, it is easy to be confused with second-class shiitake mushrooms, so a prediction error occurs. The comparison shows that the improved MobileNetV3-large network classification effect is significantly improved.
As can be seen from Table 7, the model has a good effect on the recognition of various grades of shiitake mushrooms. The recognition accuracy rate, accuracy rate, and recall rate are all higher than 99.72%, and the F1 score is higher than 99.86. Among them, the recognition accuracy rate, accuracy rate, and recall rate of first-class shiitake mushrooms are all 100%, and the F1 score is 1. It can be shown that the improved model has high accuracy and good stability for the grade recognition of shiitake mushrooms in this test and that it can provide a reference for agricultural identification of shiitake mushroom grades.

4.7. Comparison of Recognition Performance of Different Models

In order to further verify the recognition performance of the improved model, this experiment compares the recognition results of different network models. As can be seen in Table 8, the recognition accuracy rate of the VGG16 network [25] is only 69.79%, the model size is 512 M, and the overall performance is poor. GoogLeNet [26], ResNet50 [27], and MobileNetV1 [28] exhibit slightly improved recognition accuracies, but the model size is larger than the improved network, and the overall performance is not high. Both the ShuffleNetV2×1 network [29] and the MobileNetV2 network are smaller in scale than the improved MobileNetV3 network, but the accuracy rate, accuracy rate, recall rate, and F1 score are significantly reduced. The recognition accuracy of EfficientNetV2-s [30] is higher than that of the original MobileNetV3-Large and slightly lower than that of the improved model, but the network scale is not dominant. In summary, the improved MobileNetV3 network recognition in this paper has high accuracy and a small network scale, which is suitable for mushroom quality classification tasks.

4.8. Visual Result Verification

The method shown in Figure 11 is implemented by gradient-weighted class activation mapping (Grad-CAM) [31]. After the improved MobileNetV3-large network model is recognized, the characteristic extraction and comparison results of each grade of shiitake mushrooms are obtained. It can be seen that the characteristics of the cap part of first-class shiitake mushrooms have received high attention from the model; the junction of the cap and the stipe of second-class shiitake mushrooms are the area of interest to the model; and the defective part of third-class shiitake mushrooms is the key location for the model to identify. The visualization results of feature extraction are consistent with the quality classification standards for shiitake mushrooms.

4.9. Improvement of Mobilenet_V3 Compared to State-of-the-Art Methods

Table 9 demonstrates the results of comparing the recognition model in this paper with the current state-of-the-art recognition models. It can be seen that for different data sets, the authors have designed different recognition models, respectively. The data sets include Winter Jujube, Wheat, Mushrooom, Sauerkraut, and Pear. The recognition models are iResnet-50, D-VGG, and the recognition model with CNN architecture designed by the authors. Each designed model achieved better recognition results. After comparison, it can be seen that the improved MobilnetV3 designed in this paper performs well for mushroom quality classification. In addition, the recognition model designed in this paper is more robust because the collected data are captured in a real complex environment.

5. Conclusions

Deep learning techniques are widely used in precision agriculture. These techniques are used in various fields such as fruit and vegetable classification, face recognition, quality checking, yield estimation, disease prediction, etc. The success of these techniques has promoted the development of various types of deep learning models. This study is based on the MobilenetV3_large network architecture, which is improved and optimized. The application of recognizing and classifying the quality grades of shiitake mushrooms in the edible mushroom category is realized. By constructing a mixed data set, the model’s performance of recognizing mushroom subjects in a complex background is improved. Based on the original MobilenetV3_large network, the recognition accuracy of the model is improved, and the network size is reduced by smaller changes. By improving the training strategy, the recognition accuracy of the network is improved, and the training cost is reduced. The experimental results show that the improved MobileNetV3_large network in this paper has better overall performance, with a recognition accuracy of 99.91% and a model size of 11.9 M. It can be used as a reference for the development of recognition and classification technology for shiitake mushrooms cultivated in greenhouse environments.
In future work, the deployment and application of the recognition model in end devices, e.g., cell phones, computers, intelligent robots, etc., will be further enhanced. At the same time, more varieties of mushrooms will be used to achieve a wider range of mushroom quality recognition and classification. In addition, this experimental model can be employed in different target detection networks to replace the classification network among them and compare the recognition results to help realize a wider range of related application research.

Author Contributions

F.Z., Y.S., Y.Z. and J.Q. conceived and designed the experiments; F.Z., Y.S. and J.Q. performed the experiments; F.Z., Y.S. and J.Q. analyzed the data and wrote the original manuscript; F.Z., Y.S., Y.Z., W.Z. and J.Q. reviewed and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key Research and Development Program of China (2020YFD1000304-5), the Jilin Province Science and Technology Development Plan Project of China (20210202054NC) and the Jilin Province Science and Technology Development Plan Project of China (20220202029NC).

Data Availability Statement

The datasets generated and/or analyzed during the current study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Jorge, C.G.; Pardo, J.E.; Manuel, A.O.; Eulogio, L.M.; Zied, D.C.; Arturo, P.G. An intelligent mushroom strain selection model based on their quality characteristics. Food Biosci. 2023, 56, 103232. [Google Scholar]
  2. China Edible Mushroom Association (CEMA). Analysis of the survey results of the national edible fungus statistics in 2021. Chin. Edible Fungi 2023, 42, 118–127. (In Chinese) [Google Scholar]
  3. Reed, J.N.; Miles, S.J.; Butler, J.; Baldwin, M.; Noble, R. AE—Automation and Emerging Technologies: Automatic Mushroom Harvester Development. J. Agric. Eng. Res. 2001, 78, 15–23. [Google Scholar] [CrossRef]
  4. Gulzar, Y.; Ünal, Z.; Aktaş, H.; Mir, M.S. Harnessing the Power of Transfer Learning in Sunflower Disease Detection: A Comparative Study. Agriculture 2023, 13, 1479. [Google Scholar] [CrossRef]
  5. Gulzar, Y. Fruit Image Classification Model Based on MobileNetV2 with Deep Transfer Learning Technique. Sustainability 2023, 15, 1906. [Google Scholar] [CrossRef]
  6. Dhiman, P.; Kaur, A.; Balasaraswathi, V.R.; Gulzar, Y.; Alwan, A.A.; Hamid, Y. Image Acquisition, Preprocessing and Classification of Citrus Fruit Diseases: A Systematic Literature Review. Sustainability 2023, 15, 9643. [Google Scholar] [CrossRef]
  7. Mamat, N.; Othman, M.F.; Abdulghafor, R.; Alwan, A.A.; Gulzar, Y. Enhancing Image Annotation Technique of Fruit Classification Using a Deep Learning Approach. Sustainability 2023, 15, 901. [Google Scholar] [CrossRef]
  8. Liu, Q.; Fang, M.; Li, Y.S.; Gao, M.W. Deep learning based research on quality classification of shiitake mushrooms. LWT 2022, 168, 113902. [Google Scholar] [CrossRef]
  9. Liu, Y.S.; Zhou, S.B.; Han, W.; Liu, W.X.; Qiu, Z.F.; Li, C. Convolutional neural network for hyperspectral data analysis and effective wavelengths selection. Anal. Chim. Acta 2019, 1086, 46–54. [Google Scholar] [CrossRef]
  10. Wang, X.Y.; Li, Y.X.; Yang, Z.Y.; Zhang, M.; Wang, R.M.; Cui, L.X. Detection Method of Clods and Stones from Impurified Potatoes Based on Improved YOLO v4 Algorithm. Trans. Chin. Soc. Agric. Mach. 2021, 52, 241–247. [Google Scholar]
  11. Medeiros, A.D.; Bernardes, R.C.; Silva, L.J.; Freitas, B.A.L.; Santos Dias, D.C.F.; Silva, C.B. Deep learning-based approach using x-ray images for classifying crambe abyssinica seed quality. Ind. Crops Prod. 2021, 164, 113378. [Google Scholar] [CrossRef]
  12. Zhang, X.; Wang, G.H. Stud pose detection based on photometric stereo and lightweight YOLOv4. CAAI Trans. Intell. Technol. 2022, 2, 32–37. [Google Scholar] [CrossRef]
  13. Zou, Q.J.; Xiong, K.; Fang, Q.; Jiang, B.H. Deep imitation reinforcement learning for self-driving by vision. CAAI Trans. Intell. Technol. 2021, 6, 493–503. [Google Scholar] [CrossRef]
  14. Liu, Y.M.; Wang, Z.L.; Wang, R.J.; Chen, J.S.; Gao, H.J.; Jia, S.J.; Wang, P.; Jia, P.V. Flooding-based MobileNet to identify cucumber diseases from leaf images in natural scenes. Comput. Electron. Agric. 2023, 213, 108166. [Google Scholar] [CrossRef]
  15. Jia, S.J.; Wang, P.; Jia, P.Y.; Hu, S.P. Research on data augmentation for image classification based on convolution neural networks. In Proceedings of the 2017 Chinese Automation Congress (CAC), Jinan, China, 20–22 October 2017. [Google Scholar]
  16. Shi, C.; Xia, R.; Wang, L. A novel multi-branch channel expansion network for garbage image classification. IEEE Access 2020, 8, 154436–154452. [Google Scholar] [CrossRef]
  17. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  18. Leng, Z.Q.; Tan, X.M.; Liu, X.C.; Cubuk, E.D.; Shi, X.J.; Cheng, S.Y.; Anguelov, D. Polyloss: A polynomial expansion perspective of classification loss functions. arXiv 2022, arXiv:2204.12511. [Google Scholar]
  19. Pan, S.J.; Yang, Q. A survey on transfer learning. IEEE Trans. Knowl. Data Eng. 2010, 22, 1345–1359. [Google Scholar] [CrossRef]
  20. Fei, Z.G.; Wu, Z.Y.; Xiao, Y.Q.; Ma, J.; He, W.B. A new short-arc fitting method with high precision using Adam optimization algorithm. Optik 2020, 212, 164788. [Google Scholar] [CrossRef]
  21. Sun, J.; Zhu, W.D.; Luo, Y.Q. Recognizing the diseases of crop leaves in fields using improved Mobilenet-V2. Trans. CSAE 2021, 37, 161–169. (In Chinese) [Google Scholar]
  22. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional block attention module. In Proceedings of the European Conferenceon Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  23. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar]
  24. Roy, A.G.; Navab, N.; Wachinger, C. Concurrent spatial and channel‘squeeze & excitation’in fully convolutional networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Granada, Spain, 16–20 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 421–429. [Google Scholar]
  25. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 1–14. [Google Scholar]
  26. Christian, S.; Liu, W.; Jia, Y.Q.; Pierre, S.; Scott, R.; Dragomir, A.; Dumitru, E.; Vicent, V.; Andrew, R. Going deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  28. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.J.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
  29. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical guidelines for efficient cnn architecture design. In Proceedings of the 15th European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 122–138. [Google Scholar]
  30. Tan, M.; Le, Q. Efficientnetv2: Smaller Models and Faster Training. International conference on machine learning. PMLR 2021, 139, 10096–10106. [Google Scholar]
  31. Selvaraju, R.R.; Cogewell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual explanations from deep networks via gradient-based localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  32. Ban, Z.; Fang, C.; Liu, L.; Wu, Z.; Chen, C.; Zhu, Y. Detection of Fundamental Quality Traits of Winter Jujube Based on Computer Vision and Deep Learning. Agronomy 2023, 13, 2095. [Google Scholar] [CrossRef]
  33. Jiang, H.W.; Zhang, S.L.; Yang, Z.; Zhao, L.K.; Zhou, Y.; Zhou, D.X. Quality classification of stored wheat based on evidence reasoning rule and stacking ensemble learning. Comput. Electron. Agric. 2023, 214, 108339. [Google Scholar] [CrossRef]
  34. Wang, L.; Dong, P.H.; Wang, Q.; Jia, K.M.; Niu, Q.F. Dried shiitake mushroom grade recognition using D-VGG network and machine vision. Front. Nutr. 2023, 10, 1247075. [Google Scholar] [CrossRef]
  35. Du, J.; Zhang, M.; Teng, X.X.; Wang, Y.C.; Law, C.L.; Fang, D.C.; Liu, K. Evaluation of vegetable sauerkraut quality during storage based on convolution neural network. Food Res. Int. 2023, 164, 112420. [Google Scholar] [CrossRef]
  36. Yang, Z.; Li, Z.; Hu, N.; Zhang, M.; Zhang, W.; Gao, L.; Ding, X.; Qi, Z.; Duan, S. Multi-Index Grading Method for Pear Appearance Quality Based on Machine Vision. Agriculture 2023, 13, 290. [Google Scholar] [CrossRef]
Figure 1. Original shiitake mushroom images.
Figure 1. Original shiitake mushroom images.
Agronomy 13 02924 g001
Figure 2. Mushroom images before and after removing the background.
Figure 2. Mushroom images before and after removing the background.
Agronomy 13 02924 g002
Figure 3. Examples of shiitake mushroom images in the mixed data set.
Figure 3. Examples of shiitake mushroom images in the mixed data set.
Agronomy 13 02924 g003
Figure 4. Examples of shiitake mushroom images after data enhancement.
Figure 4. Examples of shiitake mushroom images after data enhancement.
Agronomy 13 02924 g004
Figure 5. Schematic diagram of the improved MobileNetV3-large network structure.
Figure 5. Schematic diagram of the improved MobileNetV3-large network structure.
Agronomy 13 02924 g005
Figure 6. The original SE module (left) and the improved SE module (right).
Figure 6. The original SE module (left) and the improved SE module (right).
Agronomy 13 02924 g006
Figure 7. Comparison of the recognition accuracy of models for different data sets.
Figure 7. Comparison of the recognition accuracy of models for different data sets.
Agronomy 13 02924 g007
Figure 8. Comparison of training loss under different loss functions.
Figure 8. Comparison of training loss under different loss functions.
Agronomy 13 02924 g008
Figure 9. Comparative test results of the SE module and migration learning.
Figure 9. Comparative test results of the SE module and migration learning.
Agronomy 13 02924 g009
Figure 10. Original MobileNetV3-large (a) and improved MobileNetV3-large (b) confusion matrix diagram.
Figure 10. Original MobileNetV3-large (a) and improved MobileNetV3-large (b) confusion matrix diagram.
Agronomy 13 02924 g010
Figure 11. Comparison results of characteristic extraction of various grades of shiitake mushrooms.
Figure 11. Comparison results of characteristic extraction of various grades of shiitake mushrooms.
Agronomy 13 02924 g011
Table 1. Shiitake Mushroom Quality Classification Standards.
Table 1. Shiitake Mushroom Quality Classification Standards.
Indicator LevelFirst ClassSecond ClassThird Class
The degree of opening of shiitake mushroomsHalf openFully open-
Mushroom cap contour curvatureFullFlat-
Appearance defectNoNoYes
Table 2. Statistical breakdown of the number of images in each data set.
Table 2. Statistical breakdown of the number of images in each data set.
Data DivisionOriginal Image/Removed Background ImageMixed Data SetEnhanced Data Set
First ClassSecond ClassThird ClassFirst ClassSecond ClassThird ClassFirst ClassSecond ClassThird Class
Training data set767472152148144300729432844
Verification data set999181818376368355
Test data set998181816375367356
Table 3. MobileNetV3-Large network structure.
Table 3. MobileNetV3-Large network structure.
InputOperatorExp Size#outSENLs
2242 × 3conv2d-16-HS2
1122 × 16bneck, 3 × 31616-RE1
1122 × 16bneck, 3 × 36424-RE2
562 × 24bneck, 3 × 37224-RE1
562 × 24bneck, 5 × 57240RE2
282 × 40bneck, 5 × 512040RE1
282 × 40bneck, 5 × 512040RE1
282 × 40bneck, 3 × 324080-HS2
142 × 80bneck, 3 × 320080-HS1
142 × 80bneck, 3 × 318480-HS1
142 × 80bneck, 3 × 318480-HS1
142 × 80bneck, 3 × 3480112HS1
142 × 112bneck, 3 × 3672112HS1
142 × 112bneck, 5 × 5672160HS2
72 × 160bneck, 5 × 5960160HS1
72 × 160bneck, 5 × 5960160HS1
72 × 160conv2d, 1 × 1-960-HS1
72 × 960pool, 7 × 7----1
12 × 960conv2d 1 × 1, NBN-1280-HS1
12 × 1280conv2d 1 × 1, NBN-k--1
Table 4. The results of training models with different data sets.
Table 4. The results of training models with different data sets.
Data SetAccuracy on the Training Data Set (%)Accuracy on the Test Set (%)
Original image81.481.1
Removed background89.186.8
Mixed data89.190.7
Mixed data enhanced97.397.2
Table 5. The results of the training model under different loss functions.
Table 5. The results of the training model under different loss functions.
Loss FunctionAccuracy (%)Precision (%)Recall (%)F1
CrossEntropyLoss97.1897.1797.2097.18
PolyCrossEntropyLoss97.8197.8397.8397.83
FocalLoss98.1898.2098.1898.19
PolyFocalLoss98.4598.4898.4598.46
Table 6. Test results of the effects of different attention modules and migration learning strategies on model recognition performance.
Table 6. Test results of the effects of different attention modules and migration learning strategies on model recognition performance.
Test NumberAttention ModuleWhether to Migrate to LearnAccuracy (%)Precision (%)Recall (%)F1Model Size (M)
1SEno97.1897.1997.2097.1916.2
2SEyes98.4598.4898.4598.4616.2
3CBAMyes99.6499.6499.6499.6411.9
4CAyes99.0999.1099.0999.0911.2
5scSEyes99.7399.7399.7399.7322.3
6Improved SEyes99.9199.9199.9199.9111.9
Table 7. Other evaluation indicators for improving the MobileNetV3-large network model.
Table 7. Other evaluation indicators for improving the MobileNetV3-large network model.
Grade CategoryAccuracy (%)Precision (%)Recall (%)F1
First class100.00100.00100.00100.00
Second class100.0099.72100.0099.86
Third class99.72100.0099.7299.86
Avg99.9199.9199.9199.91
Table 8. Comparison results of recognition performance of different models.
Table 8. Comparison results of recognition performance of different models.
ModelMigrate to LearnAccuracy (%)Precision (%)Recall (%)F1Model Size (M)
VGG16yes69.7673.8469.9271.83512
GoogLeNetno74.0482.0474.3578.0138
ResNet50yes82.2472.9082.2477.2990
MobileNetV1no90.8091.0990.8290.9512.3
MobileNetV2yes81.8882.0881.8981.988.73
MobileNetV3-Largeno97.1897.1997.2097.1916.2
ShuffleNetV2×1yes71.5871.8371.5871.704.95
EfficientNetV2-syes97.4597.4897.4597.4677.8
Improved MobileNetV3-Largeyes99.9199.9199.9199.9111.9
Table 9. Improvement of Mobilenet_V3 compared to state-of-the-art methods.
Table 9. Improvement of Mobilenet_V3 compared to state-of-the-art methods.
PaperData SetClassesImagesMethod/ModelAccuracy (%)Precision (%)Recall (%)
[32]Winter Jujube520,000iResnet-5098.3598.4098.35
[33]Wheat4108ER-Stacking88.1088.0589.31
[34]Mushrooom66775D-VGG96.2196.1896.33
[35]Sauerkraut32190CNN95.393.292.9
[36]Pear3398BP91.091.091.1
Improved MobilnetV3Mushrooom310,991MobilnetV399.9199.9199.91
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, F.; Sun, Y.; Zhang, Y.; Zhang, W.; Qi, J. An Improved MobileNetV3 Mushroom Quality Classification Model Using Images with Complex Backgrounds. Agronomy 2023, 13, 2924. https://doi.org/10.3390/agronomy13122924

AMA Style

Zhu F, Sun Y, Zhang Y, Zhang W, Qi J. An Improved MobileNetV3 Mushroom Quality Classification Model Using Images with Complex Backgrounds. Agronomy. 2023; 13(12):2924. https://doi.org/10.3390/agronomy13122924

Chicago/Turabian Style

Zhu, Fengwu, Yan Sun, Yuqing Zhang, Weijian Zhang, and Ji Qi. 2023. "An Improved MobileNetV3 Mushroom Quality Classification Model Using Images with Complex Backgrounds" Agronomy 13, no. 12: 2924. https://doi.org/10.3390/agronomy13122924

APA Style

Zhu, F., Sun, Y., Zhang, Y., Zhang, W., & Qi, J. (2023). An Improved MobileNetV3 Mushroom Quality Classification Model Using Images with Complex Backgrounds. Agronomy, 13(12), 2924. https://doi.org/10.3390/agronomy13122924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop