Next Article in Journal
True Parthenogenesis and Female-Biased Sex Ratios in Cicadomorpha and Fulgoromorpha (Hemiptera, Auchenorrhyncha)
Previous Article in Journal
Review of the Narrow-Banded Hawkmoth, Neogurelca montana (Rothschild & Jordan, 1915) (Lepidoptera: Sphingidae) in China, with Morphological and Phylogenetic Analysis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Feature Refinement Method Based on the Two-Stage Detection Framework for Similar Pest Detection in the Field

1
Science Island Branch of Graduate School, University of Science and Technology of China, Hefei 230026, China
2
Institute of Intelligent Machines, Hefei Institutes of Physical Science, Chinese Academy of Sciences, Hefei 230031, China
3
Institutes of Physical Science and Information Technology, Anhui University, Hefei 230039, China
4
Henan Yunfei Technology Development Co., Ltd., Zhengzhou 450003, China
*
Authors to whom correspondence should be addressed.
Insects 2023, 14(10), 819; https://doi.org/10.3390/insects14100819
Submission received: 29 August 2023 / Revised: 11 October 2023 / Accepted: 13 October 2023 / Published: 16 October 2023
(This article belongs to the Section Insect Pest and Vector Management)

Abstract

:

Simple Summary

The larvae of Lepidoptera pests are polyphagous insects that can cause crop mortality and severely damage crop growth, but the manual detection of such pests is a time-consuming and laborious task. We propose an automatic detection method to distinguish similar pests in the field. The proposed method is implemented based on the object detection framework, which improves the feature description ability of the network for different pests, optimizes suboptimal feature selection, and focuses the network head toward specific tasks. Our method achieves better detection results on a similar pest dataset compared with other advanced algorithms.

Abstract

Efficient pest identification and control is critical for ensuring food safety. Therefore, automatic detection of pests has high practical value for Integrated Pest Management (IPM). However, complex field environments and the similarity in appearance among pests can pose a significant challenge to the accurate identification of pests. In this paper, a feature refinement method designed for similar pest detection in the field based on the two-stage detection framework is proposed. Firstly, we designed a context feature enhancement module to enhance the feature expression ability of the network for different pests. Secondly, the adaptive feature fusion network was proposed to avoid the suboptimal problem of feature selection on a single scale. Finally, we designed a novel task separation network with different fusion features constructed for the classification task and the localization task. Our method was evaluated on the proposed dataset of similar pests named SimilarPest5 and achieved a mean average precision (mAP) of 72.7%, which was better than other advanced object detection methods.

1. Introduction

Crop pests cause serious harm to crop growth [1]. Accurate detection of different pests and the implementation of corresponding control measures can effectively improve crop yield and quality, which are crucial for agricultural production. One such destructive insect pest is the fall armyworm, Spodoptera frugiperda (Smith), flagged as a major concern by the Food and Agriculture Organization of the United Nations (FAO) [2,3]. Field investigations have revealed that S. frugiperda can easily be confused with other lepidopteran insect pests, such as Mythimna separate (Walker), Ostrinia furnacalis (Guenee), Helicoverpa armigera (Hübner), and Spodoptera litura (Fabricius), as they are similar in size, color, morphology, and living environment, particularly in their larval stage [4]. The similarity of these visual features poses a challenge for pest identification. Incorrect identification results lead to the use of unsuitable pesticides, which may not only jeopardize the growth of crops but also cause environmental pollution. Traditional pest identification tasks are completed by agricultural plant protection personnel through manual surveys, which are inefficient and susceptible to subjective factors. Fortunately, recent advancements in computer vision technology have provided new ideas and technical support for the automatic detection of some insect pests, which can effectively reduce the workload of professional plant protection personnel [5].
Traditional computer vision algorithms mainly use image processing and pattern recognition techniques for the feature extraction and classification of pest images. A large number of studies [6,7,8] have been conducted to extract color, shape, and texture features of pests and perform pest recognition tasks using support vector machines (SVMs) with certain results. In real-world outdoor scenarios, complex background environments are often present. Additionally, pest images are susceptible to various factors, such as differing light intensities, diverse weather conditions, and crop obstructions. Therefore, pest detection algorithms based on traditional computer vision face certain difficulties in adapting to complex field environments.
Compared with traditional machine learning techniques, deep learning techniques can fit the intrinsic characteristics of a large number of data well with a higher accuracy rate and strong robustness [9]. The detection-based method calculates the location and species of multiple pests in an image, which can reduce the interference of a complex background when the size of pests is relatively small in the image. The field of object detection is mainly divided into single-stage detection algorithms, represented by the YOLO series [10,11,12,13,14,15], and two-stage detection algorithms, represented by Faster RCNN [16]. Many advanced algorithms have been derived, such as Retinanet [17], Cascade RCNN [18], Double Head [19], etc. To improve the detection ability of multi-scale objects, the feature pyramid network (FPN) [20] was added into these detection frameworks as a common component. Compared with the fast single-stage algorithm, the two-stage algorithm has a slower speed but achieves higher accuracy.
Recently, many researchers have introduced deep learning technology into the field of agriculture. Many researchers [21,22,23] have achieved high recognition results on cropped datasets by using classification methods for pest recognition. For the detection of dense small-sized pests in complex environments, Li and Teng et al. [24,25] designed a coarse-to-fine network and a pyramid network to enhance the detection effect. Many researchers [26,27] constructed databases by fixed pest collection devices and utilized detectors such as YOLOv5 for pest detection. In order to enhance the discrimination ability of detectors for multiple categories of pests, feature fusion [28,29] was considered for algorithmic improvement, and it was experimentally demonstrated that feature fusion is effective in improving detection accuracy. Classification-based methods [21,22,23] focused on the global features of images. However, the proportion of pest areas in images collected in the field is usually small, which leads to more invalid information contained in global features. Some of the methods mentioned above are optimized mainly for pest detection algorithms in specific scenarios [24,25,26,27], which are not applicable to similar pest detection in complex field environments. Although the improved feature fusion methods [28,29] can improve detection results, they are optimized only from the perspective of feature fusion and still have some limitations.
As shown in Figure 1a, pest identification problems arise from inaccurate detection and recognition. The main factors that affect the accuracy of field pest identification are as follows: (1) the complex living environment of field pests and the similarity in appearance among different species of pests, which can lead to confusion in the extraction of pest features; (2) the responses of pest regions are different on multiple-scale feature maps, and incorrect feature mapping results in missed detection of pests, as shown in Figure 1b; (3) the current two-stage pest detectors adopt the same RoI (region of interest) feature for the classification task and the localization task, which may make it difficult to train the task head network with optimal parameters; and (4) due to the difficulty of pest image acquisition in complex scenes, there is a lack of similar pest datasets for real farmland scenes.
To address the aforementioned issues, some methods based on feature enhancement [30], feature fusion [31], and feature separation [19] were proposed. Inspired by this, we optimized the two-stage algorithm in terms of contextual information, adaptive feature fusion, and separating features for diverse tasks to further improve the pest detection accuracy. Firstly, a context feature enhancement module was constructed to generate multi-scale features, which were used to enhance the features extracted by the backbone network. Secondly, the attention mechanism was used to adaptively weight the fusion of pest RoI features on the multi-scale feature map to obtain more accurate features of the pest regions. Finally, different features were constructed for separating the classification and localization tasks. The multi-scale features extracted by the context feature enhancement module were used for the localization task, and the context-enhanced features fused by the FPN were used for the classification task.
The main contributions of our work can be summarized as follows:
(1)
A context feature enhancement module (CFEM) was proposed to obtain attention maps at each scale by atrous spatial pyramid pooling, which was beneficial for the detection of similar pests.
(2)
We proposed the RoI feature fusion module (RFFM) to adaptively weight and fuse pest features on multiple network layers, which was more conducive to the classification and localization of pests at different scales.
(3)
The proposed task separation module (TSM) decoupled the features of pest classification and localization networks, improving the overall performance of the detector.
(4)
A larval dataset, SimilarPest5, containing five similar pest species was established to demonstrate the effectiveness of the developed method.

2. Materials and Methods

2.1. Dataset

In recent years, several pest datasets have been published [32,33] which contain various species of pests but have either low similarity between pests or a small number of samples for each pest category. In addition, the image backgrounds in some datasets are homogeneous and significantly differ from the actual field environment. To achieve the specific task of accurately identifying similar pests in the field, we constructed a dataset named SimilarPest5, consisting of 5177 images, all of which were collected from the field environment. A comparison of multiple insect pest datasets is shown in Table 1.
The feeding habits of most Lepidoptera pests are significantly different between larval and adult stages. The larvae mainly feed on the leaves, stems, and ears of crops, causing serious damage to their growth [2]. The SimilarPest5 dataset mainly collects the larvae of five similar pests, including S. frugiperda, M. separata, O. furnacalis, H. armigera, and S. litura. The images in the SimilarPest5 dataset were collected in five different cities and counties in China, and the collection period was mainly focused on the period from July to October between 2020 and 2022. Weather conditions were mostly sunny or cloudy during the acquisition process to ensure the clarity and visibility of the images. The crop species in the images are mainly maize, and a small number of images from soybean fields are also included. To increase the generalization capability of the dataset, we used different kinds of acquisition devices, such as digital cameras and smartphones from different manufacturers. In addition, we acquired images from multiple angles and distances to obtain more visual information while ensuring that the pests were clearly visible. This diverse data collection approach contributed to a comprehensive and diverse collection of field pest images. To minimize crop damage, some obstructive objects were removed to ensure the capture of larvae on the stems and leaves. The SimilarPest5 dataset contains only similar pests in complex environments in the field, and the number of images for each pest reaches about 1000, which is different from other pest datasets. Different species of pests have similar morphology and appearance, which poses significant challenges for precise classification and localization. In addition, the complex field environment causes interference, such as obstruction and lighting, in some pest images.
We uniformly scaled the image width and height to 800 × 600 pixels and used LabelImg (https://github.com/tzutalin/labelImg) (accessed on 28 March 2023) software to annotate the pests in the images. We invited researchers from the Academy of Agricultural Sciences and agricultural experts to annotate the pest images. To ensure the accuracy of annotation, each expert focused on only one pest species. Finally, all experts collaborated to check the correctness of each annotation instance. Annotation information mainly included pest ID and location coordinates, which were stored in XML format. For training detection models, these collected pest images were divided into a training set (80%) and a testing set (20%). Table 2 reports the statistical data for each pest species.
Table 3 provides the statistical data of pests at each scale. According to the division standard of the MS COCO [34], objects smaller than 32 × 32 pixels are defined as small objects, those from 32 × 32 to 96 × 96 are considered medium, and those greater than 96 × 96 are defined as large objects. The sample scale distribution in SimilarPest5 is mainly concentrated in the range of medium and large objects.
The larval images of the target insect pests in the SimilarPest5 dataset are shown in Figure 2. Different pest species have similar morphology and appearance. Additionally, due to the living habits of pests in the field, some pests in the images are subject to interference, such as occlusion and lighting, which weakens the feature information used to distinguish between different species of pests and between foreground/background. All these factors pose significant challenges to the accurate classification and positioning of pests.

2.2. Methodologies

In general, the detection speed of the first-stage detector is faster, but the detection accuracy is not higher than that of the two-stage detector [35]. Therefore, we focused on studying the feature refinement method based on the two-stage detection framework to improve the accuracy of pest detection, and the Cascade RCNN [18] algorithm was used as the baseline network. First, pest images were fed into a backbone network to extract features. Then, the extracted feature maps were fed into the feature enhancement module (CFEM) to generate high-quality enhanced features. Next, after the enhanced feature maps were processed through the FPN, the RoI feature fusion module (RFFM) fused target region features at multiple scales. Finally, the task separation module (TSM) decoupled the features of different tasks to achieve the pest classification and localization. The overall framework of the pest detector is shown in Figure 3, and a detailed description of the modules is given in the following subsections.

2.2.1. Context Feature Enhancement Module

In order to enhance the feature description ability of the network for the target pests, we designed the CFEM to generate enhanced features of different layers. Different from the global feature enhancement based on the backbone network [30], the CFEM captured contextual information using multi-scale receptive fields, which helped the model understand semantic information of an image at different scales. As shown in Figure 4, a 1 × 1 convolution operation was conducted on each scale of features, Ci, extracted from the backbone network to ensure a uniform number of channels for each scale feature. In this paper, the number of channels was 256. In order to obtain the multi-scale context information of the target pest, atrous spatial pyramid pooling (ASPP) [36] with multiple sampling rates and effective target field of view was employed to generate the context information for the corresponding scale layers. The context heat map of the corresponding layers was obtained through the sigmoid activation function. The contextual feature maps of the different layers have differing scale biases. To enhance the information of the specific scale object, the contextual features of the corresponding layers were added to the original features using residual connections to avoid the pest features at specific scales being overwhelmed by background information.
The whole computation process can be summarized as follows:
A i = σ ( f a s p p ( φ i ( C i ) ) ) ,
M i = ( 1 A i ) φ i ( C i ) ,
where σ is the sigmoid activation function, φ i denotes a 1 × 1 convolution operation at the i t h layer, f a s p p indicates the ASPP context-aware operation, and A i denotes the context features at the i t h layer. C i represents the output features of the i t h layer of the backbone network, M i represents the enhancement features of the i t h layer, denotes element-wise addition, and denotes element-wise multiplication.

2.2.2. RoI Feature Fusion Module

In the feature pyramid structure, high-resolution feature maps have more detail and are more sensitive to small objects, while low-resolution feature maps have a high degree of semantic information and are usually employed in the detection of large objects [20]. The general two-stage object detection algorithm maps the feature of the proposal box to a specific layer of the FPN by the size of the proposal box to obtain the RoI features. However, this approach may lead to incorrect detection results because the proposal boxes of the target may not be assigned to the optimal feature map.
The effectiveness of the attention mechanism in feature fusion has been verified, and representative algorithms include channel attention networks and spatial attention networks, such as SENet [37] and CBAM [38]. The attention module can learn weight parameters adaptively, instead of mapping proposal boxes to one feature map. Unlike PANet [31], which used a fully connected layer to fuse all pyramid-level RoI features, we adopted the RFFM to adaptively aggregate the RoI features of the different scales from all feature maps. As shown in Figure 5, RoIAlign [39] was used to extract the RoI features of the P2–P5 layers, with a feature size of 7 × 7, and then these features were connected. To reduce the computational effort, the RFFM module initially performed feature dimensionality reduction. It subsequently adaptively calculated the weights of the features at each scale and finally performed weighted fusion of the features. After multiple convolutional operations and the sigmoid activation function, the spatial weights of multiple layers were obtained. The RoI features of different layers were weighted and fused with the weights of the corresponding layers to obtain the final fused features. It is important to note that the weight parameters were adaptively learned with the back-propagation of the network, which avoided the hard selection of RoI features and achieved better detection results.
The RFFM formula is expressed as follows:
w = σ ( ϕ 3 ( ψ ( ϕ 1 ( R c ) ) ) ,
R = i = 2 5 w i R i ,
where R c denotes the RoI features after the concatenate operation; ϕ 1 and ϕ 3 denote the 1 × 1 convolution and 3 × 3 convolution, respectively; ψ is the ReLU activation function; σ is the sigmoid activation function; w is the adaptive weight of the RoI features; w i is the weight of the i t h layer after splitting; R i represents the RoI features of the feature pyramid at the i t h layer; and R is the features after adaptive fusion.

2.2.3. Task Separation Module

High-level semantic information is helpful for classification, while localization is more sensitive to details [40]. Therefore, the features suitable for classification and localization may not always be consistent. The Double Head [19] algorithm was decoupled from the localization head and the classification head, which leads to better performance. However, they still share the same RoI feature.
In this study, we focused on constructing different features for pest classification and localization tasks. Notably, the fusion features within layers P2 to P5, connected from top to bottom, have enhanced semantic information. Conversely, the fusion features in the layers M2 to M5 have richer detail information. Therefore, the RoI features extracted by the RFFM on the FPN output features (P2–P5) were used for the classification task, while the RoI features extracted by the RFFM on the CFEM output features (M2–M5) were used for the localization task. As shown in Figure 6, the first stage of the two-stage object detection algorithm would output proposal boxes, which were mapped to the M2–M5 and P2–P5 feature maps. The RoI features on M2–M5 and P2–P5 were adaptively fused by their corresponding RFFMs. Then, the output features were independently used for localization and classification tasks through separate task branches while maintaining feature consistency. For the classification task, we employed the cross-entropy loss function, and for the localization task, the Smooth L1 loss function was used.

2.2.4. Parameter Settings

All the experiments were based on the SimilarPest5 dataset. In our experiments, the ResNet-50 [41] and ConvNext-B [42] models trained on the ImageNet [43] dataset were used as pre-trained models. The size of the input images to the network was proportionally adjusted to (1333, 800). The experiments were based on single-scale training and testing of the MMDetection [44] object detection framework. The experiments were conducted on the operating system Ubuntu 18.04 based on Python 3.7, PyTorch 1.10, and CUDA 11.3. In our experiments, two NVIDIA TITAN RTX GPUs with 24 GB of memory were used. All experiments were iteratively fine-tuned for 12 epochs, and the optimizer SGD (stochastic gradient descent) was adopted to train the models. The learning rate was initialized to 0.005 and reduced to one-tenth after the 8th and 11th epochs. The hyper-parameter settings are shown in Table 4, and other parameters were set to the defaults of MMDetection [44]. Due to memory constraints, all layers were fine-tuned with a stochastic gradient descent (SGD) optimizer in mini batches of size 2. The random flip operation was used in the training phase with a random scale of 0.5.

3. Results

3.1. Evaluation Metrics

The evaluation metric is an important basis for evaluating the performance of a method. To ensure the fairness of an experimental comparison, the standard evaluation metrics for the general object detection task are used. These metrics use the intersection over union (IoU) to represent accuracy in predicting bounding boxes and evaluate the performance. The average precision (AP) indicates the detection performance of each category; it is the area bounded by the precision-recall curve. The mean average precision (mAP) was used to evaluate the overall performance, and it represented the mean value of the AP for all categories, starting from 0.5 for the IoU threshold and increasing by steps of 0.05 up to 0.95. The mean recall (mRecall) represented the mean value of the recall for all categories, starting from 0.5 for the IoU threshold and increasing by steps of 0.05 up to 0.95. The calculation formulas are as follows:
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
A P = 0 1 P ( R ) d ( R )
where T P , F P , and F N denote true positives, false positives, and false negatives, respectively.

3.2. Comparison with State-of-the-Art Methods

To illustrate the overall performance of the proposed method, we conducted a comparison with other advanced object detection methods, including CNN-based one-stage methods (RetinaNet [17], YOLOF [45], YOLOV5 [13], and YOLOV8 [14]) and two-stage methods (Faster RCNN [16], Double Head [19], Libra RCNN [46], Cascade RCNN [18], and Sparse RCNN [47]), as well as the transformer-based object detectors (such as Dino [48]). Table 5 reports the experimental results for the SimilarPest5 dataset, and the training process of the model was reported in the Supplemental Materials (Figures S1 and S2). The proposed modules were integrated into the Cascade RCNN [18] framework. Compared with the other methods, our proposed method achieved the highest mAP of 72.7%, 1.2% higher than the other best-performing algorithm. Additionally, the proposed method performed better in the detection of medium- and large-scale pests with a higher recall rate. This indicated that feature optimization modules can improve the detection accuracy for similar pests in the field. As additional modules were added to the original two-stage detection framework, the overall number of parameters of the proposed method increased, resulting in a decrease in FPS (frames per second). Table S1 of the Supplementary Material reports the AP of different methods for detecting each category based on the ConvNext-B [42] backbone network.
Figure 7 shows the confusion matrix of the proposed method when the confidence threshold was equal to 0.5. The diagonal represents the percentage of each pest that was correctly identified. The complex field environment is the main reason for the misidentification between pest targets and backgrounds. The similarity in appearance led to confusing identifications between pests. As shown in Figure 7, the accuracy for S. frugiperda was reduced because some of H. armigera, O. furnacalis, and the background were misidentified as S. frugiperda. A larger proportion of H. armigera was misidentified as S. frugiperda and the background, which reduced the accuracy for H. armigera. In addition, there were different proportions of confusing identifications for M. separata, O. furnacalis, and other pests. The accuracy for S. litura was higher due to the lower percentage of misidentifications.

3.3. Ablation Experiments

We conducted several experiments on the SimilarPest5 dataset to explore the effect of the sampling rate, r, on detection accuracy in the ASPP network. The detection results were evaluated by the metrics of mRecall and mAP, as shown in Figure 8. For efficiency, the experiments were conducted using the Faster RCNN detector with our constructed CFEM by using ResNet-50 [41] as the backbone. When r equaled 1, the pest features were not sufficiently correlated with the surrounding contextual information to achieve optimal accuracy. Too large an r led to a higher correlation of pest features with distant background information, decreasing accuracy. Therefore, setting r to 2 was more suitable for our dataset, and the mAP and mRecall achieved the highest accuracy.
To verify the effectiveness of the CFEM in enhancing pest feature expression, we compared the feature response maps before and after adding the CFEM based on the Faster RCNN [16] algorithm framework in Figure 9. The features enhanced by the CFEM had stronger semantic correlation between local and contextual features of pests. The incorporation of richer contextual information made the feature responses of pest regions more obvious and accurate. Figure S3 of the Supplementary Material shows the feature response maps of our method to different pests at multiple scales.
Based on the Faster R-CNN [16] framework, we constructed ablation experiments with different features selected for the classification task and the localization task of the TSM. The results of these experiments are shown in Table 6. From the results, we could observe that the highest detection accuracy of 63.2% mAP was obtained when using M2~M5 layer features for the localization task and P2~P5 feature layers for the classification task. These results indicated that M2~M5 layer features were more suitable for localization, while P2~P5 layer features fused by the top-down pathway had stronger semantic information and were more suitable for pest classification.
To further validate the effectiveness of each module, we constructed ablation experiments for each module based on the Faster RCNN [16] framework. As shown in Table 7, the addition of any module led to a performance improvement. As shown in the second row of the table, the CFEM could effectively enhance the feature representation ability and achieved a 1.4% mAP improvement. The improvement shown in the third row indicated that the RFFM adaptively fusing multi-layer RoI features outperformed the method with single-layer RoI features in accuracy. The fourth row shows the result of introducing the TSM with a 0.3% mAP improvement, which indicated that the decoupling of tasks based on different features had a positive effect on the detection accuracy improvement. Finally, the proposed method achieved a 63.2% mAP, which was a 3.1% improvement over the original method, and was accompanied by a higher recall rate.
As the proposed method can be embedded as a plug-and-play plugin into most existing two-stage object detection frameworks, we conducted experiments to verify the applicability of the proposed modules with different methods. As shown in Table 8, the mAP of the three methods improved by 3.1%, 2.2%, and 1.6%, respectively, with the addition of our module, and the recall rate also increased. The experimental results demonstrated that the proposed feature refinement modules had good generalization ability.

3.4. Visualization Analysis

The visualization results provide us with a more intuitive way to observe the performance improvement of the proposed method. In Figure 10, we compared the detection results of some two-stage detection methods before and after optimization. Due to the similarity in color and texture between some pests and the complex background, such as M. separata and O. furnacalis, this resulted in missed detections with Double Head [19] and Cascade RCNN [18]. For the detection of S. frugiperda, Faster RCNN [16] and Double Head [19] incorrectly recognized objects in the background as the target pest. Additionally, Double Head [19] showed category confusion in identifying O. furnacalis. Although these methods were able to correctly recognize pests in detecting H. armigera and S. litura, they were not accurate enough in pest localization due to partial occlusion of the pest ontology in some training image sets. In Figure 11, we compared the detection results of the proposed method with those of the other methods. Some other methods also showed incorrect identification results, e.g., Dino [48] and YOLOV8-X [14] misidentified H. armigera as S. frugiperda, while RetinaNet [17] and YOLOV8-X [14] misidentified the background as S. litura. By comparing the detection results of different methods, it can be found that our method is more accurate in both pest classification and localization.

3.5. Discussion

Traditional computer vision algorithms are simple in design and low in computational resource consumption, but weak in feature representation for complex scenes. Therefore, many researchers have started to focus on the application of deep learning methods for pest identification. However, classification-based methods [21,22,23] usually require tedious preprocessing processes, such as manually cropping or segmenting out pest regions, while the algorithms themselves focus mainly on the classification task and have limited applications. In images of field pest datasets, the size and location of pests vary widely, and thus classification-based methods are not adapted to datasets in field environments. In pest identification methods based on object detection frameworks [24,25,26,27,28,29], researchers have designed specific algorithmic frameworks for the characteristics of different pest datasets or optimized them only from the perspective of feature fusion. However, the pests in these datasets differ significantly from the SimilarPest5 dataset in terms of scale, context, and similarity.
Although generic object detection algorithms perform well in most tasks, they often struggle to achieve outstanding results in specific tasks. From the results shown in Table 5, the detection accuracy of two-stage algorithms [16,18,19,46] is usually higher compared to single-stage algorithms [14,17,45] on similar pest datasets. Since two-stage algorithms achieve detection through a coarse-to-fine process of object classification and localization, they are more suitable for fine-grained tasks such as the detection of similar pests in the field.
In this study, we designed a feature refinement method based on the two-stage detection framework with the aim of improving the detection accuracy for similar pests in the field. The two-stage algorithm was optimized by adding feature enhancement, feature fusion, and feature selection modules, and the overall detection accuracy of this method reached 72.7% mAP. Although the proposed method achieved the best accuracy, the subtle differences between the same types of pests at different age stages, the similar appearances between different species of pests, and complex background environments posed challenges for the detection algorithm. As the S. litura sample contained some images of soybeans that had different backgrounds from the images of corn, this led to the highest detection accuracy for S. litura, while similar pests from the same crop were more likely to be confused.
To verify the effectiveness of the proposed method, several ablation experiments were constructed. As shown in Figure 9, the semantic correlation of pest region features was stronger due to the fusion of contextual information at multiple scales, and the richer feature information helped to distinguish different pests. The ablation experiment in Table 7 showed that the RFFM of adaptive fusion of multi-scale region features could improve the detection accuracy compared with selecting only single-scale region features. Furthermore, the fourth and sixth rows of Table 7 verified that feature separation of different tasks allowed different task heads to focus more on specific tasks, thereby enabling the network to train better parameters to improve accuracy. Table 8 showed the generality of the proposed method on different models, with improved accuracy for these models. Overall, compared with other excellent detection algorithms, the proposed method based on the Cascaded RCNN [18] framework was superior in terms of overall detection accuracy.

4. Conclusions

In agricultural production, many lepidopteran pests with similar appearances, represented by S. frugiperda, cause serious damage to crop growth. Therefore, we constructed a SimilarPest5 dataset with images of five similar pests in corn and soybean fields. However, since these images were obtained in actual corn and soybean field environments, they frequently exhibit challenges like complex backgrounds, uneven lighting, and occasional obstruction. These factors make it challenging for a generic model to describe the features of the pests accurately. To improve the accuracy of pest detection in real-field scenarios, we optimized the detection network from different perspectives. The effectiveness of the proposed modules was validated through ablation and comparison experiments, and state-of-the-art performance was achieved on the SimilarPest5 dataset.
The method proposed in this paper can provide intelligent recognition functions for terminal devices and reduce the burden on professionals (Figure S4). In real-world IPM scenarios, higher pest detection accuracy can provide effective decision support for pest control, while providing early warning information to agricultural personnel and helping to develop more effective IPM strategies. However, our proposed method still has some limitations. In future work, we hope to collect and construct larger datasets of field pests from different crops and extend the proposed method to other types of pest detection, such as rice pests like Nilaparvata lugens Stal, Sogatella furcifera, and ladelphax striatellus falln, to explore the effectiveness of the method in pest detection for other crops. Furthermore, we would like to further analyze the similarity of pests based on their morphological, ecological, and statistical characteristics and verify the detection accuracy of the proposed algorithm for pests with different similarity levels. Since the algorithm introduces additional modules, it leads to greater computational complexity and requires support from hardware devices with higher computational performance. Therefore, in future work, we will try to build lighter model architectures while maintaining detection performance.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/insects14100819/s1, Figure S1, Loss function curves of the proposed method based on different backbone network models; Figure S2, Comparison of the mAP of the proposed method with different “max_epochs” settings; Figure S3, Feature response maps of different pest images at multiple scales; Figure S4, Pest Identification System Client Application. Table S1. Performance of different methods as indicated by the detection results for each pest (unit: %).

Author Contributions

Conceptualization, H.C. and R.W.; methodology, H.C. and J.D.; software, H.C.; validation, H.C., T.C. and H.L.; formal analysis, H.C. and J.Z.; investigation, H.C., J.D. and T.C.; resources, J.D., T.C. and H.L.; data curation, H.C., J.D. and G.Z.; writing—original draft preparation, H.C.; writing—review and editing, H.C., J.D., T.C., H.L. and R.L.; visualization, H.C.; supervision, R.W.; project administration, J.Z.; Funding acquisition, R.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (under grant no. 32171888), the major special project of Anhui Province Science and Technology (no. 2020b06050001), and the Natural Science Foundation of Anhui Province (no. 2208085MC57).

Data Availability Statement

The original contributions presented in the study are included in the article/Supplementary Materials; further inquiries can be directed to the corresponding authors.

Acknowledgments

Thanks to all the authors cited in this article and to the referee for their helpful comments and suggestions.

Conflicts of Interest

Guotao Zhou was employed by the company Henan Yunfei Technology Development Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as potential conflict of interest.

References

  1. Oberemok, V.V.; Gal’chinsky, N.V.; Useinov, R.Z.; Novikov, I.A.; Puzanova, Y.V.; Filatov, R.I.; Kouakou, N.J.; Kouame, K.F.; Kra, K.D.; Laikova, K.V. Four Most Pathogenic Superfamilies of Insect Pests of Suborder Sternorrhyncha: Invisible Superplunderers of Plant Vitality. Insects 2023, 14, 462. [Google Scholar] [CrossRef] [PubMed]
  2. Nboyine, J.A.; Kusi, F.; Abudulai, M.; Badii, B.K.; Zakaria, M.; Adu, G.B.; Yahaya, A. A new pest, Spodoptera frugiperda (JE Smith), in tropical Africa: Its seasonal dynamics and damage in maize fields in northern Ghana. Crop Prot. 2020, 127, 104960. [Google Scholar] [CrossRef]
  3. Babendreier, D.; Koku Agboyi, L.; Beseh, P.; Osae, M.; Nboyine, J.; Ofori, S.E.K.; Frimpong, J.O.; Attuquaye Clottey, V.; Kenis, M. The Efficacy of Alternative, Environmentally Friendly Plant Protection Measures for Control of Fall Armyworm, Spodoptera Frugiperda, in Maize. Insects 2020, 11, 240. [Google Scholar] [CrossRef] [PubMed]
  4. Li, G.P.; Wang, Y.N.; Li, H.; Huang, J.R.; He, Y.Z.; Feng, H.Q. Larval Identification of Spodoptera frugiperda and Other Common Species Occurring at Seedling Stage Maize in Henan Province. Chin. J. Biol. Control 2019, 35, 747–754. [Google Scholar]
  5. Patrício, D.I.; Rieder, R. Computer vision and artificial intelligence in precision agriculture for grain crops: A systematic review. Comput. Electron. Agric. 2018, 153, 69–81. [Google Scholar] [CrossRef]
  6. Liu, T.; Chen, W.; Wu, W.; Sun, C.; Guo, W.; Zhu, X. Detection of aphids in wheat fields using a computer vision technique. Biosyst. Eng. 2016, 141, 82–93. [Google Scholar] [CrossRef]
  7. Gondal, M.D.; Khan, Y.N. Early pest detection from crop using image processing and computational intelligence. FAST-NU Res. J. 2015, 1, 59–68. [Google Scholar]
  8. Ebrahimi, M.A.; Khoshtaghaza, M.H.; Minaei, S.; Jamshidi, B. Vision-based pest detection based on SVM classification method. Comput. Electron. Agric. 2017, 137, 52–58. [Google Scholar] [CrossRef]
  9. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  10. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  11. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  12. Redmon, J.; Farhadi, A. Yolov3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  13. Jocher, G. YOLOv5 by Ultralytics, (Version 6.2); GitHub Repository. 2020. Available online: https://github.com/ultralytics/yolov5/tree/v6.2 (accessed on 15 May 2023).
  14. Jocher, G.; Chaurasia, A.; Qiu, J. YOLO by Ultralytics, (Version 8.0.0); GitHub Repository. 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 28 May 2023).
  15. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  16. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [PubMed]
  17. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  18. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6154–6162. [Google Scholar]
  19. Wu, Y.; Chen, Y.; Yuan, L.; Liu, Z.; Wang, L.; Li, H.; Fu, Y. Rethinking classification and localization for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10186–10195. [Google Scholar]
  20. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  21. Tetila, E.C.; Machado, B.B.; Astolfi, G.; de Souza Belete, N.A.; Amorim, W.P.; Roel, A.R.; Pistori, H. Detection and classification of soybean pests using deep learning with UAV images. Comput. Electron. Agric. 2020, 179, 105836. [Google Scholar] [CrossRef]
  22. Thenmozhi, K.; Reddy, U.S. Crop pest classification based on deep convolutional neural network and transfer learning. Comput. Electron. Agric. 2019, 164, 104906. [Google Scholar] [CrossRef]
  23. Turkoglu, M.; Yanikoğlu, B.; Hanbay, D. PlantDiseaseNet: Convolutional neural network ensemble for plant disease and pest detection. Signal Image Video Process. 2022, 16, 301–309. [Google Scholar] [CrossRef]
  24. Li, R.; Wang, R.J.; Xie, C.J.; Liu, L.; Zhang, J.; Wang, F.Y.; Liu, W.C. A coarse-to-fine network for aphid recognition and detection in the field. Biosyst. Eng. 2019, 187, 39–52. [Google Scholar] [CrossRef]
  25. Teng, Y.; Wang, R.; Du, J.; Huang, Z.; Zhou, Q.; Jiao, L. TD-Det: A Tiny Size Dense Aphid Detection Network under In-Field Environment. Insects 2022, 13, 501. [Google Scholar] [CrossRef]
  26. Rustia, D.J.A.; Chao, J.J.; Chiu, L.Y.; Wu, Y.F.; Chung, J.Y.; Hsu, J.C.; Lin, T.T. Automatic greenhouse insect pest detection and recognition based on a cascaded deep learning classification method. J. Appl. Entomol. 2021, 145, 206–222. [Google Scholar] [CrossRef]
  27. Tannous, M.; Stefanini, C.; Romano, D. A Deep-Learning-Based Detection Approach for the Identification of Insect Species of Economic Importance. Insects 2023, 14, 148. [Google Scholar] [CrossRef]
  28. Dong, S.; Du, J.; Jiao, L.; Wang, F.; Liu, K.; Teng, Y.; Wang, R. Automatic Crop Pest Detection Oriented Multiscale Feature Fusion Approach. Insects 2022, 13, 554. [Google Scholar] [CrossRef]
  29. Amrani, A.; Sohel, F.; Diepeveen, D.; Murray, D.; Jones, M.G. Insect detection from imagery using YOLOv3-based adaptive feature fusion convolution network. Crop Pasture Sci. 2022, 74, 615–627. [Google Scholar] [CrossRef]
  30. Takalkar, M.A.; Thuseethan, S.; Rajasegarar, S.; Chaczko, Z.; Xu, M.; Yearwood, J. LGAttNet: Automatic micro-expression detection using dual-stream local and global attentions. Knowl. Based Syst. 2021, 212, 106566. [Google Scholar] [CrossRef]
  31. Liu, S.; Qi, L.; Qin, H.; Shi, J.P.; Jia, J.Y. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  32. Wang, R.; Liu, L.; Xie, C.; Yang, P.; Li, R.; Zhou, M. AgriPest: A large-scale domain-specific benchmark dataset for practical agricultural pest detection in the wild. Sensors 2021, 21, 1601. [Google Scholar] [CrossRef] [PubMed]
  33. Wu, X.; Zhan, C.; Lai, Y.K.; Cheng, M.M.; Yang, J. IP102: A large-scale benchmark dataset for insect pest recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8787–8796. [Google Scholar]
  34. Chen, X.; Fang, H.; Lin, T.Y.; Vedantam, R.; Gupta, S.; Dollár, P.; Zitnick, C.L. Microsoft coco captions: Data collection and evaluation server. arXiv 2015, arXiv:1504.00325. [Google Scholar]
  35. Soviany, P.; Ionescu, R.T. Optimizing the trade-off between single-stage and two-stage deep object detectors using image difficulty prediction. In Proceedings of the 2018 20th International Symposium on Symbolic and Numeric Algorithms for Scientific Computing (SYNASC), Timisoara, Romania, 20–23 September 2018; pp. 209–214. [Google Scholar]
  36. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  37. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  38. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  39. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  40. Selvaraju, R.R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-cam: Visual explanations from deep networks via gradient-based localization. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  41. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  42. Liu, Z.; Mao, H.; Wu, C.Y.; Feichtenhofer, C.; Darrell, T.; Xie, S. A convnet for the 2020s. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 11976–11986. [Google Scholar]
  43. Deng, J.; Dong, W.; Socher, R.; Li, L.J.; Li, K.; Fei-Fei, L. Imagenet: A large-scale hierarchical image database. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 22–24 June 2009; pp. 248–255. [Google Scholar]
  44. Chen, K.; Wang, J.; Pang, J.; Cao, Y.; Xiong, Y.; Li, X.; Sun, S.; Feng, W.; Liu, Z.; Xu, J.; et al. MMDetection: Open mmlab detection toolbox and benchmark. arXiv 2019, arXiv:1906.07155. [Google Scholar]
  45. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13039–13048. [Google Scholar]
  46. Pang, J.; Chen, K.; Shi, J.; Feng, H.; Ouyang, W.; Lin, D. Libra r-cnn: Towards balanced learning for object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 821–830. [Google Scholar]
  47. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Luo, P. Sparse r-cnn: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14454–14463. [Google Scholar]
  48. Zhang, H.; Li, F.; Liu, S.; Zhang, L.; Su, H.; Zhu, J.; Shum, H.Y. Dino: Detr with improved denoising anchor boxes for end-to-end object detection. arXiv 2022, arXiv:2203.03605. [Google Scholar]
Figure 1. (a) Examples of detection results for similar pests in the field based on the Cascade RCNN algorithm. (b) Feature response maps of the pest image at different scales, where P3, P4, and P5 represent the feature response maps of the third, fourth, and fifth layers, respectively. The response of the pest region on the P5 layer feature map was not significant.
Figure 1. (a) Examples of detection results for similar pests in the field based on the Cascade RCNN algorithm. (b) Feature response maps of the pest image at different scales, where P3, P4, and P5 represent the feature response maps of the third, fourth, and fifth layers, respectively. The response of the pest region on the P5 layer feature map was not significant.
Insects 14 00819 g001
Figure 2. Examples of various pest images from the SimilarPest5 dataset. The pests in SimilarPest5 have a high similarity of color, texture, shape, and living environment.
Figure 2. Examples of various pest images from the SimilarPest5 dataset. The pests in SimilarPest5 have a high similarity of color, texture, shape, and living environment.
Insects 14 00819 g002
Figure 3. Overall architecture of our two-stage pest detector. The CFEM is used to enhance features at each scale, and the RFFM is used to fuse the RoI features of a multi-scale pyramid. The TSM constructs independent feature and task head networks for classification and localization.
Figure 3. Overall architecture of our two-stage pest detector. The CFEM is used to enhance features at each scale, and the RFFM is used to fuse the RoI features of a multi-scale pyramid. The TSM constructs independent feature and task head networks for classification and localization.
Insects 14 00819 g003
Figure 4. (a) Architecture of the CFEM, where Ci denotes the i t h layer features extracted by the backbone network. (b) Detailed network structure of C i to A i , where ‘r’ represents the sampling rate.
Figure 4. (a) Architecture of the CFEM, where Ci denotes the i t h layer features extracted by the backbone network. (b) Detailed network structure of C i to A i , where ‘r’ represents the sampling rate.
Insects 14 00819 g004
Figure 5. (a) Architecture of the RFFM. “N” represents the number of RoI feature maps, which is 4 in this paper. “C” represents the channel number of RoI features, which is 256 in this paper. (b) Detailed network structure of the RFFM.
Figure 5. (a) Architecture of the RFFM. “N” represents the number of RoI feature maps, which is 4 in this paper. “C” represents the channel number of RoI features, which is 256 in this paper. (b) Detailed network structure of the RFFM.
Insects 14 00819 g005
Figure 6. (a) Architecture of the task separation module. “K + 1” represents the number of pest categories and the background. “PRoI” and “MRoI” denote the adaptive fusion features of candidate regions on different feature maps. (b) Detailed network structure of the classification and localization network.
Figure 6. (a) Architecture of the task separation module. “K + 1” represents the number of pest categories and the background. “PRoI” and “MRoI” denote the adaptive fusion features of candidate regions on different feature maps. (b) Detailed network structure of the classification and localization network.
Insects 14 00819 g006
Figure 7. Confusion matrix of the proposed method with a confidence threshold equal to 0.5 (unit: %).
Figure 7. Confusion matrix of the proposed method with a confidence threshold equal to 0.5 (unit: %).
Insects 14 00819 g007
Figure 8. Detection accuracy with change in the sampling rate, r.
Figure 8. Detection accuracy with change in the sampling rate, r.
Insects 14 00819 g008
Figure 9. Comparison of feature response maps. The first row shows the feature response maps of the original pest features fused by the FPN, and the second row shows the feature response maps after adding the CFEM. All feature response maps were from the P4 layer, and ResNet-50 was used as the backbone network.
Figure 9. Comparison of feature response maps. The first row shows the feature response maps of the original pest features fused by the FPN, and the second row shows the feature response maps after adding the CFEM. All feature response maps were from the P4 layer, and ResNet-50 was used as the backbone network.
Insects 14 00819 g009
Figure 10. Visualization of the detection results of the two-stage detection methods before and after optimization. The second, fourth, and sixth rows show the detection result images with the proposed modules added on different method frameworks.
Figure 10. Visualization of the detection results of the two-stage detection methods before and after optimization. The second, fourth, and sixth rows show the detection result images with the proposed modules added on different method frameworks.
Insects 14 00819 g010
Figure 11. Visualization of the detection results of our method compared with other detectors.
Figure 11. Visualization of the detection results of our method compared with other detectors.
Insects 14 00819 g011
Table 1. Comparison of existing datasets related to insect pests. “Class” denotes the class number. “Samples number” represents the number of images. “Crop” refers to the crop species involved in the acquisition of pest images. “Stage” indicates the developmental stage of the pests. “Avail” indicates if the dataset is available. “Y” and “N” denote “yes” and “no”, respectively.
Table 1. Comparison of existing datasets related to insect pests. “Class” denotes the class number. “Samples number” represents the number of images. “Crop” refers to the crop species involved in the acquisition of pest images. “Stage” indicates the developmental stage of the pests. “Avail” indicates if the dataset is available. “Y” and “N” denote “yes” and “no”, respectively.
DatasetImageClassSamples NumberCropStageAvail Similarity
Tetila et al. [21]Insects 14 00819 i001135000SoybeanLarvaeNLow
Li et al. [24]Insects 14 00819 i00212200Wheat, rapeseedLarvae, adultN/
Jiao et al. [29]Insects 14 00819 i003212442Fixed equipmentAdultNLow
Wang et al. [32]Insects 14 00819 i0041449,707Rice, wheat, maize, rapeseedLarvae, adultNLow
IP102 [33]Insects 14 00819 i00510275,222Rice, wheat, mango, etc.Larvae, adultYLow
SimilarPest5Insects 14 00819 i00655177Maize, soybeanLarvae/High
Table 2. Statistics of the SimilarPest5 dataset. “Crop” refers to the crop species involved in the acquisition of pest images. “Samples number” indicates the number of pest images. “Instances number” indicates the number of pest targets in all images. “Training set” indicates the number of images in the training set. “Test set” indicates the number of images in the test set.
Table 2. Statistics of the SimilarPest5 dataset. “Crop” refers to the crop species involved in the acquisition of pest images. “Samples number” indicates the number of pest images. “Instances number” indicates the number of pest targets in all images. “Training set” indicates the number of images in the training set. “Test set” indicates the number of images in the test set.
Pest IDCategoriesSamples NumberInstances NumberTraining SetTest Set
1S. frugiperda10711152857214
2M. separata10231141819204
3O. furnacalis10381070831207
4H. armigera10381041832206
5S. litura10071025806201
Table 3. Statistics of pests at each scale in the SimilarPest5 dataset. “Ratio” indicates the number of pest instances at the corresponding scale as a proportion of the number of all pests.
Table 3. Statistics of pests at each scale in the SimilarPest5 dataset. “Ratio” indicates the number of pest instances at the corresponding scale as a proportion of the number of all pests.
ScaleNumberRatioAverage Pixel
Medium214939.6%5304
Large328060.4%21,869
Table 4. Training hyper-parameters.
Table 4. Training hyper-parameters.
Batch SizeGPUsEpochOptimizerLearning RateWeight DecayMomentum
2212SGD0.0050.00010.9
Table 5. Comparison experiments with other advanced object detection algorithms (unit: %). “Params” indicates the number of parameters (unit: M). “FPS” represents the number of frames processed per second.
Table 5. Comparison experiments with other advanced object detection algorithms (unit: %). “Params” indicates the number of parameters (unit: M). “FPS” represents the number of frames processed per second.
MethodBackbonemAPAPMAPLmRecallParams(M)FPS
RetinaNet [17]ResNet5056.948.860.066.436.2933.4
YOLOF [45]ResNet5062.855.265.970.442.2752.6
Faster RCNN [16]ResNet5060.151.963.467.141.1732.3
Libra RCNN [46]ResNet5062.452.466.270.441.4330.8
Double Head [19]ResNet5063.352.967.270.046.7614.2
Cascade RCNN [18]ResNet5064.550.769.170.868.9526.8
Sparse RCNN [47]ResNet5063.052.467.474.1105.9525.6
Dino [48]ResNet5064.352.069.670.647.5917.3
OursResNet5066.155.770.272.278.5210.3
RetinaNet [17]ConvNext-B65.558.968.472.897.6515.1
YOLOF [45]ConvNext-B66.258.769.873.5105.9718.5
Faster RCNN [16]ConvNext-B69.360.572.874.0104.9414.8
Libra RCNN [46]ConvNext-B69.961.373.375.0105.2014.5
Double Head [19]ConvNext-B71.461.875.176.1110.529.3
Cascade RCNN [18]ConvNext-B71.360.975.476.2132.7313.7
Sparse RCNN [47]ConvNext-B67.958.871.974.2169.7413.8
Dino [48]ConvNext-B71.561.275.876.3111.629.8
YOLOV5-L [13]CSPDarknet67.056.770.773.446.1663.9
YOLOV5-X [13]CSPDarknet67.957.271.274.086.2543.7
YOLOV8-X [14]CSPDarknet67.255.270.173.568.1559.8
OursConvNext-B72.762.476.376.7143.286.2
Table 6. Ablation studies with different features were selected for the classification task and the localization task of the TSM (unit: %).
Table 6. Ablation studies with different features were selected for the classification task and the localization task of the TSM (unit: %).
ClassificationLocalizationmAPmRecall
M2~M5M2~M561.067.8
P2~P5P2~P562.669.2
M2~M5P2~P562.268.6
P2~P5M2~M563.269.3
Table 7. Ablation experiments based on the Faster RCNN algorithm (unit: %). “Params” indicates the number of parameters (unit: M).
Table 7. Ablation experiments based on the Faster RCNN algorithm (unit: %). “Params” indicates the number of parameters (unit: M).
CFEMRFFMTSMmAPmRecallParams(M)
60.167.141.17
61.5 (+1.4)68.6 (+1.5)50.06
61.9 (+1.8)69.0 (+1.9)41.41
62.2 (+2.1)68.7 (+1.6)41.69
62.6 (+2.5)69.2 (+2.1)50.33
63.2 (+3.1)69.3 (+2.2)50.61
Table 8. The performance of various detection methods with or without our module (unit: %).
Table 8. The performance of various detection methods with or without our module (unit: %).
Methodw/oursmAPmRecall
Faster RCNN [16] 60.167.1
63.2 (+3.1)69.3 (+2.2)
Double Head [19] 63.370.0
65.5 (+2.2)71.1 (+1.1)
Cascade RCNN [18] 64.570.8
66.1 (+1.6)72.2 (+1.4)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, H.; Wang, R.; Du, J.; Chen, T.; Liu, H.; Zhang, J.; Li, R.; Zhou, G. Feature Refinement Method Based on the Two-Stage Detection Framework for Similar Pest Detection in the Field. Insects 2023, 14, 819. https://doi.org/10.3390/insects14100819

AMA Style

Chen H, Wang R, Du J, Chen T, Liu H, Zhang J, Li R, Zhou G. Feature Refinement Method Based on the Two-Stage Detection Framework for Similar Pest Detection in the Field. Insects. 2023; 14(10):819. https://doi.org/10.3390/insects14100819

Chicago/Turabian Style

Chen, Hongbo, Rujing Wang, Jianming Du, Tianjiao Chen, Haiyun Liu, Jie Zhang, Rui Li, and Guotao Zhou. 2023. "Feature Refinement Method Based on the Two-Stage Detection Framework for Similar Pest Detection in the Field" Insects 14, no. 10: 819. https://doi.org/10.3390/insects14100819

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop