Next Article in Journal
Simultaneous Stereo Matching and Confidence Estimation Network
Previous Article in Journal
Investigating Contrastive Pair Learning’s Frontiers in Supervised, Semisupervised, and Self-Supervised Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms

College of Physics and Electronic Information, Gannan Normal University, Ganzhou 341000, China
*
Author to whom correspondence should be addressed.
J. Imaging 2024, 10(8), 197; https://doi.org/10.3390/jimaging10080197
Submission received: 16 July 2024 / Revised: 8 August 2024 / Accepted: 10 August 2024 / Published: 13 August 2024

Abstract

:
Currently, existing deep learning methods exhibit many limitations in multi-target detection, such as low accuracy and high rates of false detection and missed detections. This paper proposes an improved Faster R-CNN algorithm, aiming to enhance the algorithm’s capability in detecting multi-scale targets. This algorithm has three improvements based on Faster R-CNN. Firstly, the new algorithm uses the ResNet101 network for feature extraction of the detection image, which achieves stronger feature extraction capabilities. Secondly, the new algorithm integrates Online Hard Example Mining (OHEM), Soft non-maximum suppression (Soft-NMS), and Distance Intersection Over Union (DIOU) modules, which improves the positive and negative sample imbalance and the problem of small targets being easily missed during model training. Finally, the Region Proposal Network (RPN) is simplified to achieve a faster detection speed and a lower miss rate. The multi-scale training (MST) strategy is also used to train the improved Faster R-CNN to achieve a balance between detection accuracy and efficiency. Compared to the other detection models, the improved Faster R-CNN demonstrates significant advantages in terms of [email protected], F1-score, and Log average miss rate (LAMR). The model proposed in this paper provides valuable insights and inspiration for many fields, such as smart agriculture, medical diagnosis, and face recognition.

1. Introduction

Object detection has long been a research focus for computer vision, and it has been extensively applied in areas such as face recognition, medical image diagnosis, and road detection [1,2,3]. At present, deep learning-based target detection methods can be broadly categorized into two main classes [4,5]. One class is the two-stage object detection approach typified by Region-based Convolutional Neural Networks (R-CNNs), and the other class is the one-stage approach typified by You Only Look Once (YOLO). These two types of algorithms have their own characteristics and advantages [6]. One-stage target detection algorithms detect targets directly on the original image without a region proposal step, so these algorithms are relatively faster, but the detection accuracy decreases when detecting different multi-scale targets. Two-stage object detection algorithms exhibit relatively higher detection accuracy but at the cost of slower processing speeds. Driven by the rapid advancement of deep learning, target detection algorithms have achieved impressive gains in both accuracy and processing speed. R-CNN [7] is the seminal work in object detection algorithms, which computes over candidate regions generated by the selective search method, and further applies SVM classification and bounding box regression. Consequently, R-CNN consumes too much time in image processing, resulting in low detection efficiency. He et al. [8] introduced the Faster R-CNN on the basis of R-CNN, which introduced the Region Proposal Network (RPN) to generate candidate regions and utilized shared convolutional features to further improve detection accuracy and efficiency. Cai et al. [9] introduced Cascade R-CNN, which is best characterized by cascading classifiers and multi-stage training to improve detection accuracy and speed. Wan et al. [10] proposed an improved version of Faster R-CNN with optimized convolutional and pooling layers for detecting a wide range of fruits and achieving higher accuracy. Yang et al. [11] introduced an improved strawberry detection algorithm based on Mask R-CNN, which resulted in a substantial improvement in model generalization and robustness. After the R-CNN family of algorithms, Divvala et al. [12] proposed YOLO as an alternative to R-CNN. Unlike R-CNN, YOLO directly predicts classifications and regressions from features, using a single fully connected layer for both tasks. This design enhances speed and efficiency in processing. However, the disadvantage of YOLO is that its generalization ability and robustness are not strong, and it is easy to miss small targets. With the aim of improving the above problems effectively, Liu et al. [13] introduced the Single Shot MultiBox Detector (SSD) family of algorithms. Zhu et al. [14] used SSD to detect fruits on mango trees with an F1 of 0.91. Anagnostis et al. [15] used SSD to categorize infected trees in walnut orchards and the method detected whether walnut leaves were infected with 87% accuracy. Tian et al. [16] proposed EasyRP-R-CNN, a convolution-based framework for cyclone detection. The method was improved based on Region of Interest and achieved satisfactory detection accuracy. Li et al. [17] proposed a lightweight convolutional neural network, WearNet, to achieve automatic detection of scratches on contact sliding parts such as metal molding, and the classification accuracy of the method can reach 94.16%.
All of the methods mentioned above have some problems. (1) These methods detect and recognize a single or no target categories, which cannot satisfy the multi-target detection task, and are not able to accurately localize and recognize small targets. (2) These methods do not have strong feature extraction capabilities for small targets and cannot extract enough information about the target features. Thus, they can generate noise in the detection region, resulting in a decrease in accuracy. (3) These methods do not achieve a balance between detection accuracy and speed to meet the real-time demands of detection tasks. For the purpose of improving the detection accuracy in a multi-scale target environment, after considering accuracy and detection efficiency, this paper chooses to use Faster R-CNN as a baseline to detect different multi-scale targets on the Pascal VOC (Visual Object Classes) dataset [18]. In this paper, the following modifications are made while reducing model computation and improving model detection performance:
(1)
ResNet101 [19] is employed as the trunk network in the improved Faster R-CNN, which enhances the feature extraction capabilities of the model.
(2)
The Online Hard Example Mining (OHEM) algorithm [20] is used to help the model learn hard-to-classify samples more efficiently, which in turn enhances the model’s capacity for generalization. The Soft non-maximum suppression (Soft-NMS) algorithm [21] and the Distance Intersection Over Union (DIOU) algorithm [22] are used to optimize the excessive bounding boxes generated by the RPN and their overlap degree, which enhances the accuracy of detecting small targets and improves the issue of missed target detection.
(3)
The RPN structure is optimized by adding an anchor box with a scale of 64 and using a smaller convolutional kernel to achieve bounding box regression. Employing the multi-scale training (MST) method to train the improved Faster R-CNN [23] achieves a balance between detection accuracy and speed.

2. Methodology and Modeling

In this chapter, we first introduce the overall structure of the proposed model. Then, we provide a detailed explanation of the improvements and integrated modules within the model. Finally, we briefly describe the performance evaluation metrics used in the experiments.

2.1. The Adjusted Faster R-CNN Network

Figure 1 shows the overall framework of the improved Faster R-CNN [24]. By comparing the four networks, VGG16, ResNet34, ResNet50, and ResNet101 [19], ResNet101 is chosen as the trunk network. The introduction of DIOU can increase the effectiveness of the improved Faster R-CNN regarding the problems of slow convergence of the target detection loss function and target regression localization accuracy. The introduction of OHEM enables the improved Faster R-CNN to mine difficult samples in the dynamic training process. This can improve the issue of imbalance between positive and negative samples during the training process. As depicted in Figure 1, the feature map from the ResNet101 trunk network is input into the optimized RPN. At this point, a large number of candidate proposal boxes are generated on the feature map. Soft-NMS is used to eliminate redundant target proposal boxes. It can reduce the miss rate of small targets by gradually decreasing the confidence score of overlapping proposal boxes.

2.1.1. Improved Backbone Network

VGG16 [25] serves as the trunk network for the original Faster R-CNN. In general, data expansion and increasing network depth methods can be used to improve model performance, and network depth is very important for optimizing network performance [26]. ResNet [27] makes it possible for information to skip certain layers directly by introducing direct connections across layers in the network. This effectively improves the problems of gradient vanishing and gradient explosion. ResNet mainly uses convolutional operations instead of fully connected layers, which reduces the number of network parameters and can effectively avoid overfitting problems. Common ResNet structures are ResNet34, ResNet50, and ResNet101 [28]. One of the most significant features of ResNet50 compared to ResNet34 is the introduction of a new bottleneck residual block structure. It comprises a sequence of 1 × 1 convolutional layers, followed by a 3 × 3 convolutional layer, and then another 1 × 1 convolutional layer. This structure allows ResNet50 to have stronger feature representation while maintaining the depth of the model. ResNet101 has an additional set of convolutional blocks compared to ResNet50, which contains multiple residual units. The deeper network structure allows ResNet101 to further enhance its expressive and learning capabilities, allowing it to better capture image details and semantic information, as shown in Figure 2.
ResNet101 consists of two fundamental blocks, as depicted in Figure 2a and Figure 2b, respectively, named Conv Block and Identity Block. The Conv Block’s input and output dimensions are different and cannot be connected in series. Its function is to change the network dimension. The input dimensions and output dimensions of the Identity Block are the same and can be connected in series to deepen the network. ResNet101 consists of multiple residual units in each convolutional block, as illustrated in Figure 2c. Each residual unit performs three convolutional operations. The shortcut connections of residual units and the identity mapping help address the issues of gradient vanishing or exploding, which can lead to a decrease in detection accuracy. As shown in Figure 2d, ResNet101 consists of five convolutional layers and has a depth of 101 layers, which can provide stronger feature extraction capabilities.
In the fourth chapter, comparative experiments are conducted for VGG16, ResNet34, ResNet50, and ResNet101. Experimental results indicate that ResNet101 outperforms the other three networks in overall detection performance. Therefore, ResNet101 has been selected as the trunk network for the improved Faster R-CNN.

2.1.2. Modifying the Region Proposal Network

In the object detection process, CNNs are usually used to extract image features. These features are convolved and pooled through several convolution and pooling operations to produce a smaller feature map (also referred to as an activation map). The activation map contains semantic information and location information taken from the image, and then the activation map is input into the RPN. The RPN [8,29] will first further extract features from the input activation map through a shared convolutional layer, and then it will initially generate a set of predefined anchor boxes at each spatial position of the activation map to accomplish object detection. These predefined anchor boxes generated during the detection process usually have different width-to-height ratios as well as areas. Figure 3 shows the schematic of the optimized RPN.
For each anchor box, the RPN network uses a binary classifier to predict whether it contains a target. It outputs the probability that each anchor box belongs to the foreground or background to complete the initial classification prediction. The conventional RPN generates nine anchor boxes at each spatial position of the activation map with aspect ratios of 1:1, 1:2, and 2:1 and scales of 128, 256, and 512, respectively, as shown in Figure 3a. In order to improve the accuracy of detecting small targets and reduce the leakage rate, a new anchor box with a scale of 64 is added to the RPN while the aspect ratio remains unchanged [30], so that there are a total of 12 anchor boxes at each position of the feature map, which is then able to better capture small-scale targets, as shown in Figure 3b. In addition to the classification predictions, the RPN completes an initial bounding box regression on the positive samples (containing the target’s anchor boxes) with the goal of accurately predicting the target’s bounding box position. The output is the translation and scaling parameters relative to each positive sample anchor box. As depicted in Figure 3c, the structure of the RPN is simplified, and only the 3 × 3 convolutional kernel is used to generate 256 feature maps. The benefits of using a 3 × 3 convolution kernel include: reducing the number of parameters and computational burden, improving computational efficiency, capturing local features, effectively utilizing boundary information, and simplifying the network design. The 3 × 3 convolution kernel is able to maintain model simplicity and computational efficiency while ensuring the effectiveness and accuracy of feature extraction.

2.1.3. Optimization of Detection Boxes and Sample Imbalance

In object detection, the bounding box mechanism is used to locate and identify objects in an image using rectangular boxes. This is mainly achieved by predicting the center coordinates, width, and height of the bounding boxes to ensure accurate object localization and classification. Regression loss functions and post-processing algorithms are employed to optimize the position of the bounding boxes and remove redundant detections, enhancing the accuracy and efficiency of detection. In the improved Faster R-CNN, several common strategies are adopted to enhance the bounding box mechanism. Below is a brief introduction to these strategies.
(1) DIOU
IOU [31] is a widely used performance index in object detection, indicating the degree of overlap between the predicted box and the actual box. Figure 4 shows the schematic diagram of the bounding box with different degrees of overlap. Currently, there are two main disadvantages to using IOU as a loss function for object detection. One is that if there is no overlap between the two detection boxes, IOU = 0, which means that the two detection boxes can no longer participate in deep learning training. Secondly, it can be possible for two detection boxes to have different degrees of overlap but yield the same IOU. This means that IOU does not give accurate feedback on the degree of overlap between the detection boxes, as shown in Figure 4a.
With the aim of improving the above two problems effectively, DIOU [22] is used to better refine the positioning of the bounding box. DIOU can effectively address the issues of slow convergence and regression localization accuracy that exist in IOU, and its formula is defined as Equation (1).
L D I O U = 1 D I O U = 1 ( I O U ρ 2 ( b , b g t ) c 2 ) = 1 I O U + ρ 2 ( b , b g t ) c 2
where ρ 2 ( b , b g t ) denotes the distance between the centers of the prediction box b and the ground truth b g t , and c denotes the diagonal length of the smallest region that contains both the prediction box and the ground truth, as shown in Figure 4b. Since the DIOU loss function has the characteristic of being able to minimize the distance between two prediction boxes directly, it converges and regresses faster than IOU. DIoU improves the traditional IoU [32] metric by addressing the insensitivity of IoU to the distance between bounding boxes. By introducing a penalty term in the loss function based on the center distance of the bounding box, DIoU provides a more comprehensive and accurate measure of bounding box similarity, which leads to better performance in target detection tasks. Using DIOU as the loss function of the improved Faster R-CNN can make the model pay more attention to hard-to-detect targets during the training process, which is especially important for the detection of small and irregularly shaped targets.
(2) OHEM
During the object detection training process, the activation maps from the trunk network are input into the RPN, at which time a large number of candidate proposal boxes are generated. Generally, the IOU threshold is set to 0.5 and proposal boxes with an IOU above 0.5 are retained as positive samples, while those below 0.5 are treated as negative samples. This leads to significantly more positive samples than negative ones. As a result, the model may overlook difficult negative samples that are challenging to detect. These difficult samples can contribute higher loss values, improving the overall detection performance of the model.
In order to enable more difficult samples to be used in the dynamic process of training, OHEM is introduced into the improved Faster R-CNN [20]. Figure 5 illustrates the working principle of OHEM. OHEM [33] improves the training effect of the model by dynamically selecting and processing those difficult samples that the model currently has difficulty classifying during the training process. Specifically, it first detects the input image during each training session and selects the samples with higher loss as difficult samples according to the loss value. These difficult samples are then utilized for backpropagation and parameter updating so that the model learns to correctly classify difficult cases faster, thus improving its overall performance. The OHEM training strategy in the dynamic process of training can send the target samples that are difficult to detect into the network again for deep learning training. This makes the network more sensitive to the detection target, which in turn improves the target detection accuracy.
(3) Soft-NMS
NMS is a very important algorithm in target detection. Its basic idea is to sort the proposal boxes by their confidence scores and retain the one with the highest score. In this process, if the overlap between two proposal boxes exceeds a set threshold (generally 0.5), the box with the lower score is discarded, and the one with the higher score is retained. Therefore, The NMS score is based solely on the classification confidence, without considering the localization accuracy of the bounding box. This means that the classification and localization confidences are not positively correlated.
To effectively address some of the limitations of NMS, Soft-NMS is adopted in the improved Faster R-CNN [21]. Its linear weighted equation is defined as Equation (2).
S i = S i I O U ( M , b i ) < N t S i ( 1 I O U ( M , b i ) I O U ( M , b i ) N t
Here, S i is the proposal box confidence score and M is the proposal box with the highest score. The set of all proposal boxes during training is denoted by b , the area of overlap between M and the proposal boxes in set b is denoted by I O U ( M , b i ) , and N t is the IOU threshold set at the beginning. The difference between Soft-NMS and NMS is the way overlapping bounding boxes are handled. Traditional NMS directly removes other bounding boxes that overlap with the highest scoring box above a certain threshold, which may result in some correct detections being mistakenly removed. In contrast, Soft-NMS does not remove these overlapping bounding boxes, but gradually reduces their scores. Soft-NMS works by attenuating the scores proportionally to the degree of overlap between the bounding box and the highest-scoring box [34]. The larger the overlap, the more the score is attenuated. In this way, Soft-NMS is able to retain more valuable detection results, reduce missed detections, and improve the accuracy of target detection.

2.1.4. Multi-Scale Training

The MST [23] method can enhance the model’s adaptability and generalization capabilities when detecting a variety of target sizes. An image pyramid called MST [35] is used in the training process of CNN. As shown in Figure 6, image training is achieved by randomly inputting images of different scales within a given segment of image size, which enables the target detection model to adapt to targets of different scales. In the testing phase, the same image at different scales is input for multiple detections. Finally, Soft-NMS is employed to integrate all detection results, which enables the detection model to cover as many targets as possible at multiple scales and improves the robustness of the detection model. During the feature extraction phase, the generated activation map will be significantly smaller than the original image. This can make it challenging for the model to focus on the details of small targets. Therefore, by providing the model with larger and richer images, its detection capabilities can be enhanced effectively [36]. In this paper, MST is used to train the improved Faster R-CNN. The training samples have image lengths ranging from 380 to 640, and image widths spanning 300 to 450.

2.2. Evaluation Metrics

To assess the model’s detection capabilities, several widely used evaluation metrics are used here [37,38]. These metrics include Average Precision (AP), F1 Score (F1), Log average miss rate (LAMR), and mean Average Precision (mAP). The specific formula is given in Equation (3).
X T P represents the number of targets accurately identified by the model, X F P indicates the number of targets mistakenly identified by the model, and X F N denotes the number of targets missed by the model. AP is the area surrounded by the PR curve. LAMR is used to measure the miss rate of the target. mAP is used as a comprehensive evaluation metric, and its value is the average of the APs of all detection categories.
p r e c i s i o n ( P ) = X T P X T P + X F P × 100 % r e c a l l ( R ) = X T P X T P + X F N × 100 % A P = 0 1 p ( r ) d r F 1 = 2 P R P + R × 100 % L A M R = 1 M i = 1 M log ( M R i ) m A P = i = 1 K A P i K × 100 %

3. Experiments and Results

In this chapter, we first introduce the datasets used in the experiments. Then, we provide a detailed explanation of the data augmentation methods applied to these datasets. Next, we conduct experiments on datasets with different proportional distributions and present the recorded results.

3.1. Data and Preprocessing

3.1.1. Data Collection

To accentuate the model’s performance more prominently, this experiment uses the Pascal VOC [18] dataset, which contains a total of 20 detection categories. Each target has three types: large, medium, and small, which can fulfill the purpose of detecting different multi-scale targets, as illustrated in Table 1. The Pascal VOC 2007 dataset provides rich object classes and diverse scenes, which makes it an important benchmark for evaluating target detection and other computer vision tasks [39]. It covers a wide range of categories from humans and animals to vehicles and indoor objects, with diverse lighting conditions, viewing angles, and occlusions, which makes it an important tool for testing the robustness and generalizability of algorithms. The numbers 1–20 for subsequent related experiments correspond to the detection targets in the table.

3.1.2. Data Augmentation

Before training the model, after analyzing the dataset, it was found that the dataset was very uneven and that the image size was not uniform. Therefore, the dataset needed to be processed in the preprocessing stage. Image augmentation is a widely adopted technique to expand the dataset and strengthen the model’s robustness [40,41]. Data distribution can be changed regularly by image enhancement techniques, weakening the features of unimportant objects. There are two types of commonly used image enhancement methods. One is image enhancement based on image processing techniques and the other is deep learning based image enhancement algorithms. This paper adopts the first one, which mainly includes rotation, flipping, random cropping, and mosaic stitching. In this case, mosaic stitching is the proportional recombination of four images from the training set into a new image. This enhances the object’s contextual background, allowing the model to learn richer features from a single image. The effect of data enhancement is shown in Figure 7. After data enhancement, the dataset was expanded from 9963 to 15,000 images. The image size is uniformly resized to 400 × 400.

3.2. Results

All experiments are conducted on the Colab cloud server platform, using a Tesla T4 graphics card with 16 GB of video memory and Windows as the operating system. The deep learning framework used is PyTorch 1.9, with 100 training epochs and all images uniformly resized to 400 × 400. To improve model robustness and efficiency, the expanded dataset is divided into three categories, A, B, and C [30], and the proportional allocation of the dataset in each category is depicted in Table 2.
The improved Faster R-CNN is first applied to the training set of three datasets and the results are recorded. During the training process, model error is minimized by adjusting various training parameters over 100 iterations of training. At last, the improved Faster R-CNN is trained using test data from the three datasets. Figure 8 shows the experimental result curves and data distributions during the training process.
By observing the changes and positions of the curves, we know that the accuracy curves of dataset A are steadily increasing and located at the top of the three curves. This indicates that the improved Faster R-CNN has the best performance on dataset A. Meanwhile, the loss value curve of dataset A decreases steadily. The fluctuation is small, located at the bottom of the three curves, and gradually tends to be stable. This also indicates that the improved Faster R-CNN has stronger stability and robustness on dataset A. The variation of these curves can be seen in Figure 8a,c,e,g. Accordingly, by observing the data distribution of the experimental results, it can be found that, compared to the training and testing sets of datasets B and C, the accuracy values of dataset A are more centrally distributed and have higher positions, while the loss values are at lower distribution positions and more compactly distributed. The improved Faster R-CNN is adequately trained on dataset A with better generalization performance. This is because 80% of the images in dataset A are used for training so that the model can extract richer information about image features. Based on these experimental results, dataset A is used for all subsequent experiments.

4. Discussion

4.1. Comparison of Trunk Networks

In this section, comparative experiments [24] are conducted for VGG16, ResNet34, ResNet50, and ResNet101. A total of 15,000 images of dataset A are divided in the ratio of 8:1:1, and SGD is employed to optimize the model [42]. The momentum is 0.9, the learning rate is 0.001, and the F1 threshold is 0.5. After 100 epochs, the model training curve converges, indicating that the model has reached an optimal solution. The improved Faster R-CNN improves the capability of target detection by combining different trunk networks. Table 3 shows the experimental results on the test set.
As can be seen in Table 3, the trunk network with a residual structure has stable detection performance in the multi-scale target category. With the increase of network layers, the detection accuracy of ResNet50 surpasses that of VGG16 and there are fewer missed detections for targets. Since the residual unit module inside the ResNet101 network is connected by a multilevel residual network, this makes it better able to capture the local feature information of multi-scale targets. ResNet101 has the best detection effect, with [email protected] reaching 74.9% and the LAMR reaching 29.5%. Therefore, from Table 3, it can be seen that the improvement of the Faster R-CNN trunk network is effective. ResNet101 performs better in multi-scale target category detection, and the miss rate for small targets is lower.
Figure 9 compares the detection results of the four trunk networks. Focusing on the first two columns, when using ResNet101 as the feature extractor, detection performance is notably improved compared to the other trunk networks. The target detection scores are generally higher, the target localization is more accurate, and there are no target misdetections or missed detections. VGG16 and ResNet50 mistakenly detected the ponytail as a person, while ResNet34 missed the couch and did not detect the chair in the upper right corner. In the third resultant image containing only small targets, ResNet101 can accurately detect the person and the occluded bicycle in the image. This further demonstrates better detection performance when using ResNet101.

4.2. Comparison of Different Object Detectors

To demonstrate that the proposed method has better detection effectiveness and accuracy, experiments are also performed on the test set of the expanded dataset for RetinaNet [43], Faster R-CNN [8], Mask R-CNN [44], YOLOv4 [45], and Cascade R-CNN [9]. The SGD optimizer is used to optimize the model with a momentum of 0.9 and a learning rate of 0.001, with the learning rate decaying 0.1 times every 20 epochs. By continuously adjusting the training parameters, after 100 epochs, the training curves of each comparison model gradually level off. This indicates that the model training process is relatively smooth. Table 4 shows the mean values of each metric for the 20 targets detected by the six comparison models on the test set. Figure 10 shows a comparison of the metric values and their data distributions for the 20 detection targets of the six comparison models on the test set.
Table 4 shows the mean values of each metric for the 20 targets detected by the six comparison models on the test set. Figure 10 shows a comparison of the metric values and their data distributions for the 20 detection targets of the six comparison models on the test set. The improved Faster R-CNN improves [email protected] to 77.8%, F1 to 60.6%, and the LAMR to 26.5%. Compared to the other five detection models, the proposed method has better performance. Specifically, compared with the original Faster R-CNN, the [email protected] is improved by 5.1%, the F1 is improved by 5.3%, and the LAMR is reduced by 4.9%. This indicates that the proposed method is effective in improving the detection rate of small targets as well as reducing the miss rate. In terms of detection speed, the proposed approach is 0.016 s slower than Faster R-CNN. The potential reason may be due to the increase in the number of anchor boxes in the RPN, which has resulted in a longer overall detection time for the proposed approach, slightly decreasing speed but obtaining more accurate detection results. This helps to achieve a balance between time and accuracy.
Observing Figure 10a,b, it can be seen that 15 AP values of the proposed approach are located in the first position, and the data distribution of the prediction results is more centralized than the other compared models. This indicates that the overall detection accuracy of the model is higher. From Figure 10c,d, it can be seen that 17 F1 values are in the leading position, among which two F1 values exceed 0.8, and the distribution intervals of the F1 values are relatively higher. This indicates that the model is more adaptable when facing multi-scale target categories. As shown in Figure 10e,f, the prediction results of the proposed approach all have LAMR values below 0.5, and the miss rate is also reduced for small targets such as pottedplant (1), sheep (4), and bottle (5). The reason for this may be that the bounding box optimization mechanism, as well as the introduced multi-scale training strategy, plays a role. The data distribution interval of the LAMR is also relatively lower. This indicates that the model is able to focus on small targets that are difficult to detect when performing multi-scale target category detection, and the model is more robust and stable. Figure 11 shows the visualization of the detection image after mosaic data enhancement. Compared to the other detection models, although the proposed approach is not the best in each target category, the overall effect is excellent. The very small car in the upper right corner, as well as the occluded chair in the lower left corner, can be accurately detected, and both target detection scores are relatively quite high. This demonstrates that the proposed approach can better adapt to the different sizes and shapes of targets in different multi-scale detection target environments and improve the accuracy of small target detection.

4.3. The Ablation Experiments and Analysis

In this section, we set up seven sets of ablation experiments on the test set to verify the validity of each module. The setting of the ablation experiments is detailed in Table 5. The SGD optimizer is used to optimize each combined model [42]. The momentum is 0.9 and the batch size is 64. To mitigate overfitting during training, the regularization parameter is 0.01 and the learning rate is 0.001. The number of training times for each combination of models is still 100 epochs, and to achieve timely adjustments to the training parameters, the training results are taken every 10 epochs. When the training curves of each combined model gradually level off, it indicates that the model can converge stably. The results of the ablation study are presented in Table 5.
As depicted in Table 5, the addition of the modified RPN to the Faster R-CNN (ResNet101) increases the [email protected] by 0.3% and the F1 by 0.2%. The introduction of DIOU improves the [email protected] by 0.2% and the F1 by 0.1%. It can be seen that the introduction of OHEM increases the [email protected] by 0.5% and the F1 by 0.4%. This shows its improvement for the problems of sample imbalance and insufficient training of hard case samples in the target detection task. After replacing NMS with Soft-NMS, the miss rate is reduced, while performance and accuracy are improved. Finally, by combining the modules with a multi-scale training strategy, an improved Faster R-CNN is obtained with a mAP of 77.8%, which is 5.1% higher than the original Faster R-CNN, and the F1 is improved by 5.3%. It is worth noting that the LAMR of the proposed approach compared to Improve5 is lower. This improvement is attributed to the efficacy of the MST [16] in reducing the miss rate and enhancing the model’s recognition capabilities for small targets. The partial detection results for Faster R-CNN, Improved5, and Improved Faster R-CNN are shown in Figure 12. A comparison shows that Faster R-CNN is not very effective in detecting cows and bottles with small sizes and misses some small targets. While Improved5 improves this phenomenon, the network can detect some small targets that were originally missed. The best detection was achieved by Improved Faster R-CNN, which had more accurate edge localization, detected more small targets, and largely correctly identified overlapping and low-contrast targets. The detection performance of this model is much stronger.

5. Conclusions

This paper proposes a novel two-stage object detection model for detecting multi-scale objects from diverse categories. The model introduces improvements to the Faster R-CNN and its trunk feature extraction network. DIOU, OHEM, and Soft-NMS are used to improve the problems of unbalanced positive and negative samples and target miss rate during model training. The RPN is also optimized and the proposed approach is trained by employing a multi-scale training strategy. Comparison experiments with trunk networks verify that using the ResNet101 feature extraction network is more advantageous. The validity of the proposed approach is further confirmed by comparison experiments with other detection models. Ablation experiments are also conducted to verify that the modules in the proposed approach can indeed be useful. The experiments show that the proposed method has a [email protected] of 77.8% and an F1 of 60.6%, which are 5.1% and 5.3% higher than the original Faster R-CNN, respectively. The experimental results show that the proposed method can improve the accuracy and performance of object detection in a multi-scale target detection environment. In the future, we will further optimize, extend, and experiment on more datasets so that the model can be better applied to different types of object detection scenarios and provide a roadmap for continued advancements in the field of multi-scale object detection.

Author Contributions

Q.C. and M.L. designed the study; Q.C. and Z.L. performed the research; M.L. and J.Z. conceived the idea; Q.C., J.Z. and Z.L. provided and analyzed the data; Q.C., M.L. and L.G. helped perform the analysis with constructive discussions. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant Numbers: 51663001, 52063002, 42061067 and 61741202).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset is available free of charge at kaggle. (https://www.kaggle.com/datasets/qianyongchen/dataset, accessed on 10 August 2024).

Acknowledgments

The authors thank the anonymous reviewers and editors for their valuable comments and constructive suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zeng, N.; Wu, P.; Wang, Z.; Li, H.; Liu, W.; Liu, X. A Small-Sized Object Detection Oriented Multi-Scale Feature Fusion Approach with Application to Defect Detection. IEEE Trans. Instrum. Meas. 2022, 71, 1–14. [Google Scholar] [CrossRef]
  2. Deng, Y.; Hu, X.L.; Li, B.; Zhang, C.X.; Hu, W.M. Multi-scale self-attention-based feature enhancement for detection of targets with small image sizes. Pattern Recogn. Lett. 2023, 166, 46–52. [Google Scholar] [CrossRef]
  3. Ma, Y.L.; Wang, Q.Q.; Cao, L.; Li, L.; Zhang, C.J.; Qiao, L.S.; Liu, M.X. Multi-Scale Dynamic Graph Learning for Brain Disorder Detection with Functional MRI. IEEE Trans. Neur. Syst. Rehabil. 2023, 31, 3501–3512. [Google Scholar] [CrossRef] [PubMed]
  4. Menezes, A.G.; de Moura, G.; Alves, C.; de Carvalho, A. Continual Object Detection: A review of definitions, strategies, and challenges. Neural Netw. 2023, 161, 476–493. [Google Scholar] [CrossRef] [PubMed]
  5. Xu, S.B.; Zhang, M.H.; Song, W.; Mei, H.B.; He, Q.; Liotta, A. A systematic review and analysis of deep learning-based underwater object detection. Neurocomputing 2023, 527, 204–232. [Google Scholar] [CrossRef]
  6. Goswami, P.K.; Goswami, G. A Comprehensive Review on Real Time Object Detection using Deep Learing Model. In Proceedings of the 2022 11th International Conference on System Modeling & Advancement in Research Trends (SMART), Moradabad, India, 16–17 December 2022; pp. 1499–1502. [Google Scholar]
  7. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Coiumbus, OH, USA, 23–28 June 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 580–587. [Google Scholar]
  8. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. 2017, 28, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  9. Cai, Z.; Vasconcelos, N. Cascade r-cnn: Delving into high quality object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Salt Lake City, UT, USA, 18–22 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 6154–6162. [Google Scholar]
  10. Wan, S.; Goudos, S. Faster R-CNN for multi-class fruit detection using a robotic vision system. Comput. Netw. 2020, 168, 107036. [Google Scholar] [CrossRef]
  11. Yu, Y.; Zhang, K.; Yang, L.; Zhang, D. Fruit detection for strawberry harvesting robot in non-structural environment based on Mask-RCNN. Comput. Electron. Agric. 2019, 163, 104846. [Google Scholar] [CrossRef]
  12. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 779–788. [Google Scholar]
  13. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the 2016 14th European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  14. Liang, Q.; Zhu, W.; Long, J.; Wang, Y.; Sun, W.; Wu, W. A real-time detection framework for on-tree mango based on SSD network. In Proceedings of the 2018 11th International Conference on Intelligent Robotics and Applications (ICIRA), Newcastle, NSW, Australia, 9–11 August 2018; pp. 423–436. [Google Scholar]
  15. Anagnostis, A.; Tagarakis, A.C.; Asiminari, G.; Papageorgiou, E.; Kateris, D.; Moshou, D.; Bochtis, D. A deep learning approach for anthracnose infected trees classification in walnut orchards. Comput. Electron. Agric. 2021, 182, 105998. [Google Scholar] [CrossRef]
  16. Tian, X.X.; Bi, C.K.; Han, J.; Yu, C. EasyRP-R-CNN: A fast cyclone detection model. Vis. Comput. 2024, 40, 4829–4841. [Google Scholar] [CrossRef]
  17. Li, W.; Zhang, L.C.; Wu, C.H.; Cui, Z.X.; Niu, C. A new lightweight deep neural network for surface scratch detection. Int. J. Adv. Manuf. Tech. 2022, 123, 1999–2015. [Google Scholar] [CrossRef] [PubMed]
  18. Tong, K.; Wu, Y. Rethinking PASCAL-VOC and MS-COCO dataset for small object detection. J. Vis. Commun. Image R. 2023, 93, 103830. [Google Scholar] [CrossRef]
  19. Demir, A.; Yilmaz, F.; Kose, O. Early detection of skin cancer using deep learning architectures: Resnet-101 and inception-v3. In Proceedings of the 2019 Medical Technologies Congress (TIPTEKNO 2019), Izmir, Turkey, 3–5 October 2019; pp. 1–4. [Google Scholar]
  20. Shrivastava, A.; Gupta, A.; Girshick, R. Training region-based object detectors with online hard example mining. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 761–769. [Google Scholar]
  21. Bodla, N.; Singh, B.; Chellappa, R.; Davis, L.S. Soft-NMS--improving object detection with one line of code. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 5561–5569. [Google Scholar]
  22. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the 2020 20th AAAI Conference on Artificial Intelligence (AAAI-20), New York, NY, USA, 7–12 February 2020; pp. 12993–13000. [Google Scholar]
  23. Tian, R.; Shi, H.; Guo, B.; Zhu, L. Multi-scale object detection for high-speed railway clearance intrusion. Appl. Intell. 2022, 52, 3511–3526. [Google Scholar] [CrossRef]
  24. Wang, H.; Xiao, N. Underwater object detection method based on improved Faster RCNN. Appl. Sci. 2023, 13, 2746. [Google Scholar] [CrossRef]
  25. Lu, X.; Wang, H.; Zhang, J.J.; Zhang, Y.T.; Zhong, J.; Zhuang, G.H. Research on J wave detection based on transfer learning and VGG16. Biomed. Signal. Process. 2024, 95, 106420. [Google Scholar] [CrossRef]
  26. Pal, S.K.; Pramanik, A.; Maiti, J.; Mitra, P. Deep learning in multi-object detection and tracking: State of the art. Appl. Intell. 2021, 51, 6400–6429. [Google Scholar] [CrossRef]
  27. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition(CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 770–778. [Google Scholar]
  28. Corso, M.P.; Stefenon, S.F.; Singh, G.; Matsuo, M.V.; Perez, F.L.; Leithardt, V.R.Q. Evaluation of visible contamination on power grid insulators using convolutional neural networks. Electr. Eng. 2023, 105, 3881–3894. [Google Scholar] [CrossRef]
  29. Chan, S.X.; Tao, J.; Zhou, X.L.; Bai, C.; Zhang, X.Q. Siamese Implicit Region Proposal Network with Compound Attention for Visual Tracking. IEEE Trans. Image Process. 2022, 31, 1882–1894. [Google Scholar] [CrossRef] [PubMed]
  30. Sha, G.; Wu, J.; Yu, B. The improved faster-RCNN for spinal fracture lesions detection. J. Intell. Fuzzy Syst. 2022, 42, 5823–5837. [Google Scholar] [CrossRef]
  31. Zhou, D.; Fang, J.; Song, X.; Guan, C.; Yin, J.; Dai, Y.; Yang, R. Iou loss for 2d/3d object detection. In Proceedings of the 2019 International Conference on 3D Vision (3DV), Québec City, QC, Canada, 16–19 September 2019; pp. 85–94. [Google Scholar]
  32. Shen, Y.Y.; Zhang, F.Z.; Liu, D.; Pu, W.H.; Zhang, Q.L. Manhattan-distance IOU loss for fast and accurate bounding box regression and object detection. Neurocomputing 2022, 500, 99–114. [Google Scholar] [CrossRef]
  33. Wang, Z.H.; Jiang, Q.P.; Zhao, S.S.; Feng, W.S.; Lin, W.S. Deep Blind Image Quality Assessment Powered by Online Hard Example Mining. IEEE Trans. Multimed. 2023, 25, 4774–4784. [Google Scholar] [CrossRef]
  34. Li, W.B.; Wang, Q.; Gao, S. PF-YOLOv4-Tiny: Towards Infrared Target Detection on Embedded Platform. Intell. Autom. Soft Comput. 2023, 37, 921–938. [Google Scholar] [CrossRef]
  35. Xiao, L.; Wu, B.; Hu, Y. Surface defect detection using image pyramid. IEEE Sens. J. 2020, 20, 7181–7188. [Google Scholar] [CrossRef]
  36. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-End Object Detection with Learnable Proposals. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 14449–14458. [Google Scholar]
  37. Fang, H.; Ding, L.; Wang, L.; Chang, Y.; Yan, L.; Han, J. Infrared Small UAV Target Detection Based on Depthwise Separable Residual Dense Network and Multiscale Feature Fusion. IEEE Trans. Instrum. Meas. 2022, 71, 1–20. [Google Scholar] [CrossRef]
  38. Smart, P.D.S.; Thanammal, K.K.; Sujatha, S.S. An Ontology Based Multilayer Perceptron for Object Detection. Comput. Syst. Sci. Eng. 2023, 44, 2065–2080. [Google Scholar] [CrossRef]
  39. Zhang, X.; Zhao, C.; Luo, H.Z.; Zhao, W.Q.; Zhong, S.; Tang, L.; Peng, J.Y.; Fan, J.P. Automatic learning for object detection. Neurocomputing 2022, 484, 260–272. [Google Scholar] [CrossRef]
  40. Chen, J.A.; Tam, D.; Raffel, C.; Bansal, M.; Yang, D.Y. An Empirical Survey of Data Augmentation for Limited Data Learning in NLP. Trans. Assoc. Comput. Linguist 2023, 11, 191–211. [Google Scholar] [CrossRef]
  41. Shi, J.; Ghazzai, H.; Massoud, Y. Differentiable Image Data Augmentation and Its Applications: A Survey. IEEE Trans. Pattern Anal. 2024, 46, 1148–1164. [Google Scholar] [CrossRef]
  42. Gower, R.M.; Loizou, N.; Qian, X.; Sailanbayev, A.; Shulgin, E.; Richtárik, P. SGD: General analysis and improved rates. In Proceedings of the International Conference on Machine Learning (ICML), Long Beach, CA, USA, 9–15 June 2019; pp. 5200–5209. [Google Scholar]
  43. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  44. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask r-cnn. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2961–2969. [Google Scholar]
  45. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
Figure 1. General framework of the improved Faster R-CNN.
Figure 1. General framework of the improved Faster R-CNN.
Jimaging 10 00197 g001
Figure 2. Resnet101 network composition. (a) Conv Block. (b) Identity Block. (c) Residual block. (d) Resnet101 network structure.
Figure 2. Resnet101 network composition. (a) Conv Block. (b) Identity Block. (c) Residual block. (d) Resnet101 network structure.
Jimaging 10 00197 g002
Figure 3. Optimized RPN schematic. (a) RPN sliding window and anchors. (b) Modified RPN anchors. (c) Modified RPN.
Figure 3. Optimized RPN schematic. (a) RPN sliding window and anchors. (b) Modified RPN anchors. (c) Modified RPN.
Jimaging 10 00197 g003
Figure 4. Schematic representation of the different levels of overlap of the bounding box. (a) Three possible scenarios when IoUs are identical. (b) DIOU loss for bounding box regression.
Figure 4. Schematic representation of the different levels of overlap of the bounding box. (a) Three possible scenarios when IoUs are identical. (b) DIOU loss for bounding box regression.
Jimaging 10 00197 g004
Figure 5. Diagram of the OHEM algorithm.
Figure 5. Diagram of the OHEM algorithm.
Jimaging 10 00197 g005
Figure 6. Diagram of the multi-scale training strategy.
Figure 6. Diagram of the multi-scale training strategy.
Jimaging 10 00197 g006
Figure 7. Data enhancement effect diagram.
Figure 7. Data enhancement effect diagram.
Jimaging 10 00197 g007
Figure 8. The result curves and data distribution for datasets A, B, and C. (a) Training accuracy curve. (b) Distribution of training accuracy data. (c) Training loss curve. (d) Distribution of training loss data. (e) Test accuracy curve. (f) Distribution of test accuracy data. (g) Test loss curve. (h) Distribution of test loss data.
Figure 8. The result curves and data distribution for datasets A, B, and C. (a) Training accuracy curve. (b) Distribution of training accuracy data. (c) Training loss curve. (d) Distribution of training loss data. (e) Test accuracy curve. (f) Distribution of test accuracy data. (g) Test loss curve. (h) Distribution of test loss data.
Jimaging 10 00197 g008
Figure 9. Visual comparison of detection results for four backbone networks.
Figure 9. Visual comparison of detection results for four backbone networks.
Jimaging 10 00197 g009
Figure 10. Comparison of experimental result values and their data distribution for 20 detection targets for each comparison model. (a,b) Comparisons of AP values. (c,d) Comparisons of F1 values. (e,f) Comparisons of LAMR values.
Figure 10. Comparison of experimental result values and their data distribution for 20 detection targets for each comparison model. (a,b) Comparisons of AP values. (c,d) Comparisons of F1 values. (e,f) Comparisons of LAMR values.
Jimaging 10 00197 g010
Figure 11. Visual comparison of detection results for six comparison models.
Figure 11. Visual comparison of detection results for six comparison models.
Jimaging 10 00197 g011
Figure 12. Visual comparison of detection results for ablation experiments.
Figure 12. Visual comparison of detection results for ablation experiments.
Jimaging 10 00197 g012
Table 1. Dataset distribution of Pascal VOC 2007.
Table 1. Dataset distribution of Pascal VOC 2007.
Object ClassesTrainValidationTestTotal
Pottedplant133112224469
Chair224221417862
Sofa111118223452
sheep484897193
bottle139105212456
diningtable97103190390
Bird180150282612
boat81100172353
aeroplane112126204442
motorbike120125222467
tvmonitor128128229485
person102598320074015
train127134259520
bicycle116127239482
Cow6972127268
Dog203218418839
Cat163174322659
Bus9789174360
Car3763377211434
horse139148274561
total2501251049529963
Table 2. Proportional distribution of the three datasets.
Table 2. Proportional distribution of the three datasets.
Data SetTrainValidationTest
A (8:1:1)12,00015001500
B (7:1.5:1.5)10,50022502250
C (6:2:2)900030003000
Table 3. Predictive performance of different trunk networks.
Table 3. Predictive performance of different trunk networks.
Methods[email protected] (%)F1 (%)LAMR (%)
VGG1672.755.331.4
ResNet3473.356.430.8
ResNet5073.856.230.2
ResNet10174.957.229.5
Table 4. Experimental mean results for six comparison models.
Table 4. Experimental mean results for six comparison models.
Methods[email protected] (%)F1 (%)LAMR (%)T (s)
RetinaNet75.658.529.20.155
Faster R-CNN72.755.331.40.147
Mask R-CNN75.858.228.50.153
YOLOv476.559.227.20.132
Cascade R-CNN76.259.728.10.139
This Paper77.860.626.50.163
Table 5. Setting of ablation experiments and experimental results.
Table 5. Setting of ablation experiments and experimental results.
MethodResNet101RPNDIOUOHEMSoft-NMSMST[email protected] (%)F1 (%)LAMR (%)
Faster R-CNN(VGG16) 72.755.331.4
Faster R-CNN(ResNet101) 74.957.229.5
Improve1 75.257.429.2
Improve2 75.157.329.3
Improve3 75.457.628.9
Improve4 75.057.429.3
Improve5 75.558.128.7
This paper77.860.626.5
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Q.; Li, M.; Lai, Z.; Zhu, J.; Guan, L. A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms. J. Imaging 2024, 10, 197. https://doi.org/10.3390/jimaging10080197

AMA Style

Chen Q, Li M, Lai Z, Zhu J, Guan L. A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms. Journal of Imaging. 2024; 10(8):197. https://doi.org/10.3390/jimaging10080197

Chicago/Turabian Style

Chen, Qianyong, Mengshan Li, Zhenghui Lai, Jihong Zhu, and Lixin Guan. 2024. "A Multi-Scale Target Detection Method Using an Improved Faster Region Convolutional Neural Network Based on Enhanced Backbone and Optimized Mechanisms" Journal of Imaging 10, no. 8: 197. https://doi.org/10.3390/jimaging10080197

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop