Next Article in Journal
Exploiting On-Chip Voltage Regulators for Leakage Reduction in Hardware Masking
Previous Article in Journal
An Efficient CS-Based Spectral Peak Search Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

PDC: Pearl Detection with a Counter Based on Deep Learning

1
College of Mechanical Engineering, Guangdong Ocean University, Zhanjiang 524088, China
2
Agricultural Machinery Appraisal and Extension Station in Hainan, Haikou 570206, China
3
Guangdong Marine Equipment and Manufacturing Engineering Technology Research Center, Zhanjiang 524088, China
4
South China of Marine Science and Engineering Guangdong Laboratory, Zhanjiang 524088, China
*
Authors to whom correspondence should be addressed.
Sensors 2022, 22(18), 7026; https://doi.org/10.3390/s22187026
Submission received: 14 August 2022 / Revised: 1 September 2022 / Accepted: 8 September 2022 / Published: 16 September 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Pearl detection with a counter (PDC) in a noncontact and high-precision manner is a challenging task in the area of commercial production. Additionally, sea pearls are considered to be quite valuable, so the traditional manual counting methods are not satisfactory, as touching may cause damage to the pearls. In this paper, we conduct a comprehensive study on nine object-detection models, and the key metrics of these models are evaluated. The results indicate that using Faster R-CNN with ResNet152, which was pretrained on the pearl dataset, [email protected] = 100% and [email protected] = 98.83% are achieved for pearl recognition, requiring only 15.8 ms inference time with a counter after the first loading of the model. Finally, the superiority of the proposed algorithm of Faster R-CNN ResNet152 with a counter is verified through a comparison with eight other sophisticated object detectors with a counter. The experimental results on the self-made pearl image dataset show that the total loss decreased to 0.00044. Meanwhile, the classification loss and the localization loss of the model gradually decreased to less than 0.00019 and 0.00031, respectively. The robust performance of the proposed method across the pearl dataset indicates that Faster R-CNN ResNet152 with a counter is promising for natural light or artificial light peal detection and accurate counting.

1. Introduction

For centuries, people have been captivated by the natural beauty of pearls. With their lustrous sheen and sumptuous colors, pearls, as shown in Figure 1, are regarded as one of nature’s most glorious treasures. Pearls are created to protect mollusks against irritants, and they are processed into exquisite necklaces after cultivation. Pearl detection with a counter (PDC) is an important and necessary step during pearl processing. Automatic counting systems [1,2] for pearls reduce the human working time significantly, compared with traditional manual counting methods. A noncontact counting system for pearls is recommended to achieve clean and precise measurements [3]. In general, the computer vision-based noncontact counting system consists of two stages, object detection [4], and object counting [5].
Stage 1 involves object detection based on deep learning. Object detection is a computer technology related to computer vision and image processing that deals with detecting instances of semantic objects of a certain class (such as humans, buildings, or cars) in digital images and videos. Fortunately, however, the most successful approaches to object detection are currently extensions of deep learning-based models [6,7]. In 2021, Google released a new object detection model zoo for TensorFlow2, and then the prebuilt models and weights for a few specific models could be applied, which are as follows: single-shot multibox detector (SSD) with MobileNets v1 and v2, Faster-RCNN with ResNet 50,101 and 152, SSD ResNet 50, 101 and 152. According to the statistics from Web of Science, in the past five years, SSD MobileNet was adopted in a total of 107 papers [8,9], accounting for 43.7%, Faster R-CNN with ResNet in 77 papers, accounting for 31.4% [10,11], and SSD ResNet in 61 papers [12,13], accounting for 24.9%, which are three commonly used models, as shown in Figure 2.
Stage 2 involves object counting based on computer vision. The goal of object counting is to count the number of object instances in a single image or video sequence. It has many real-world applications, such as automatic bird counting [14], vehicle counting [15], and counting surgical instruments [16]. According to the Web of Science Core Collection Database, Figure 3 shows the statistics about the number of object counting papers published in the computer vision field. Compared with the period from 2013 to 2016, it is obvious that the relevant research results have increased significantly in the past five years. This is mainly due to the deep learning applications for computer vision, such as shrimp egg counting [17], people counting [18], and distribution line pole counting [19].
The Industry 4.0 paradigm will be the next step in the evolution of manufacturing, which is based on monitoring and pervasive data collection. However, the manufacturing industry has faced growing challenges in recent years. Computer vision can alleviate the biggest challenges. A subset of artificial intelligence allows computers to take in information from digital images and then make decisions based on that information. Mohamed et al. proposed a deep learning-based method, which was capable of detecting the defective paddy rice seedlings with the highest precision. It can facilitate a significant amount of work in agriculture, especially in paddy cultivation [20]. To promote the farming industry, Xinze Zheng et al. used multiple deep neural network models in the object detection and sex classification of ducks. In addition, the accuracy score of 99.29%, the F1 score of 98.60%, and 269.68 fps were achieved [21]. With the advance of Industry 4.0 and deep learning technology, Haonan Yin et al. proposed an adaptive and robust calculating algorithm to greatly save labor and material costs, which successfully facilitates inventories for stacked goods in the stereoscopic warehouse [22].
PyTorch and TensorFlow are the two most popular deep learning frameworks in 2022 by far. TensorFlow has been the go-to framework for deployment-oriented applications since its inception, and for good reason. However, PyTorch used to be extremely lackluster from a deployment perspective, although it has worked on closing this gap in recent years. From the perspective of the ease of deployment of the object detection models, we will conduct an in-depth evaluation of the nine commonly used target detection models in the TensorFlow2 model zoo, such as Faster R-CNN, SSD MobileNet, and SSD ResNet.
Balaji et al. proposed an application of a deep convolutional network (DeepSort) for the sorting and counting of haploid seeds [23]. Similarly, to solve the problems of low capability of real-time detection and counting in traffic flow statistics, Liang Mu et al. built a Yolov5 with an improved DeepSort algorithm [24]. However, despite the efforts and advances made on applying object detection models with DeepSort, it is still slow due to the base of object counting and tracking in a frame. This study is focused on the problems such as tracking. To summarize, the main contributions of this study are as follows:
(i)
A model is developed to automatically recognize pearls based on CNN deep learning,
(ii)
The accuracy of pearl detection models is evaluated using images taken from nature and artificial light images,
(iii)
An efficient algorithm is presented based on CNN and a corresponding computer vision setup is developed for pearl counting in densely distributed cases.

2. Related Work

2.1. Object Detection

Deep learning-based methods have demonstrated great success in many computer vision fields, including multiple-object recognition [25,26], identifying objects using cameras [27,28] and object classification [29,30]. In general, there are two different approaches for object detection, it can either make a fixed number of predictions on a grid (one stage) or leverage a proposal network to find objects and then use a second network to fine-tune these proposals and output a final prediction (two stages). The state-of-the-art one-stage object detection methods achieve good performance on datasets such as MS COCO [31], PASCAL VOC [32] and customer personal collection [33]. One-stage detectors that can predict all the bounding boxes and class probabilities in a single pass with high inference speeds are more suitable for real-time applications. Two-stage detectors have low inference speeds for the intermediate layer, which is used to propose possible object regions, but the two-stage detectors still take the lead in accuracy. The most common examples of one-stage object detectors are YOLO [34,35], SSD [36,37], SqueezeDet [38], and DetectNet [39,40]. Two-stage methods prioritize detection accuracy, and typical models include Faster R-CNN [41,42], Mask R-CNN [43,44] and Cascade R-CNN [45,46].

2.2. Object Count

The goal of the deep learning-based object counting task is to count the number of object instances in a single image or video sequence. It has many real-world applications. Balakrishnan presented the Faster RCNN ResNet101 algorithm, which was used to detect the number of false ceilings [47]. A deep learning recognition and counting method for the vehicle-road collaborative information was conducted by Hongbin Jiao [48]. Zishuo Huang and Qinyou Hu proposed an improved single shot multibox detector and DeepSORT algorithms and performed numerical experiments, indicating that the ship counting system is more accurate [49]. Beibei Xu, et al. adopted Mask R-CNN with feature extraction to present a low-cost approach for livestock counting in animals in real time [50].
The object count consists of two categories, which are elaborated in the following section.
With regard to statistical images method counting (SIMC), in [51], a computer-vision-based system was used to count fish by combining the information from blob detection, a mixture of Gaussians and a Kalman filter. According to the authors, the proposed method is a feasible approach for automatic fish counting, reducing costs and boosting production, as it increases labor availability. Two of the most common methods, Faster R-CNN with InceptionV2 and single-shot multibox detector (SSD) with MobileNet, have been tested, as in [52]. In fruit statistical image counting, the accuracy of Vasconez’s system reaches up to 93% (overall for all fruits) using Faster-RCNN with InceptionV2, and 90% (overall for all fruits) using SSD with MobileNet.
With regard to object tracking and counting ID (OTC), to realize automatic detection and retrain the feature extractor for multi-object tracking, a real-time vehicle tracking counter was proposed for vehicles that combines vehicle detection and vehicle tracking algorithms to realize traffic flow detection [53]. By using the ant colony algorithm and A∗ algorithm, Zihan Jiang [54] proposed an improved A∗ algorithm and combined it with the DeepSORT algorithm and YOLOv5 algorithm to carry out variable path planning. Some of the relative works are shown in Table 1.

3. Methods

3.1. Image Acquisition and Annotation

To design an efficient machine vision-based detection and count system, every aspect of the procedure from the pearl dataset to the implementation phase should be considered carefully. Such a small pearl cannot provide enough information for detection, as it has few pixels. It would be difficult to perfectly recognize every pearl through a whole image. In this study, to collect a proper dataset, several important terms were considered, including natural and artificial light source, angle of view, and the number of pearls.
In this working environment, the pearl images were captured by a 640 × 480 maximum photo resolution computer USB webcam (Figure 4a). It covered a total of 3000 images, which are divided into 2 categories, namely natural light images and artificial light images, as listed in Table 2. After the annotation process was performed, the 3000 images in the original image dataset were randomly separated into training and test image sets. To increase the model accuracy, these images were carefully labelled (Figure 4b). After the annotation process was performed, the ratio to the total image was 9:1, 90% of the pearl images were used for training, and the remaining 10% were maintained for testing, which consisted of 13,850 pearls and 1563 pearls.

3.2. Detector and Counting Framework

The algorithm is intended to reduce the computational burden and increase the accuracy of the base model. The detector and counter consist of two pipeline stages, as shown in Figure 5.
(i)
Detector: Faster R-CNN ResNet152v1. A pearl image was first input to the convolutional backbone that can extract image features and output a feature map. ResNet152 was used as a substructure and greatly improves the effect of target detection. The region proposal network (RPN) was introduced to generate candidate regions, and the RoI pooling and classifier layer received both the optimized candidate boxes of RPN output and Conv feature map output. Finally, the classification and bounding box regression were predicted.
(ii)
Counter. As shown in Figure 5, the threshold was usually between 0.7 and 0.8. If the confidence value of detected people was below the threshold, the pearls were not counted. In this paper, the default threshold value of 0.8 was used for the sake of simplicity. Then, more bounding boxes of pearls were generated, and we found that pearl counting works better with more detected boxes and higher confidence values by a series of experiments.

3.3. Faster R-CNN ResNet152 as Detector

In this study, the faster region-based convolutional neural network (R-CNN) ResNet 152 is used for pearl detection and classification tasks. As a two-stage detector, Faster R-CNN has a higher localization and object recognition accuracy and comprises the RPN and Fast-RCNN modules. As shown in Figure 6, Faster R-CNN with ResNet consists of 152 convolution layers. Remarkably, ResNet152 is a pretrained deep CNN algorithm that performs the feature extraction task, and the 152-layer ResNet (11.3 billion FLOPs) still has lower complexity than VGG-16/19 nets (15.3/19.6 billion FLOPs) [62]. Here, ResNet152 contains the following 6 blocks: Conv1, Con2_x, Con3_x, Con4_x, Con5_x, and a fully connected layer. Because each unit contains 3 convolution layers, the total number of ResNet-152 layers is (3 + 8 + 36 + 3) × 3 = 150. If the first convolution layer and the last full connection layer are added, the total number of ResNet-152 layers is 150 + 2 = 152 layers.
A region proposal network (RPN) takes an image (of any size) as input and outputs a set of rectangular object proposals, which is an alternative to “selective search” in Fast-RCNN. Anchor boxes are predetermined boundary boxes in the RPN with a specific height and width. RPN uses the final Conv feature map as input and performs a 3 × 3 Conv sliding window operation spatially. Nine anchor boxes are generated according to center point of the sliding window. The feature map is reshaped into the softmax layer, which applies a softmax activation function to the next reshape block. The proposals generated by the RPN and the shared convolutional features are fed into the region of interest (RoI) pooling layer. The positive samples are taken up by the proposal region, with the labelled ground-truth intersection over union (IoU) value being larger than the set threshold. Finally, the extracted characteristics are transformed into a constant feature size of 7 × 7 through RoI pooling.
The scaling of RoI includes the following steps:
  • Producing the fixed-size feature maps from non-uniform inputs,
  • Finding a feature map obtained from a CNN after multiple convolutions and pooling layers
  • Indicating the index and the proposal coordinates.
We have our RoI mapped onto the feature map that takes every ROI from the input and a section of input feature map, which corresponds to that ROI and converts that feature-map section into a fixed dimension map. In this study, we applied max pooling, so this process is completed on the whole RoI matrix rather than only on the topmost layer. The output fixed dimension of the ROI pooling solely depends on the layer parameters.
All bounding boxes with high confidence in the anchor box are selected for pearl detection tasks and sent to the full convolution layer through RoI pooling to obtain the category confidence and regression box of the detected image. In this study, the value of classification = 1; therefore, the positioning module uses fully connected layers to calculate the final category confidence and target position coordinates.

3.4. Counting Pearl

In the core of the pearl counter code, the images are analyzed by the software, looking for detection scores, boxes and classes. As shown in Figure 7, we develop all outputs as batch tensors, which are converted to NumPy arrays, and take index to remove the batch dimension. This framework allowed the input of a pearl image patch of any size within the memory of the GPU, and we slightly modified the data input of the network to fit the pearl dataset. The output detection scores, boxes and classes are fed into the object counter, and the counting range is set up between Min_Conf_Thresh and 1. It predicts a bounding box that involves the (Xmin, Ymin), (Xmax, Ymax) coordinates and the width and height and a metric of valuation of the quality confidence score.
As a rectangle has exactly four corners, we can simply count the number of corners in the detected pearl image. The approach includes the following steps:
(i)
Obtain a binary pearl image. Load detected pearl image, greyscale, Gaussian blur, and Otsu’s threshold;
(ii)
Remove small noise. We find contours and then filter them by contour area filtering with cv2.contourArea and remove the noise by filling in the contour with cv2.drawContours;
(iii)
Find corners. The Shi-Tomasi Corner Detector that has already been implemented, as cv2.goodFeaturesToTrack is used for corner detection;
(iv)
Increase count until Max_Conf_Thresh = 1. The total number of these possible rectangles is obtained, and the interpreter can return the coordinates that are the edges of the image dimensions.
Finally, the object counter can adapt to the different detection scores and boxes and provide a high counting accuracy performance. Furthermore, the label text, boxes and the total number of detections will be obtained from the model with high precision.

3.5. Evaluation Metrics

3.5.1. Intersection over Union, IoU

Bounding box regression is an extremely important and necessary task in pearl processing detection. Intersection over union (IoU), as shown in Figure 8, illustrates the graphical view of the equation below.
IoU = Area   of   Overlap Area   of   Union
In general, when IoU 0.5 , the pearl prediction is regarded as correct. The confidence score is also calculated (see Figure 9), which reflects how likely the box contains a pearl and how accurate the bounding box is.

3.5.2. The Detection of Pearl Accuracy, Recall, Precision and mAP

The detection of pearl accuracy is the ratio of correct pearl predictions to the total number of input samples.
accuracy = correct   pearl   predictions   all   pearl   predictions
The detection of pearl precision is the ratio of relevant examples (true positives) among all the examples, which are defined by the following equation:
precision = true   positives true   positives + false   positives
On the other hand, the recall is the ratio of examples, which implicates how well we find positives.
recall = true   positives true   positives + false   negatives
When working with pearl detection, a common approach to evaluate the effect is used to calculate the mean average precision (mAP). The average precision (AP) of the pearl class is calculated by measuring the precision for all occurrences of the object class in the pearl test dataset.

3.5.3. Loss Function

L p i , t i is the Faster R-CNN loss function that optimizes the classification of each region of interest (ROI). Our loss function for a pearl image is calculated through
L p i , t i = 1 N c l s i L c l s ( p i , p i * ) + λ 1 N r e g i p i * L r e g ( t i , t i * )
L c l s ( p i , p i * ) = ln p i p i * + ( 1 p i ) ( 1 p i * )
L r e g ( t i , t i * ) = s m o o t h L 1 ( t i , t i * )
where L c l s ( p i , p i * ) and L r e g ( t i , t i * ) are classification loss and regression loss, respectively. L c l s ( p i , p i * ) is defined by L c l s ( p i , p i * ) = ln p i p i * + ( 1 p i ) ( 1 p i * ) , and L r e g ( t i , t i * ) is defined by L r e g ( t i , t i * ) = s m o o t h L 1 ( t i , t i * ) , where
s m o o t h L 1 ( x ) = 0.5 × x 2 ,     i f x < 1 x 0.5 ,     o t h e r w i s e
For the regression loss, we use L r e g ( t i , t i * ) = R ( t i t i * ) , where R is the robust loss function (smoothL1). λ is a hyperparameter that controls the balance between the losses, normally set to λ = 1 . t i and t i * are the positions of the predicted bounding boxes and ground-truth boxes on the feature maps, obtained by Equations (9)–(12),
t x = ( x x a ) ω a ,   t y = y y a h a
t w = log ( ω ω a ) ,   t h = log ( h h a )
t x * = ( x * x a ) ω a ,   t y * = ( y * y a ) h a
t w * = log ( ω * ω a ) ,   t h * = log ( h * h a )
where each grid cell predicts a bounding box that involves the coordinates x , y , ω , and h , as well as the width and height. Each variable x , x a , and x * for the predicted box, anchor box, and ground-truth box are determined. A measure of IoU can determine whether two bounding boxes in the same pearl image actually show the same object or a bounding box can be compared with ground-truth annotations to calculate accuracy or precision.

3.5.4. Softmax

The softmax function is used in Faster R-CNN for classification among pearl possible labels. Specifically, in multinomial logistic regression and linear discriminant analysis, the input to the function is the result of K distinct linear functions, and the predicted probability for the j class given a sample vector x and a weighting vector w is as follows:
p ( y ( i ) = j x i ; θ ) = e θ j T x ( i ) l = 1 k e θ T x ( i )
where the composition of K linear functions x x T w 1 ,…, x x T w K and among the softmax function, x T w denotes the inner product of x and w . The operation is equivalent to applying a linear operator defined by w and vectors x , thus transforming the original, probably highly dimensional, input to vectors in a K -dimensional space k .

4. Experimental Results

The network model was run on an Intel Xeon 5218R CPU workstation with an NVIDIA GeForce GTX 2080Ti (12G) GPU, using Ubuntu 20.04.4 LTS 64-bit. The pear image object detection program was developed in TensorFlow V2.8.

4.1. Faster R-CNN with Object Counter Performance

In this section, we collected the pearls dataset to train and test our system, as shown in Table 2. Figure 10 shows the results of applying Faster R-CNN ResNet152 with an object counter on the input images. We can classify the pearls identified by Faster R-CNN in the image and count the number of boxes for each pearl. Figure 10a–c show the results of the model for different counting performance evaluation indicators on the pearl dataset. As shown in Figure 10, the counting accuracy is maintained when the number of pearls in the image increases, which indicates that the Faster R-CNN ResNet152 model with an object counter is suitable for counting densely distributed pearls and its accuracy is basically independent of the number of pearls in the image. Therefore, this model has high detection and counting accuracy.

4.2. The Visualization of Training

Figure 11 shows the classification loss, localization loss and total loss of the training process. The model was trained in 30,000 steps, and when the curve of Figure 11c was examined, the total loss decreased to 0.00044 at the end of 30,000 steps. As the number of epochs increased, the classification loss (Figure 11a) and the localization loss (Figure 11b) of the model gradually decreased to less than 0.00019 and 0.00031, respectively. This shows that the model will have high accuracy in pearl detection.

5. Discussion

The comparison results of the seven models on the integrated test set measured by the total loss curves are shown in Figure 12. A curve at the bottom of the total loss curves indicates a better performance. It can be observed that all the networks provided acceptable performances for pearl detection, and the final experiments show that Faster R-CNN with ResNet152 had good performance in various indicators. Furthermore, we adopted the mAP and speed as quantitative metrics to evaluate the accuracy of the nine models. Although the accuracy of the nine models in evaluating [email protected] was high, ranging from 0.9118 to 0.9883, it was found that the identification [email protected] = 0.9883 was similar to that observed by Yu-Ting Li (2021) [63]. In Table 3, we show how the proposed system performs with different amounts of pearls in images. The performance results of the above cases are trained from 3000 images, with the total amount of pearls being 15,413. Compared with the other eight models, the proposed Faster R-CNN ResNet152 v1 with a counter achieved 98.83% in [email protected], which is the best result. Additionally, the average recall (AR) of Faster R-CNN ResNet152 v1 with a counter is also better than that of the other models (recall/AR@100(small) = 0.8643). It can be observed in Table 3 that the size of Faster R-CNN ResNet152 v1 with the recognition model proposed in the study was only 15 MB, which can effectively realize the detection. The high inference speed as the impact of key factors for deep learning has been widely realized in detection [64,65,66]; however, the loading time of a model is always overlooked. Nevertheless, we cannot always depend on our model for inference in a cache, and we still need a short loading time for predictions. Table 3 shows the loading time of the model measured on a Dell 7920 workstation with Ubuntu and the newest TensorFlow 2.8 and Python 3.10. The first time that we load the model, it takes a long time because the OpenCL kernel for the GPU stack will compile. However, each subsequent inference of the same model will be much faster, and the experimental results are shown in Table 3.
The results shown in Table 3 indicate that these values of different models present no clear tendency. Therefore, the bar charts of the model loading time and model size are also plotted, as shown in Figure 13 and Figure 14. It can be noted that the loading time and side of Faster R-CNN ResNet152 v1 in this study are intermediate levels among the nine detection models. However, as the two most important indicators, the proposed Faster R-CNN ResNet 152 v1 ensures the best detection recall/AR100(small) = 0.8643 (see Figure 15). Meanwhile, the [email protected] = 0.9883 of Faster R-CNN ResNet 152 v1 shows the second-best precision after SSD MobleNetv2, which is superior to the other seven models (see Figure 16). As pearls are valuable, it is necessary to guarantee that there are not any missing in counting. Therefore, to balance the recall and mAP, the Faster R-CNN ResNet152 v1 with a counter is more suitable for pearl auto detection and counting.
To further illustrate the effect of the three classification networks of Faster R-CNN, the overall performances of ResNet50, ResNet101 and ResNet152 were comparatively analyzed. In Figure 17, the red line is the Faster R-CNN ResNet152v1 model, and it can be clearly observed that ResNet152 has a smaller fluctuation when convergence is achieved according to the total loss curves. In conclusion, compared with ResNet50 and ResNet101, Faster R-CNN ResNet152 has better detection and recognition performance for pearls in natural sunlight and artificial light sources, so as to verify that the Faster R-CNN with ResNet152 is optimal.

6. Conclusions and Future Work

An accurate and hygienic PDC method is necessary for pearl production management. Current solutions for pearl counting mainly rely on manual counting, which is generally not accessible to small and medium producers. In this study, a new pearl counting technique is introduced, and the model is composed of object detection and a counting algorithm. Through the comparison of nine object models, encouraging results are obtained. The accuracy of [email protected] of Faster R-CNN ResNet152v1 is 98.83%, which achieves the expected effect. Moreover, the results show that the pearl detection has favorable robustness, which indicates that this method can be extended to the rapid detection of the pearls in natural light and artificial light environments, if appropriate training datasets are available. The proposed deep learning network achieved a recall/AR@100(medium) of 95% in pearl identification, and 100% accuracy in pearl counting. Finally, it is of great significance in the exploration of noncontact and high-throughput object counting. There are mainly two tasks that should be carried out in future work: to evaluate and compare more object detection models and to measure the object sizes and reduce these method limitations.

Author Contributions

Conceptualization, M.H.; Methodology, G.Y.; Software, J.L.; Supervision, X.D.; Writing Original Draft Preparation, R.D.; Validation, X.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the Fund for the National Natural Science Foundation of China (62171143), the Competitive Allocation Project of the Special Fund for Science and Technology in Guangdong Province in 2021 (2021A0528), Guangdong Regional Joint Fund Project (2019B1515120017), Guangdong Marine Economic Development (six major marine industries) Special Project (GDNRC [2021]42), Zhanjiang Key Laboratory of Modern Marine Fishery Equipment (2021A05023), Key R&D Program of Hainan Province (ZDYF2020025), the Scientific Research Start-Up Funds of Guangdong Ocean University (R20037), and the Innovation and Entrepreneurship Team Induced Navigation Plan Project of Zhanjiang (2020LHJH003).

Acknowledgments

The authors would like to thank the reviewers and editors for their time to help improve this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Ren, P.; Wang, L.; Fang, W.; Song, S.; Djahel, S. A novel squeeze YOLO-based real-time people counting approach. Int. J. Bio-Inspired Comput. 2020, 16, 94–101. [Google Scholar] [CrossRef]
  2. Sun, Y.; Li, Z.; He, H.; Guo, L.; Zhang, X.; Xin, Q. Counting trees in a subtropical mega city using the instance segmentation method. Int. J. Appl. Earth Obs. Geoinf. 2022, 106, 102662. [Google Scholar] [CrossRef]
  3. Furuta, M.; Shimizu, K.; Maeta, T.; Miyashita, M.; Izunome, K.; Kubota, H. Noncontact evaluation for interface states by photocarrier counting. Jpn. J. Appl. Phys. 2018, 57, 031301. [Google Scholar] [CrossRef]
  4. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  5. Zhang, Y.; Zhou, C.; Chang, F.; Kot, A.C. Multi-resolution attention convolutional neural network for crowd counting. Neurocomputing 2019, 329, 144–152. [Google Scholar] [CrossRef]
  6. Murthy, C.B.; Hashmi, M.F.; Keskar, A.G. Optimized MobileNet plus SSD: A real-time pedestrian detection on a low-end edge device. Int. J. Multimed. Inf. Retr. 2021, 10, 171–184. [Google Scholar] [CrossRef]
  7. Velumani, K. An automatic method based on daily in situ images and deep learning to date wheat heading stage. Field Crop. Res. 2020, 252, 107793. [Google Scholar] [CrossRef]
  8. Hu, X.; Li, H.; Li, X.; Wang, C. MobileNet-SSD MicroScope using adaptive error correction algorithm: Real-time detection of license plates on mobile devices. IET Intell. Transp. Syst. 2020, 14, 110–118. [Google Scholar] [CrossRef]
  9. Chen, L.; Lin, S.; Lu, X.; Cao, D.; Wu, H.; Guo, C.; Liu, C.; Wang, F.-Y. Deep Neural Network Based Vehicle and Pedestrian Detection for Autonomous Driving: A Survey. IEEE Trans. Intell. Transp. Syst. 2021, 22, 3234–3246. [Google Scholar] [CrossRef]
  10. Han, A.; Zhang, Y.; Liu, Q.; Dong, Q.; Zhao, F.; Shen, X.; Liu, Y.; Yan, S.; Zhou, S. Application of refinements on faster-RCNN in automatic screening of diabetic foot wagner grades. Acta Med. Mediterr. 2020, 36, 661–665. [Google Scholar]
  11. Liu, S.; Ban, H.; Song, Y.; Zhang, M.; Yang, F. Method for Detecting Chinese Texts in Natural Scenes Based on Improved Faster R-CNN. Int. J. Pattern Recognit. Artif. Intell. 2020, 34, 2053002. [Google Scholar] [CrossRef]
  12. Gao, X.; Xu, J.; Luo, C.; Zhou, J.; Huang, P.; Deng, J. Detection of Lower Body for AGV Based on SSD Algorithm with ResNet. Sensors 2022, 22, 2008. [Google Scholar] [CrossRef]
  13. Chen, D.; Liang, M.; Jin, C.; Sun, Y.; Xu, D.; Lin, Y. Coronary Calcium Detection Based on Improved Deep Residual Network in Mimics. J. Med. Syst. 2019, 43, 119. [Google Scholar]
  14. Vy, N.; Duc, N.T. Single-image crowd counting: A comparative survey on deep learning-based approaches. Int. J. Multimed. Inf. Retr. 2020, 9, 63–80. [Google Scholar]
  15. Liu, M.; Wang, X.; Zhou, A.; Fu, X.; Ma, Y.; Piao, C. UAV-YOLO Small Object Detection on Unmanned Aerial Vehicle Perspective. Sensors 2020, 20, 2238. [Google Scholar] [CrossRef] [PubMed]
  16. Abdullah, S.S.; Rajasekaran, M.P. Automatic detection and classification of knee osteoarthritis using deep learning approach. Radiol. Med. 2022, 127, 398–406. [Google Scholar] [CrossRef] [PubMed]
  17. Zhang, J.; Yang, G.; Sun, L.; Zhou, C.; Zhou, X.; Li, Q.; Bi, M.; Guo, J. Shrimp egg counting with fully convolutional regression network and generative adversarial network. Aquac. Eng. 2021, 94, 102175. [Google Scholar] [CrossRef]
  18. Rahmaniar, W.; Wang, W.J.; Chiu, C.-W.E.; Hakim, N.L.L. Real-time bi-directional people counting using an RGB-D camera. Sens. Rev. 2021, 41, 341–349. [Google Scholar] [CrossRef]
  19. Chen, B.; Miao, X. Distribution Line Pole Detection and Counting Based on YOLO Using UAV Inspection Line Video. J. Electr. Eng. Technol. 2019, 77, 125, reprinted in J. Electr. Eng. Technol. 2020, 15, 997–997. [Google Scholar]
  20. Anuar, M.M.; Halin, A.A.; Perumal, T.; Kalantar, B. Aerial Imagery Paddy Seedlings Inspection Using Deep Learning. Remote Sens. 2022, 14, 274. [Google Scholar] [CrossRef]
  21. Zheng, X.; Li, F.; Lin, B.; Xie, D.; Liu, Y.; Jiang, K.; Gong, X.; Jiang, H.; Peng, R.; Duan, X. A Two-Stage Method to Detect the Sex Ratio of Hemp Ducks Based on Object Detection and Classification Networks. Animals 2022, 12, 1177. [Google Scholar] [CrossRef] [PubMed]
  22. Yin, H.; Chen, C.; Hao, C.; Huang, B. A Vision-based inventory method for stacked goods in stereoscopic warehouse. Neural Comput. Appl. 2022, 7, 1–18. [Google Scholar] [CrossRef]
  23. Veeramani, B.; Raymond, J.W.; Chanda, P. DeepSort: Deep convolutional networks for sorting haploid maize seeds. BMC Bioinform. 2018, 19, 289. [Google Scholar] [CrossRef] [PubMed]
  24. Mu, L.; Zhao, H.; Li, Y.; Liu, X.; Qiu, J.; Sun, C. Traffic Flow Statistics Method Based on Deep Learning and Multi-Feature Fusion. Comput. Model. Eng. Sci. 2021, 129, 465–483. [Google Scholar] [CrossRef]
  25. Kirby, E.; Zenha, R.; Jamone, L. Comparing Single Touch to Dynamic Exploratory Procedures for Robotic Tactile Object Recognition. IEEE Robot. Autom. Lett. 2022, 7, 4252–4258. [Google Scholar] [CrossRef]
  26. Kosuge, A.; Suehiro, S.; Hamada, M.; Kuroda, T. mmWave-YOLO: A mmWave Imaging Radar-Based Real-Time Multiclass Object Recognition System for ADAS Applications. IEEE Trans. Instrum. Meas. 2022, 71, 1–10. [Google Scholar] [CrossRef]
  27. Huang, X.; He, P.; Rangarajan, A.; Ranka, S. Machine-Learning-Based Real-Time Multi-Camera Vehicle Tracking and Travel-Time Estimation. J. Imaging 2022, 8, 101. [Google Scholar] [CrossRef]
  28. Premachandra, C.; Tamaki, M. A Hybrid Camera System for High-Resolutionization of Target Objects in Omnidirectional Images. IEEE Sens. J. 2021, 21, 10752–10760. [Google Scholar] [CrossRef]
  29. Mittal, P.; Sharma, A.; Singh, R.; Sangaiah, A.K. On the performance evaluation of object classification models in low altitude aerial data. J. Supercomput. 2022, 78, 1305–1336. [Google Scholar] [CrossRef]
  30. Yang, Z.; Bai, Y.-M.; Sun, L.-D.; Huang, K.-X.; Liu, J.; Ruan, D.; Li, J.-L. SP-ILC: Concurrent Single-Pixel Imaging, Object Location, and Classification by Deep Learning. Photonics 2021, 8, 400. [Google Scholar] [CrossRef]
  31. Wang, H.; Li, D.; Song, Y.; Gao, Q.; Wang, Z.; Liu, C. Single-Shot Object Detection with Split and Combine Blocks. Appl. Sci. 2020, 10, 6382. [Google Scholar] [CrossRef]
  32. Moussa, M.M.; Shoitan, R.; Abdallah, M.S. Efficient common objects localization based on deep hybrid Siamese network. J. Intell. Fuzzy Syst. 2021, 41, 3499–3508. [Google Scholar] [CrossRef]
  33. Zhao, J.; Zhang, X.; Yan, J.; Qiu, X.; Yao, X.; Tian, Y.; Zhu, Y.; Cao, W. A wheat spike detection method in uav images based on improved yolov5. Remote Sens. 2021, 13, 3095. [Google Scholar] [CrossRef]
  34. Jia, W.; Xu, S.; Liang, Z.; Zhao, Y.; Min, H.; Li, S.; Yu, Y. Real-time automatic helmet detection of motorcyclists in urban traffic using improved YOLOv5 detector. IET Image Process. 2021, 2021, 3623–3637. [Google Scholar]
  35. Zhan, J.; Hu, Y.; Cai, W.; Zhou, G.; Li, L. PDAM-STPNNet: A Small Target Detection Approach for Wildland Fire Smoke through Remote Sensing Images. Symmetry 2021, 13, 2260. [Google Scholar] [CrossRef]
  36. Yang, D.; Bi, C.; Mao, L.; Zhang, R. Contour feature fusion SSD Algorithm. In Proceedings of the 38th Chinese Control Conference (CCC), Guangzhou, China, 27–30 July 2019; pp. 3423–3426. [Google Scholar]
  37. Zhou, S.; Qiu, J. Enhanced SSD with interactive multi-scale attention features for object detection. Multimed. Tools Appl. 2021, 80, 11539–11556. [Google Scholar] [CrossRef]
  38. Wu, B.; Iandola, F.; Jin, P.H.; Keutzer, K. SqueezeDet: Unified, Small, Low Power Fully Convolutional Neural Networks for Real-Time Object Detection for Autonomous Driving. In Proceedings of the 30th IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Honolulu, HI, USA, 21–26 July 2016; pp. 446–454. [Google Scholar]
  39. Hu, X. Football Player Posture Detection Method Combining Foreground Detection and Neural Networks. Sci. Program. 2021, 2021, 4102294. [Google Scholar] [CrossRef]
  40. Kuwada, C.; Ariji, Y.; Fukuda, M.; Kise, Y.; Fujita, H.; Katsumata, A.; Ariji, E. Deep learning systems for detecting and classifying the presence of impacted supernumerary teeth in the maxillary incisor region on panoramic radiographs. Oral Surg. Oral Med. Oral Pathol. Oral Radiol. 2020, 130, 464–469. [Google Scholar] [CrossRef]
  41. Liu, Z.; Lyu, Y.; Wang, L.; Han, Z. Detection Approach Based on an Improved Faster RCNN for Brace Sleeve Screws in High-Speed Railways. IEEE Trans. Instrum. Meas. 2020, 69, 4395–4403. [Google Scholar] [CrossRef]
  42. Xie, H.; Chen, Y.; Shin, H. Context-aware pedestrian detection especially for small-sized instances with Deconvolution Integrated Faster RCNN (DIF R-CNN). Appl. Intell. 2019, 49, 1200–1211. [Google Scholar] [CrossRef]
  43. He, K.; Gkioxari, G.; Dollar, P.; Girshick, R. Mask R-CNN. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 386–397. [Google Scholar] [CrossRef] [PubMed]
  44. Wu, M.; Yue, H.; Wang, J.; Huang, Y.; Liu, M.; Jiang, Y.; Ke, C.; Zeng, C. Object detection based on RGC mask R-CNN. IET Image Process. 2020, 14, 1502–1508. [Google Scholar] [CrossRef]
  45. Sun, C.-Y.; Hong, X.-J.; Shi, S.; Shen, Z.-Y.; Zhang, H.-D.; Zhou, L.-X. Cascade Faster R-CNN Detection for Vulnerable Plaques in OCT Images. IEEE Access 2021, 9, 24697–24704. [Google Scholar] [CrossRef]
  46. Zhong, L.; Li, J.; Zhou, F.; Bao, X.; Xing, W.; Han, Z.; Luo, J. Integration between Cascade Region-Based Convolutional Neural Network and Bi-Directional Feature Pyramid Network for Live Object Tracking and Detection. Traitement Signal 2021, 38, 1253–1257. [Google Scholar] [CrossRef]
  47. Ramalingam, B.; Tun, T.; Mohan, R.E.; Gomez, B.F.; Cheng, R.; Balakrishnan, S.; Rayaguru, M.M.; Hayat, A.A. AI Enabled IoRT Framework for Rodent Activity Monitoring in a False Ceiling Environment. Sensors 2021, 21, 5326. [Google Scholar] [CrossRef]
  48. Jiao, H. Intelligent Research Based on Deep Learning Recognition Method in Vehicle-Road Cooperative Information Interaction System. Comput. Intell. Neurosci. 2022, 2022, 4921211. [Google Scholar] [CrossRef]
  49. Huang, Z.; Hu, Q.; Mei, Q.; Yang, C.; Wu, Z. Identity recognition on waterways: A novel ship information tracking method based on multimodal data. J. Navig. 2021, 74, 1336–1352. [Google Scholar] [CrossRef]
  50. Xu, B. Livestock classification and counting in quadcopter aerial images using Mask R-CNN. Int. J. Remote Sens. 2020, 2020, 8121–8142. [Google Scholar] [CrossRef]
  51. Albuquerque, P.L.F.; Garcia, V.; Oliveira, A.D.S., Jr.; Lewandowski, T.; Detweiler, C.; Goncalves, A.B.; Costa, C.S.; Naka, M.H.; Pistori, H. Automatic live fingerlings counting using computer vision. Comput. Electron. Agric. 2019, 167, 105015. [Google Scholar] [CrossRef]
  52. Vasconez, J.P.; Delpiano, J.; Vougioukas, S.; Cheein, F.A. Comparison of convolutional neural networks in fruit detection and counting: A comprehensive evaluation. Comput. Electron. Agric. 2020, 173, 105348. [Google Scholar] [CrossRef]
  53. Chen, C.; Liu, B.; Wan, S.; Qiao, P.; Pei, Q. An Edge Traffic Flow Detection Scheme Based on Deep Learning in an Intelligent Transportation System. IEEE Trans. Intell. Transp. Syst. 2021, 22, 1840–1852. [Google Scholar] [CrossRef]
  54. Jiang, Z.; Shi, B.; Du, F.; Xue, B.; Lei, M.; Yang, Z.; Sun, H. Intelligent Plant Cultivation Robot Based on Key Marker Algorithm Using Visual and Laser Sensors. IEEE Sens. J. 2022, 22, 879–889. [Google Scholar] [CrossRef]
  55. Yu, X.; Wang, Y.; An, D.; Wei, Y. Counting method for cultured fishes based on multi-modules and attention mechanism. Aquac. Eng. 2022, 96, 102215. [Google Scholar] [CrossRef]
  56. Syazwani, R.W.N.; Asraf, M.H.; Amin, M.A.M.S.; Dalil, K.A.N. Automated image identification, detection and fruit counting of top-view pineapple crown using machine learning. Alex. Eng. J. 2022, 61, 1265–1276. [Google Scholar] [CrossRef]
  57. Hansen, M.F.; Oparaeke, A.; Gallagher, R.; Karimi, A.; Tariq, F.; Smith, M.L. Towards Machine Vision for Insect Welfare Monitoring and Behavioural Insights. Front. Vet. Sci. 2022, 9, 835529. [Google Scholar] [CrossRef]
  58. Mahmud, M.S.; Zahid, A.; Das, A.K.; Muzammil, M.; Khan, M.U. A systematic literature review on deep learning applications for precision cattle farming. Comput. Electron. Agric. 2021, 187, 106313. [Google Scholar] [CrossRef]
  59. Mu, Y.; Chen, T.-S.; Ninomiya, S.; Guo, W. Intact Detection of Highly Occluded Immature Tomatoes on Plants Using Deep Learning Techniques. Sensors 2020, 20, 2984. [Google Scholar] [CrossRef]
  60. Kouyoumdjieva, S.T.; Danielis, P.; Karlsson, G. Survey of Non-Image-Based Approaches for Counting People. IEEE Commun. Surv. Tutor. 2020, 22, 1305–1336. [Google Scholar] [CrossRef]
  61. Durve, M.; Bonaccorso, F.; Montessori, A.; Lauricella, M.; Tiribocchi, A.; Succi, S. Tracking droplets in soft granular flows with deep learning techniques. Eur. Phys. J. Plus 2021, 136, 864. [Google Scholar] [CrossRef]
  62. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  63. Li, Y.-T.; Kuo, P.; Guo, J.-I. Automatic Industry PCB Board DIP Process Defect Detection System Based on Deep Ensemble Self-Adaption Method. IEEE Trans. Compon. Packag. Manuf. Technol. 2021, 11, 312–323. [Google Scholar] [CrossRef]
  64. Huang, K.; Li, S.; Deng, W.; Yu, Z.; Ma, L. Structure inference of networked system with the synergy of deep residual network and fully connected layer network. Neural Netw. 2022, 145, 288–299. [Google Scholar] [CrossRef] [PubMed]
  65. Urban, C.J.; Bauer, D.J. A Deep Learning Algorithm for High-Dimensional Exploratory Item Factor Analysis. Psychometrika 2021, 86, 1–29. [Google Scholar] [CrossRef] [PubMed]
  66. Chang, Y.-B.; Tsai, C.; Lin, C.-H.; Chen, P. Real-Time Semantic Segmentation with Dual Encoder and Self-Attention Mechanism for Autonomous Driving. Sensors 2021, 21, 8072. [Google Scholar] [CrossRef] [PubMed]
Figure 1. The finest quality of pearls. (a) Pearl in a mollusk shell, (b) pearl necklace.
Figure 1. The finest quality of pearls. (a) Pearl in a mollusk shell, (b) pearl necklace.
Sensors 22 07026 g001
Figure 2. The application proportions of the three models from 2018 to 2022, including the Faster R-CNN ResNet, SSD MobileNet and SSD ResNet models. (From Web of Science Core Collection).
Figure 2. The application proportions of the three models from 2018 to 2022, including the Faster R-CNN ResNet, SSD MobileNet and SSD ResNet models. (From Web of Science Core Collection).
Sensors 22 07026 g002
Figure 3. The number of object counting papers published in the field of computer vision (from Web of Science).
Figure 3. The number of object counting papers published in the field of computer vision (from Web of Science).
Sensors 22 07026 g003
Figure 4. Illustration of the image capturing process and annotations. (a) Pearl image collections. (b) Annotations by LabelImg.
Figure 4. Illustration of the image capturing process and annotations. (a) Pearl image collections. (b) Annotations by LabelImg.
Sensors 22 07026 g004
Figure 5. The framework of Faster R-CNN with a counter.
Figure 5. The framework of Faster R-CNN with a counter.
Sensors 22 07026 g005
Figure 6. The original Faster R-CNN with RestNet152.
Figure 6. The original Faster R-CNN with RestNet152.
Sensors 22 07026 g006
Figure 7. Object counter architecture of pearls.
Figure 7. Object counter architecture of pearls.
Sensors 22 07026 g007
Figure 8. IoU.
Figure 8. IoU.
Sensors 22 07026 g008
Figure 9. The higher the detection IoU, the better the performance. (a) IoU = 0.396. (b) IoU = 0.886. (c) IoU = 0.9838.
Figure 9. The higher the detection IoU, the better the performance. (a) IoU = 0.396. (b) IoU = 0.886. (c) IoU = 0.9838.
Sensors 22 07026 g009
Figure 10. Partial prediction results of Faster R-CNN ResNet152 with an object counter. (a) The total number of detected pearls is 8. (b) The total number of detected pearls is 16. (c) The total number of detected pearls is 20.
Figure 10. Partial prediction results of Faster R-CNN ResNet152 with an object counter. (a) The total number of detected pearls is 8. (b) The total number of detected pearls is 16. (c) The total number of detected pearls is 20.
Sensors 22 07026 g010aSensors 22 07026 g010b
Figure 11. The loss curve of Faster R-CNN with ResNet152: (a) classification loss; (b) localization loss; (c) total loss.
Figure 11. The loss curve of Faster R-CNN with ResNet152: (a) classification loss; (b) localization loss; (c) total loss.
Sensors 22 07026 g011aSensors 22 07026 g011b
Figure 12. The total loss curves of all algorithms (Faster R-CNN ResNet 152, Faster R-CNN Inception ResNet, SSD MobileNet v1, SSD MobileNet v2, SSD ResNet50, SSD ResNet101 and SSD ResNet152).
Figure 12. The total loss curves of all algorithms (Faster R-CNN ResNet 152, Faster R-CNN Inception ResNet, SSD MobileNet v1, SSD MobileNet v2, SSD ResNet50, SSD ResNet101 and SSD ResNet152).
Sensors 22 07026 g012
Figure 13. Comparison of detection models in loading time.
Figure 13. Comparison of detection models in loading time.
Sensors 22 07026 g013
Figure 14. Comparison of detection models in sizes.
Figure 14. Comparison of detection models in sizes.
Sensors 22 07026 g014
Figure 15. Values of Recall/AR100(small) in different detection models.
Figure 15. Values of Recall/AR100(small) in different detection models.
Sensors 22 07026 g015
Figure 16. Values of [email protected] in different detection models.
Figure 16. Values of [email protected] in different detection models.
Sensors 22 07026 g016
Figure 17. Comparison of total loss curves among Faster R-CNN with ResNet: ResNet50, ResNet101 and ResNet152.
Figure 17. Comparison of total loss curves among Faster R-CNN with ResNet: ResNet50, ResNet101 and ResNet152.
Sensors 22 07026 g017
Table 1. Some of the representative studies in recent years.
Table 1. Some of the representative studies in recent years.
NumAuthorsOrganizationFundamentalsTechnologiesApplicationYear
1Xiaoning Yu [55]China Agricultural UniversityDeep learningMulti-modules and attention mechanismAquaculture2022
2Wan Nurazwin R. [56]Shah Alam SelangorImage recognitionMachine
learning classifiers
Pineapple crown2021
3Mark F. Hansen [57]UWE BristolMachine visionBlack Soldier Fly (BSF)insect farming2022
4Md Sultan Mahmud [58]The Pennsylvania State UniversityDeep learningLSTM + Mask RCNNCattle agriculture2021
5J.P. Vasconz [52]Universidad Técnica Federico Santa MaríaCNNFaster-RCNN with
Inception V2
Fruit counting2020
6Yue Mu [59]Nanjing Agricultural UniversityDeep learningR-CNN with ResNet101Tomato counting2020
7Sylvia T. Kouyoumdjieva [60]KTH Royal Institute of TechnologyNo-image
recognition
RSSI + CSIPeople counting2020
8Mihir Durve [61]Fondazione Istituto Italiano di TecnologiaDeep learningYOLO + DeepSORTTrack moving droplets2021
Table 2. Detailed information of pearl image sets.
Table 2. Detailed information of pearl image sets.
Image SetAmount of ImagesAmount of PearlsImage Conditions
Train image set270013,850Natural light imagesArtificial light images
13501350
Test image set3001563Natural light imagesArtificial light images
150150
Table 3. Metrics comparison of different models.
Table 3. Metrics comparison of different models.
ModelmAP
@0.5IoU
mAP
@0.75IoU
Recall/AR
@100(medium)
Recall/AR
@100(small)
Model Loading TimeSpeed
Inference Time with Counter
Size
Faster R-CNN ResNet152 v1
(ours)
10.98830.950.86435.478 s15.8 ms15 MB
Faster R-CNN ResNet101 v110.96420.93250.85715.625 s14.9 ms11.1 MB
Faster R-CNN ResNet50 v110.95420.93120.854.306 s15.8 ms7.2 MB
Faster R-CNN Inception ResNet v210.91180.9250.72147.869 s16.0 ms18.1 MB
SSD MobileNet v10.99780.98380.940.81434.855 s14.5 ms7.4 MB
SSD MobileNet v210.99420.93750.85715.102 s15.1 ms9.6 MB
SSD ResNet50
v1
10.98300.9250.854.361 s13.4 ms10.2 MB
SSD ResNet101 v10.98410.98410.90.81435.679 s15.2 ms15.8 MB
SSD ResNet 152 v10.98620.98620.88750.82866.471 s15.5 ms21.3 MB
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Hou, M.; Dong, X.; Li, J.; Yu, G.; Deng, R.; Pan, X. PDC: Pearl Detection with a Counter Based on Deep Learning. Sensors 2022, 22, 7026. https://doi.org/10.3390/s22187026

AMA Style

Hou M, Dong X, Li J, Yu G, Deng R, Pan X. PDC: Pearl Detection with a Counter Based on Deep Learning. Sensors. 2022; 22(18):7026. https://doi.org/10.3390/s22187026

Chicago/Turabian Style

Hou, Mingxin, Xuehu Dong, Jun Li, Guoyan Yu, Ruoling Deng, and Xinxiang Pan. 2022. "PDC: Pearl Detection with a Counter Based on Deep Learning" Sensors 22, no. 18: 7026. https://doi.org/10.3390/s22187026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop