Next Article in Journal
Artificial Intelligence-Aided Endoscopy and Colorectal Cancer Screening
Previous Article in Journal
Photodynamic Therapy Used to Treat an HIV Patient with an Efflorescence of Viral Warts after HAART Therapy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv5s-SA: Light-Weighted and Improved YOLOv5s for Sperm Detection

1
School of Electronic Engineering, Beijing University of Posts and Telecommunications, Beijing 100876, China
2
SAS Medical Technology (Beijing) Co., Ltd., Changping District, Beijing 102200, China
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(6), 1100; https://doi.org/10.3390/diagnostics13061100
Submission received: 20 February 2023 / Revised: 7 March 2023 / Accepted: 11 March 2023 / Published: 14 March 2023
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)

Abstract

:
Sperm detection performance is particularly critical for sperm motility tracking. However, there are a large number of non-sperm objects, sperm occlusion and poorly detailed texture features in semen images, which directly affect the accuracy of sperm detection. To solve the problem of false detection and missed detection in sperm detection, a multi-sperm target detection model, Yolov5s-SA, with an SA attention mechanism is proposed based on the YOLOv5s algorithm. Firstly, a depthwise, separable convolution structure is used to replace the partial convolution of the backbone network, which can ensure stable precision and reduce the number of model parameters. Secondly, a new multi-scale feature fusion module is designed to enhance the perception of feature information to supplement the positional information and high-resolution of the deep feature map. Finally, the SA attention mechanism is integrated into the neck network before the output of the feature map to enhance the correlation between the feature map channels and improve the fine-grained feature fusion ability of YOLOv5s. Experimental results show that compared with various YOLO algorithms, the proposed algorithm improves the detection accuracy and speed to a certain extent. Compared with the YOLOv3, YOLOv3-spp, YOLOv5s and YOLOv5m models, the average accuracy increases by 18.1%, 15.2%, 6.9% and 1.9%, respectively. It can effectively reduce the missed detection of occluded sperm and achieve lightweight and efficient multi-sperm target detection.

1. Introduction

Infertility is a common health problem, affecting 15% of couples worldwide [1]. Overall, 30% of infertility cases are related to male infertility [2]. A fundamental challenge in the diagnosis and treatment of infertility lies in the rapid and accurate analysis of semen samples provided by the laboratory. Sperm concentration, morphology [3,4] and DNA fragments are important indicators for semen analysis [5]. In addition, another important indicator affecting male fertility is sperm motility [6,7,8,9], and high-quality detection methods are the key to accurately assessing sperm motility.
Sperm detection is a complex and challenging process. In semen micro videos, sperm cells are usually unstained, colorless and transparent. Color, size and texture information were not effective in sperm detection and are basically ineffective in sperm detection based on color, size and texture information. On the other hand, sperm cells have highly similar morphology, which greatly reduces the feasibility of detection algorithms to extract information from the morphology [10,11,12]. At the same time, semen samples are doped with a lot of epithelial cells or other impurities, which put forward higher requirements for accurate detection of sperm in video frames.
As far as we know, the initial sperm detection methods relied on manual marking by clinical experts using various medical counting devices [13,14,15]. However, manual evaluation is highly subjective, labor-intensive and subject to differences within and among technicians, who lack rigorous and objective data support. To compensate for this, the computer-assisted semen analysis (CASA) [16,17,18] system has been introduced to provide a more accurate, objective and standardized sperm assessment for this field of diagnosis. However, due to the strict parameter definition and many uncertainties of the algorithm used, the methods of sperm detection still vary greatly. New ways are needed to automate and standardize the sperm detection process. In the literature [19], Jati G et al. detected the position of sperm based on the Kalman filter algorithm and inferred the motion trajectory of each sperm by using the Hungarian method. However, the problem of sperm detection in the low contrast region of an image sequence has not been solved. Hasikin et al. [20] proposed a region-based adaptive threshold segmentation technique that used the intensity distribution to classify image pixels, group and separately process pixels, generate multiple thresholds according to the classification and grouping, and realize spermatozoon segmentation in low-contrast areas. When motile sperm collide with impurities with similar contrast, it is easy to lose the sperm target, resulting in a certain percentage of sperm-tracking loss. Alameri et al. [21] adopted an improved Gaussian mixture model to detect sperm and automatically update, including new sperm that re-entered the field of vision. When the number of swimming sperm is large enough, the calculation speed of the algorithm is greatly reduced. Qi et al. [22] detected moving objects through background modeling, and at the same time, realized the classification and counting of sperm with progressive motion, non-progressive motion and non-motility. For detection of small targets, Alabdulla et al. [23] proposed a new method to optimize non-uniform illumination images, and adopted an improved branch-bound algorithm for sperm location trajectory allocation, which had a low error rate in sperm detection and tracking. Somasundaram et al. [24] proposed a FRCNN based on an elliptic scan detection algorithm (ESA) for human sperm motility analysis, improving the accuracy of computer-assisted semen analysis. In recent years, with the wide application of deep learning, experts and researchers at home and abroad continue to deepen and expand the research on small-target detection technology and have made some achievements. Yet, only a handful of algorithms use deep learning to detect sperm cells. Prabaharan et al. [25] used image morphological processing to reduce noise, segmented sperm images by the E-Ostu threshold method, and then used a convolutional neural network to detect abnormal regions. The method achieved 98.99% accuracy and was able to effectively detect abnormal sperm morphology. Abbasi et al. [26] applied a novel combination of deep transfer learning and multitasking learning to sperm detection. The literature [27] uses classical CNN to detect and segment sperm cells. Compared with traditional algorithms, one of the advantages of this work is that it can process stained semen images. In 2022, Qixian et al. [28] established a new dataset consisting of 1207 sperm cell images from more than 20 male infertility patients. They proposed an efficient deep learning algorithm integrating extended convolution into the U-Net network structure for automatic segmentation of the human sperm head. However, the introduction of extended convolution is prone to a grid effect, which will lose the continuity of feature information. The object detection algorithm based on deep learning uses a large number of datasets to train the model, which has a stronger generalization ability and is more suitable for complex and changeable scenes.
In this paper, the proposed method is based on the deep convolution neural network of YOLOv5s [29] architecture for sperm detection. The main contributions of this study are as follows:
  • We propose a novel YOLOv5s algorithm and introduce Shuffle Attention (SA) mechanism [30] to enhance the model’s attention to small target sperm.
  • We propose to replace partial convolution in the backbone network with depthwise separable convolution (DWConv) [31] to improve the convergence speed of the model.
  • The proposed method can effectively process semen images with low quality and a complex background, improve the detection performance of sperm and provide the feasibility of accurate detection using human or other animal sperm cells.
The rest of the paper is organized as follows. Section 2 describes the creation of the sperm dataset and the detailed structure of the YOLOv5s-SA detection model. Experimental comparisons of the state-of-the-art detection methods are performed in Section 3. Finally, Section 4 states our conclusions and discusses future work.

2. Materials and Methods

2.1. Date Set

The quality of the dataset is critical to the outcome of deep learning. Currently, the only open-source dataset for sperm motility prediction is VISEM [32], but this dataset does not provide detailed data consistent with deep learning detection. Therefore, we built a dataset. In the R&D department of SAS Medical Technology (Beijing) Co., LTD. (Beijing, China), there were microscopic videos of sperm from 10 different men with unstained samples. To capture sperm movement, we used a phase contrast microscope with a total magnification of 100 times. Videos range in length from 10 s to 130 s, are shot at 30 fps and have a resolution of 1280 by 960 pixels. The average number of sperm per frame is 274.6.
In this study, datasets were constructed from video samples of 10 different males’ sperm, and the first 100 frames were captured from the first 6 male semen video samples, 80% of which were used for the training dataset and the rest for the validation dataset. The training set had 480 images and 133,921 sperms. The validation set consisted of 32,768 sperms. In addition, 30 video frames were randomly captured from the remaining 4 videos and used as a test dataset. The test set had 120 pictures and contained 33,194 sperm cells. Table 1 illustrates the dataset proportions. The training set and the test set are strictly independent. Although there are only 10 videos in the data, there are a significant number of annotated objects.

2.2. YOLOv5s-SA Architecture

The structure of the YOLOv5s-SA target detection network is shown in Figure 1, which consists of the backbone feature extraction network, neck network and head target prediction network. (1) The backbone network is mainly composed of the focus, CSP module and spatial pyramid pooling (SPP) [33] module. In order to reduce the number of parameters in the model, partial convolution in the backbone network is replaced by DWConv [31]. (2) The feature pyramid network (FPN) [34] and path aggregation network (PAN) [35] are used in the neck to realize multi-scale feature fusion. In order to further enhance deep and shallow semantic feature fusion and suppress redundant background noise, we introduced an SA attention mechanism [30] to enhance the model’s attention to small target sperm. (3) The head prediction network integrates features of the acquired enhanced feature maps and introduces EIoU instead of CIoU [36] to calculate the bounding box regression losses and obtain the prediction results.

2.2.1. Lightweight Backbone Feature Extraction Module

Skip connections between different layers of the original YOLOv5 backbone network can alleviate the gradient disappearance caused by the deepening of the network. However, due to the excessive number of CSP structures, the texture details are easily lost. We designed a lightweight backbone network to reduce the weight parameters and computation amount of the model. DWConv of MobileNet [31] replaces the partial convolution of the backbone network, which can guarantee the sperm detection accuracy is not affected while the number of parameters and calculation of the model are not increased. The specific structure of DWConv is shown in Figure 2, which is a combination of depthwise convolution and pointwise convolution, with a low parameter quantity and calculation cost. In addition, the selection of the activation function in the convolution operation has a certain impact on the detection performance inspired by the literature [37]. We replace the LeakyReLU [38] activation function in the convolution layer with SiLU [39] with good generalization ability, whose expression is defined as follows:
f ( x ) = x s i g m o i d ( β x )
where β is a constant or trainable parameter. SiLU has the characteristics of non-monotonic, lower limit, no upper limit and smooth, resulting in a stronger regularization effect. The specific structure of the lightweight backbone is shown in Figure 1.

2.2.2. Multi-Scale Feature Fusion Enhancement Module

The original YOLOv5s network extracted features from the shallow layer to the deep layer. Shallow layer features have higher resolution and rich geometric information, but lack detailed feature information and have a small receptive field. On the contrary, the deep feature layer has high semantic information and a strong receptive field due to multiple convolution, but at the cost of losing part of the resolution and location information. After the analysis of the self-built dataset, there is only one target category to be detected in the sample, and each sperm to be identified in the sample is basically a small target. The deeper features are fused with the shallower features. The dimensions of the feature maps’ output from different network layers are unified by an up-sampling or down-sampling operation. The concat operation is used to fuse multi-scale feature information. The network model has a high resolution, strong receptive field and rich semantic information.
Let the input channels of the branches at both ends of the concat operation be X 1 , X 2 , , X c and Y 1 , Y 2 , , Y c . Then, the output of concat is shown in Formula (2).
Z c o n c a t = i = 1 c X i K i + i = 1 c Y i K i + c
where denotes the convolution operation. As can be seen from the above formula, the concat layer only realizes the concatenation of two or more feature maps on the channel dimensions. That is, the features of the image itself are increased, but no attention is paid to the relationships between each feature in the channel. Cross-scale feature fusion easy causes an aliasing effect, resulting in semantic information attenuation. Inspired by the SA module [30], the spatial attention and channel attention mechanisms can be efficiently combined in feature fusion networks. After inserting the SA module into the three output channels of the neck network, we introduced group convolution to reduce the amount of computation and assigned corresponding weights to each group to enhance the perception of details.
Feature fusion algorithms based on channel attention and spatial attention are applied to the neck network. The optimized multi-scale feature fusion network is shown in Figure 3. After extracting image features from the original image through the backbone network, three feature maps with different depths are obtained: H2, H3 and H4. The feature map is then input to the neck network. First, the FPN structure conveys strong semantic features from the top-down, incorporating deep feature information into the shallow layer. After concat and the up-sampling of the deep feature map H4 and middle layer feature H3, N3 is generated by the convolution operation. Feature map N3 is then fused with shallow feature map H2 by up-sampling to generate the enhanced feature map N2. Because FPN is limited by unidirectional feature fusion, shallow feature information is not fully utilized. The PAN structure further carries out feature fusion on the feature layer and integrates the shallow feature information into the deep layer. N2 is down-sampled and fused with feature layer N3 to obtain the enhanced feature figure P3. P3 is down-sampled and fused with feature layer N4 to obtain the enhanced feature figure P4. Finally, feature maps P2, P3 and P4, combined with the efficient SA modules, enable the neck network to accurately learn the correlation of feature maps between channels, alleviate background information and noise interference from non-sperm cells and enhance the feature extraction effect of target sperm.

2.3. Parameter Setting and Experimental Environment

In addition to the neural network architecture, network parameters are also crucial for training speed and accuracy. We set the size of video frames sent into network training as 608 × 608 pixels, batch_size as 8, epoch as 1000 and initial learning rate as 0.01. We set the momentum factor to 0.9 and the weight attenuation parameter to 0.0005. Mosaic data enhancement is adopted to increase the diversity of the training samples and improve the generalization ability of the network.
The training and testing environment of the YOLOv5s-SA algorithm is as follows. The GPU is NVIDIA GeForce GTX 1080ti, and the processor is Intel(R) Core(TM) i7-6700. The operating system is Windows 10, and pytorch 1.9.1 + CUDA 10.2 is selected as the deep learning framework.

3. Experiments

3.1. Evaluation Metrics

In order to evaluate the performance of the algorithm, the precision ( P ), recall ( R ), average precision ( A P ) and F 1 score ( F 1 ) are used as four evaluation indexes of the model detection accuracy. Their definitions are as follows:
P = T P F P + T P
R = T P F N + T P
A P = 0 1 P ( R ) d R
F 1 = 2 × P × R P + R
where T P represents the number of sperm detected and correctly predicted, T N represents the number of non-sperm detected, F P represents the number of sperm detected but incorrectly predicted, F N represents the number of sperm missed, and A P represents the area of a certain category under the P R curve. In addition, frames per second (fps) was used to measure the detection speed of the model. The larger the fps, the faster the detection speed of the model.

3.2. YOLOv5s Ablation Study

We designed an ablation experiment to verify the effectiveness of the DWConv and SA modules. Each group of experiments used the same parameters and was trained in the same network environment. Table 2 shows the influence of different optimization methods on the detection performance of YOLOv5s on the test dataset. It can be seen from the experimental results that only DWConv is used to replace part of the convolution structure of the YOLOv5s backbone network, and the sperm detection performance is not affected, while the number of network parameters is reduced. The detection speed also increases with the reduction of parameters, from 52.5 fps to 55.8 fps, increasing the recall rates by 1.8 percentage points. With the addition of only the SA attention module, although the optimization method increases the number of parameters and the calculation amount, the accuracy of the sperm detection is increased by 2.1 percentage points, and the recall rate is increased by 2.2 percentage points. The two improvements alone are not significant. However, when the two improvements are added to the original model, the sperm detection accuracy and recall rate are 5.7% and 8.6% higher, respectively, than that of the original YOLOv5s model, and the AP is increased by 6.9% from 80.9% to 87.5%. This shows that the improvement of the two methods has greatly improved the accuracy of the sperm detection.
Figure 4 shows the comparison of the sperm detection results before and after the network model optimization. It can be seen from Figure 4a that the original YOLOv5s model is prone to missed detection and false detection. Figure 4b shows that the optimized YOLOv5s-SA model significantly improves the defects of the original YOLOv5s sperm detection, and the introduction of the SA attention mechanism can make the model more fine-grained to express multi-scale features, pay more attention to small targets and improve the accuracy of the network sperm detection.

3.3. Evaluation of Sperm Detection Methods

The YOLOv5s-SA algorithm is compared with traditional image processing, YOLOv3, YOLOv3-spp, YOLOv5s and YOLOv5m algorithms to objectively evaluate the superiority of YOLOv5s-SA. The same training parameters and datasets are used for training, and the same test set is used for comparison experiments after training. The comparison of evaluation indicators of the different algorithms is shown in Table 3. It can be seen that the detection performance of the traditional image processing methods is not as good as that based on deep learning. YOLOv3 achieved an 83.0 AP on the validation dataset, but the accuracy rate decreased on the test dataset. The YOLOv3 algorithm has only an AP of 69.4% on the test dataset, which is not enough to detect sperm targets effectively, and has poor robustness for the detection of densely distributed sperm. Compared with the YOLOv3-spp algorithm, the YOLOv5s and YOLOv5m algorithms increased by 2.5% and 4.3%, respectively, on the validation dataset. The average precision of the proposed method on the validation dataset is 93.1%, and the average precision on the test dataset is 87.5%, which is 1.9% and 1.9% higher, respectively, than YOLOv5m. In general, the YOLOv5 algorithm can basically recognize the majority of sperm, but it will seriously miss detection in dense and occluded conditions. The detection performance of the YOLOv5s-SA algorithm is improved compared with the YOLOv5 algorithm, and it can maintain a high average precision when the number of parameters is low. In addition, for the case of dense distribution, the number of missed detections is significantly reduced, and the detection effect of the proposed algorithm is the best, and all sperm are basically completely identified.
Speed is a highly critical indicator of sperm target detection. The traditional digital image processing method has reached 1.59 fps, though it only uses the CPU. However, it provides low detection accuracy. Obviously, the deep learning method improves the detection speed (dozens of times as much as the traditional image processing method). In addition, the accuracy of the application of the deep learning method has also been greatly improved. We used a feature extraction network with a more lightweight structure to further improve the speed of sperm detection. Although the SA attention module was introduced in the multi-scale feature fusion process, the detection speed did not decrease significantly. Compared with traditional image processing methods and YOLO series algorithms, the YOLOv5s-SA algorithm can obtain a higher detection accuracy with fewer parameters, and the speed is about three times faster than YOLOv3.

3.4. Partial Occlusion Handling

The small size and fast movement of sperm make it easy to obstruct each other and to create obstructions between sperm moving forward and between sperm and impurities. However, partial occlusion is one of the major challenges resulting from the limited accuracy of sperm detection. Compared with other methods, the improved YOLOv5s-SA model proposed performs well in dealing with partial occlusions. Figure 5 shows a frame of video in the test dataset, and Table 4 shows the comparison details.
In the first state, YOLOv3-spp and YOLOv5s failed to detect partially occluded sperm cells. Our improved YOLOv5s-SA and traditional image processing methods were able to completely detect partially occluded sperm. However, traditional image processing methods failed to detect some obvious sperm correctly. In the second state, YOLOv5s can detect partially occluded sperm. At the same time, it detected some sperm-like impurities as sperm cells, leading to a false detection. Other methods still misjudge or fail to detect. It can be seen that YOLOv5s-SA can accurately detect sperm and has a better effect in handling non-sperm objects, although in a complex background.

3.5. Failure Case Analysis

In short, the YOLOv5s-SA model we proposed can detect sperm better than other methods. However, we still encountered some error detection issues. Table 5 shows some detection failures highlighted with arrows.
Taken together, the reasons for these failures are: (1) A large number of fragments in semen are more or less similar to sperm cells (Case 1). (2) The sperm head or neck is attached to the edge of the video frame, making it look like an artifact (Case 2). (3) Sperm is almost completely occluded (case 3).

4. Conclusions

As the YOLO algorithm still has insufficient representation ability and low accuracy when applied to small-target detection tasks, we adopted the improved YOLOv5s algorithm for small-target detection of sperm, which is robust for partially blocked sperm, a large number of non-sperm objects and low contrast. DWConv is used to replace partial convolution in the backbone network, which reduces network redundant computation and improves the convergence speed of the model. By integrating the SA attention mechanism into the neck network, the weight relationship between the channels of different scale feature maps was optimized, which alleviated the semantic information attenuation in the process of multi-scale feature fusion and enhanced the multi-scale sperm detection performance. The experimental results show that the proposed YOLOv5s-SA model has significantly improved the detection accuracy of sperm targets, and the detection speed is better than other mainstream methods on the basis of maintaining the lightweight calculation cost. In the future, we hope to extend the proposed method to detect sperm cells in other animals, such as pigs, cows and fish. We believe that the YOLOv5s-SA algorithm has strong generalization ability and robustness, and should also perform well in similar situations of small-size target detection.

Author Contributions

Conceptualization, J.H.; Methodology, R.Z.; Software, E.H.; Validation, J.Z.; Formal analysis, J.H.; Investigation, Y.C. and H.L.; Resources, E.H.; Data curation, Z.Z.; Writing—original draft, R.Z.; Writing—review & editing, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Key R&D Program of China, grant number 2017YFB0503700.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zauner, G.; Girardi, G. Potential causes of male and female infertility in Qatar. J. Reprod. Immunol. 2020, 141, 103173. [Google Scholar] [CrossRef] [PubMed]
  2. Selvam, M.K.P.; Ambar, R.F.; Agarwal, A.; Henkel, R. Etiologies of sperm DNA damage and its impact on male infertility. Andrologia 2021, 53, 1. [Google Scholar]
  3. Matson, P.; Kitson, M.; Zuvela, E. Human sperm morphology assessment since 2010: Experience of an Australian external quality assurance programme. Reprod. Biomed. Online 2022, 44, 340–348. [Google Scholar] [CrossRef] [PubMed]
  4. Miahi, E.; Mirroshandel, S.A.; Nasr, A. Genetic Neural Architecture Search for automatic assessment of human sperm images. Expert Syst. Appl. 2022, 188, 115937. [Google Scholar] [CrossRef]
  5. World Health Organization. WHO Laboratory Manual for the Examination and Processing of Human Semen, 5th ed.; World Health Organization: Geneva, Switzerland, 2010; pp. 280–286. [Google Scholar]
  6. Valiuskaite, V.; Raudonis, V.; Maskeliunas, R.; Damasevicius, R.; Krilavicious, T. Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination. Sensors 2020, 21, 72. [Google Scholar] [CrossRef]
  7. Ilhan, H.O.; Yuzkat, M.; Aydin, N. Sperm Motility Analysis by using Recursive Kalman Filters with the smartphone based data acquisition and reporting approach. Export Syst. Appl. 2021, 186, 115774. [Google Scholar] [CrossRef]
  8. Garcia-Grau, E.; Lleberia, J.; Costa, L.; Guitart, M.; Yeste, M.; Benet, J.; Amengual, M.J.; Ribas-Maynou, J. Decline of Sperm Quality over the Last Two Decades in the South of Europe: A Retrospective Study in Infertile Patients. Biology 2023, 12, 70. [Google Scholar] [CrossRef]
  9. Ozyer, G.T.; Ozyer, B.; Negin, F.; Alarabi, I.; Agahian, S. A hybrid IMM-JPDAF algorithm for tracking multiple sperm targets and motility analysis. Neural Comput. Appl. 2022, 34, 17407–17421. [Google Scholar] [CrossRef]
  10. Liu, G.L.; Shi, H.; Zhang, H.; Zhou, Y.T.; Sun, Y.J.; Li, W.; Huang, X.F.; Jiang, Y.Q.; Fang, Y.L.; Yang, G. Fast Noninvasive Morphometric Characterization of Free Human Sperms Using Deep Learning. Microsc. Microanal. 2022, 28, 1767–1779. [Google Scholar] [CrossRef]
  11. Ilhan, H.O.; Serbes, G. Sperm morphology analysis by using the fusion of two-stage fine-tuned deep networks. Biomed. Signal Process. Control 2022, 7, 103246. [Google Scholar] [CrossRef]
  12. Spencer, L.; Fernando, J.; Akbaridoust, F.; Ackermann, K.; Nosrati, R. Ensembled Deep Learning for the Classification of Human Sperm Head Mophology. Adv. Intell. Syst. 2022, 4, 2200111. [Google Scholar] [CrossRef]
  13. Mahmoud, A.M.; Depoorter, B.; Piens, N.; Comhaire, F.H. The performance of 10 different methods for the estimation of sperm concentration. Fertil. Steril. 1997, 68, 345. [Google Scholar] [CrossRef] [PubMed]
  14. Davis, R.O.; Rothmann, S.A.; Overstreet, J.W. Accuracy and precision of computer-aided sperm analysis in multicenter studies. Fertil. Steril. 1992, 57, 648–653. [Google Scholar] [CrossRef]
  15. Talarczyk-Desole, J.; Berger, A.; Taszarek, G.; Hauke, J.; Pawelczyk, L.; Jedrzejczak, P. Manual vs. computer-assisted sperm analysis: Can CASA replace manual assessment of human semen in clinical practice? Ginekol. Pol. 2017, 88, 56–60. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Larsen, L.; Scheike, T.; Jensen, T.K.; Bonde, J.P.; Ernst, E.; Hjollund, N.H.; Zhou, Y.; Skakkebaek, N.E.; Giwercman, A. Computer-assisted semen analysis parameters as predictors for fertility of men from the general population. Human Reprod. 2000, 15, 1562–1567. [Google Scholar] [CrossRef] [Green Version]
  17. Tomlinson, M.J.; Pooley, K.; Simpson, T.; Newton, T.; Hopkisson, J.; Jayaprakasan, K.; Jayaprakasan, R.; Naeem, A.; Pridmore, T. Validation of a novel computer-assisted sperm analysis (CASA) system using multitarget-tracking algorithms. Fertil. Steril. 2010, 93, 1911–1920. [Google Scholar] [CrossRef]
  18. Hicks, S.A.; Andersen, J.M.; Witczak, O.; Thambawita, V.; Halvorsen, P.; Hammer, H.L.; Haugen, T.B.; Riegler, M.A. Machine learning-based analysis of sperm videos and participant data for male fertility prediction. Sci. Rep. 2019, 9, 16770. [Google Scholar] [CrossRef] [Green Version]
  19. Jati, G.; Gunawan, A.A.; Lestari, S.W. Multi-sperm tracking using Hungarian kalman filter on low frame rate video. In Proceedings of the Advanced Computer Science and Information Systems (ICACSIS), Malang, Indonesia, 15–16 October 2016; pp. 530–535. [Google Scholar]
  20. Hasikin, K.; Ii, N.A.M.; Mohamed, M. A new region-based adaptive thresholding for sperm motility segmentation. Malays. J. Comput. Sci. 2016, 29, 272–286. [Google Scholar] [CrossRef]
  21. Alameri, M.; Hasikin, K.; Kadri, N.A. Multistage Optimization Using a Modified Gaussian Mixture Model in Sperm Motility Tracking. Comput. Math. Methods Med. 2021, 8, 14. [Google Scholar] [CrossRef]
  22. Qi, S.; Nie, T.; Li, Q.; He, Z.; Xu, D.; Chen, Q. A Sperm Cell Tracking Recognition and Classification Method. In Proceedings of the 2019 International Conference on Systems, Signals and Image Processing (IWSSIP), Osijek, Croatia, 5–7 June 2019; pp. 163–167. [Google Scholar] [CrossRef]
  23. Alabdulla, A.A.A.; Haşıloğlu, A.; Aksu, E.H. A robust sperm cell tracking algorithm using uneven lighting image fixing and improved branch and bound algorithm. IET Image Process. 2021, 15, 2068–2079. [Google Scholar] [CrossRef]
  24. Somasundaram, D.; Nirmala, M. Faster region convolutional neural network and semen tracking algorithm for sperm analysis. Comput. Methods Programs Biomed. 2021, 200, 105918. [Google Scholar] [CrossRef] [PubMed]
  25. Prabaharan, L.; Raghunathan, A. An improved convolutional neural network for abnormality detection and segmentation from human sperm images. J. Ambient Intell. Humaniz. Comput. 2021, 12, 3341–3352. [Google Scholar] [CrossRef]
  26. Abbasi, A.; Miahi, E.; Mirroshandel, S.A. Effect of Deep Transfer and Multi task Learning on Sperm Abnormality Detection. Comput. Biol. Med. 2021, 128, 104121. [Google Scholar] [CrossRef]
  27. Nissen, S.M.; Krause, O.; Almstrup, K.; Kjærulff, S.; Nielsen, T.T.; Nielsen, M. Convolutional neural networks for segmentation and object detection of human semen. In Proceedings of the 20th Scandinavian Conference on Image Analysis (SCIA), Tromsø, Norway, 12–14 June 2017; pp. 397–406. [Google Scholar]
  28. Qixian, L.V.; Yuan, X.R.; Qian, J.Z.; Li, X.K.; Zhang, H.Y.; Zhan, S. An Improved U-Net for Human Sperm Head Segmentation. Neural Process. Lett. 2022, 54, 537–557. [Google Scholar]
  29. Qiao, G.C.; Yang, M.X.; Wang, H. A water level measurement approach based on YOlOv5s. Sensors 2022, 10, 3714. [Google Scholar] [CrossRef]
  30. Zhang, Q.L.; Yang, Y.B. SA-Net: Shuffle attention for deep convolutional neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 2235–2239. [Google Scholar]
  31. Howard, A.; Sandler, M.; Chen, B. Searching for MobileNet V3. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–28 October 2019; pp. 1314–1324. [Google Scholar]
  32. Haugen, T.B.; Hicks, S.A.; Andersen, J.M.; Witczak, O.; Hammer, H.L.; Borgli, R.; Halvorsen, P.; Riegler, M.A. Visem: A multimodal video dataset of human spermatozoa. In Proceedings of the 10th ACM on Multimedia Systems Conference, ACM, Amherst, MA, USA, 18–21 June 2019. [Google Scholar]
  33. Huang, Z.C.; Wang, J.L. DC- SPP- YOLO: Dense connection and spatial pyramid pooling based YOLO for object detection. Inf. Sci. 2020, 522, 241–258. [Google Scholar] [CrossRef] [Green Version]
  34. Lin, T.Y.; Dollar, P.; Girshick, R. Feature pyramid networks for object detection. In Proceedings of the Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  35. Liu, S.; Qi, L.; Qin, H.F.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 21–23 June 2018; pp. 8759–8768. [Google Scholar]
  36. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU loss: Faster and better learning for bounding box regression. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 12993–13000. [Google Scholar]
  37. Murtaza, M.; Sharif, M.; Yasmin, M.; Fayyaz, M.; Kadry, S.; Lee, M.Y. Clothes Retrieval Using M-AlexNet with Mish Function and Feature Selection Using Joint Shannon’s Entropy Pearson’s Correlation Coefficient. IEEE Access 2022, 2, 115469–115490. [Google Scholar] [CrossRef]
  38. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics, JMLR Workshop and Conference Proceedings, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  39. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoidweighted linear units for neural network function approximation in reinforcement learning. Neural Netw. 2018, 107, 3–11. [Google Scholar] [CrossRef]
Figure 1. Structure of YOLOv5s-SA model.
Figure 1. Structure of YOLOv5s-SA model.
Diagnostics 13 01100 g001
Figure 2. DWConv model.
Figure 2. DWConv model.
Diagnostics 13 01100 g002
Figure 3. The optimized multi-scale fusion structure.
Figure 3. The optimized multi-scale fusion structure.
Diagnostics 13 01100 g003
Figure 4. Comparison of model improvement before and after. (a) YOLOv5s sperm detection; (b) YOLOv5s-SA sperm detection.
Figure 4. Comparison of model improvement before and after. (a) YOLOv5s sperm detection; (b) YOLOv5s-SA sperm detection.
Diagnostics 13 01100 g004
Figure 5. The partial occlusion analysis. (Numbers 1 and 2 represent cases 1 and 2 in Table 4, respectively.)
Figure 5. The partial occlusion analysis. (Numbers 1 and 2 represent cases 1 and 2 in Table 4, respectively.)
Diagnostics 13 01100 g005
Table 1. Dataset proportions.
Table 1. Dataset proportions.
SamplesVideo1∼Video6Video7Video8Video9Video10
Number of framesTrainingValidationTesting
48012030303030
Table 2. Ablation experiments. (● represents using the optimization method, ○ represents not using the optimization method).
Table 2. Ablation experiments. (● represents using the optimization method, ○ represents not using the optimization method).
ModelDWConvSAP
(%)
R
(%)
AP
(%)
Params
(M)
FPS
YOLOv5s77.479.680.67.1952.5
YOLOv5s + DWConv76.381.180.67.0655.8
YOLOv5s + SA76.881.581.37.2753.0
YOLOv5s + DWConv + SA83.188.287.57.1457.3
Table 3. Comparative experimental results.
Table 3. Comparative experimental results.
ModelValidation Set
AP (%)
Test SetParams (M)FPS
P (%)R (%)AP (%)F1
Digital image processing79.255.755.960.155.8n/a1.59
YOLOv383.067.366.069.466.661.8519.6
YOLOv3-spp86.975.071.472.373.263.5317.2
YOLOv5s89.477.479.680.678.57.1952.5
YOLOv5m91.282.285.485.683.821.0746.1
YOLOv5s-SA93.183.188.287.585.67.1457.3
Table 4. Partial occlusion handling comparison.
Table 4. Partial occlusion handling comparison.
CaseConventional Image ProcessingYOLOv3-sppYOLOv5sYOLOv5s-SA
1Diagnostics 13 01100 i001Diagnostics 13 01100 i002Diagnostics 13 01100 i003Diagnostics 13 01100 i004
2Diagnostics 13 01100 i005Diagnostics 13 01100 i006Diagnostics 13 01100 i007Diagnostics 13 01100 i008
Table 5. Some detection failures.
Table 5. Some detection failures.
CaseOriginal FrameDetection Result
1. Failure example from the validation dataset (false positive)Diagnostics 13 01100 i009Diagnostics 13 01100 i010
2. Failure example from the test dataset (false positive)Diagnostics 13 01100 i011Diagnostics 13 01100 i012
3. Failure example from the test dataset (false negative)Diagnostics 13 01100 i013Diagnostics 13 01100 i014
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhu, R.; Cui, Y.; Huang, J.; Hou, E.; Zhao, J.; Zhou, Z.; Li, H. YOLOv5s-SA: Light-Weighted and Improved YOLOv5s for Sperm Detection. Diagnostics 2023, 13, 1100. https://doi.org/10.3390/diagnostics13061100

AMA Style

Zhu R, Cui Y, Huang J, Hou E, Zhao J, Zhou Z, Li H. YOLOv5s-SA: Light-Weighted and Improved YOLOv5s for Sperm Detection. Diagnostics. 2023; 13(6):1100. https://doi.org/10.3390/diagnostics13061100

Chicago/Turabian Style

Zhu, Ronghua, Yansong Cui, Jianming Huang, Enyu Hou, Jiayu Zhao, Zhilin Zhou, and Hao Li. 2023. "YOLOv5s-SA: Light-Weighted and Improved YOLOv5s for Sperm Detection" Diagnostics 13, no. 6: 1100. https://doi.org/10.3390/diagnostics13061100

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop