Next Article in Journal
Binaural Heterophasic Superdirective Beamforming
Previous Article in Journal
Twist-n-Sync: Software Clock Synchronization with Microseconds Accuracy Using MEMS-Gyroscopes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination

by
Viktorija Valiuškaitė
1,
Vidas Raudonis
1,
Rytis Maskeliūnas
2,
Robertas Damaševičius
3,4,* and
Tomas Krilavičius
3
1
Department of Control Systems, Kaunas University of Technology, 51423 Kaunas, Lithuania
2
Department of Multimedia Engineering, Kaunas University of Technology, 51423 Kaunas, Lithuania
3
Department of Applied Informatics, Vytautas Magnus University, 44404 Kaunas, Lithuania
4
Faculty of Applied Mathematics, Silesian University of Technology, 444-100 Gliwice, Poland
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(1), 72; https://doi.org/10.3390/s21010072
Submission received: 6 October 2020 / Revised: 15 December 2020 / Accepted: 21 December 2020 / Published: 24 December 2020
(This article belongs to the Section Intelligent Sensors)

Abstract

:
We propose a deep learning method based on the Region Based Convolutional Neural Networks (R-CNN) architecture for the evaluation of sperm head motility in human semen videos. The neural network performs the segmentation of sperm heads, while the proposed central coordinate tracking algorithm allows us to calculate the movement speed of sperm heads. We have achieved 91.77% (95% CI, 91.11–92.43%) accuracy of sperm head detection on the VISEM (A Multimodal Video Dataset of Human Spermatozoa) sperm sample video dataset. The mean absolute error (MAE) of sperm head vitality prediction was 2.92 (95% CI, 2.46–3.37), while the Pearson correlation between actual and predicted sperm head vitality was 0.969. The results of the experiments presented below will show the applicability of the proposed method to be used in automated artificial insemination workflow.

1. Introduction

Infertility is recognized clinically when a man and woman are unable to conceive after one year of regular unprotected sex, or when a woman has had two or more failed pregnancies. Around 15–20% of couples face this problem [1]. One of the solutions is artificial insemination. It is a procedure in which a woman′s egg is fertilized by male semen in a test tube rather than in the womb. This method has been used for several decades (the first baby was born after the artificial insemination procedure in 1978 [2]), but the probability of fertilization is still very low; only 30–50% of tests give a positive result. Infertility has previously been considered a woman’s problem, but research has shown that male infertility accounts for up to 47% of cases in infertile couples [3]. Sometimes infertility can be cured by changing one′s habits (e.g., giving up tobacco, alcohol and heavy food) or by removing stress-causing factors. Such changes may favor several crucial quality indicators: sperm motility (the percentage of motile and partially motile sperm should exceed 32% in a healthy sample), volume (normal ejaculation of sperm should be 1.5 mL), viability (a desired value of at least 58%), or concentration (typically about 39 million sperm are found in the sample) [4]. Threshold values for sperm concentration, morphology, and motility can be employed to identify men as subfertile or fertile [5]. Good motility is characterized by the ability of a sperm head to move quickly in one direction, while poor motility is observed when a sperm head moves slowly and in an unorderly fashion (Figure 1). Although infertility can be treated with hormonal medications, such treatment is only effective for a very small proportion of patients, and the treatment itself lasts up to three months.
In cases where neither lifestyle changes nor medication help, artificial insemination is used, i.e., the fusion of a woman′s eggs and a man′s sperm outside the woman′s body, which is implemented using assisted reproductive technology (ART). Possible methods of fertilization are: in vitro fertilization (IVF) [6], where high-quality sperm is poured on the ovum, and intracytoplasmic sperm injection (ICSI), where morphologically healthy-looking sperm is selected with the help of a syringe and injected into the ovum. The latter procedure requires a particularly powerful microscope that would magnify the image up to 8000 times. Not all fertility clinics can afford it, so this method of fertilization is used only in exceptional cases. Although the incidence of intracytoplasmic sperm injection and the percentage of women who give birth are as high as 60%, this procedure is more invasive than IVF, as there is a chance (2%) that the egg will be damaged during sperm injection, making the egg non-viable. In addition, preparing for the ICSI procedure requires more diligence; embryologists, looking through a microscope, must select one healthy sperm and take it with a syringe needle. For this reason, the ICSI method is used only if the IVF procedure does not give the desired results. For artificial insemination, a sample of healthy sperm can be selected with hyaluron binding, centrifugation, magnets, or seemingly simple, compact devices [7]. For the in vitro fertilization procedure, such sampling is sufficient, but the incidence of the procedure itself and the percentage of women who have given birth remain very low, ranging from 30 to 35%. Currently, the selection of sperm cells for artificial insemination takes from 30 min to 120 min [8].
To increase the conception rate of artificial insemination, a computer-aided method to determine the suitability (i.e., motility) of a sperm sample for artificial insemination based on the microscope footage is required. Recently, artificial intelligence (AI)-based methods, such as convolutional neural networks (CNNs) and deep learning combined with computer vision methods, were adopted for multiple biomedical imaging applications, such as disease classification [9,10], edge detection [11], image segmentation [12], knowledge inference [13], image reconstruction [14], shape recognition [15], and others [16].
For ART in general, and for the analysis of sperm quality in particular, the AI techniques have also been adopted for classifying semen data and images [17]. For example, Nissen et al. [18] compared common convolutional neural network (CNN) architectures for human sperm cell-segmentation and recognition in semen sample images, achieving 93.87% precision and 91.89% recall for the best-analyzed network architecture. Hicks et al. [19] predicted the percentage of progressive, non-progressive, and immotile sperm heads from sperm images using ResNet-18 and ResNet-50 models and transfer learning. Movahed et al. [20] used a combined CNN-kmeans-SVM approach to segment the exterior and interior parts of the sperm heads for quantitative morphological analysis of sperm heads. Mohammed et al. [21] adopted a twin support vector machine (TWSVM) for the selection of features from a wider set of texture analysis-based image features, while a back-propagation neural network (BPNN) was used for the classification of healthy/unhealthy human sperm heads, achieving very good results. Their dataset however, was rather small. Thambawita et al. [22] suggested a two-stage architecture, where an autoencoder was used to extract sperm image features, and the pre-trained Resnet-34 CNN was employed for predicting motility and morphology of sperm heads. Butola et al. [23] used feedforward deep neural networks (DNNs) to recognize normal and stress-affected sperm cells. They achieved an accuracy of 85.6% on a dataset of 10,163 interferometric images of sperm cells. Ilhan et al. [24] used multi-stage cascade connected image preprocessing, region-based descriptors and a non-linear kernel support vector machine (SVM) for the recognition of sperm morphology from stained sperm images. The preprocessing techniques such as wavelet-based local adaptive de-noising, overlapping group shrinkage, and automatic directional masking increased the accuracy by up to 10%. In another paper, Ilhan et al. [25] suggested using the Mobile-Net neural network, which has a relatively low number of parameters, for the classification of normal/ab-normal sperms, achieving 87% accuracy. Iqbal et al. [26] developed a custom convolutional neural network (CNN) architecture to categorize human sperm heads, achieving 88% recall on the SCIAN dataset and 95% recall on the HuSHeM dataset. Javadi and Mirroshandel [27] trained a custom CNN to recognize morphological abnormalities of sperm heads and achieved over 83% accuracy in recognizing acrosome, head, and vacuole defects. Kandel et al. [28] used an optimized U-Net network to label the pixels in the sperm image as “head”, “midpiece”, “tail”, or “background”. The results were used to evaluate the dry mass as a proxy for infertility assessment. Riordon et al. [29] adopted VGG16, a deep CNN originally trained on ImageNet, and a transfer learning approach to retrain the network for sperm classification. The study has achieved 94.1% accuracy on the HuSHeM dataset and 62% accuracy on the SCIAN dataset. Shaker et al. [30] adopted a dictionary learning approach to create a dictionary of sperm head shapes. The latter was then employed to categorize the sperm heads into four different classes, achieving 92.2% accuracy on the HuSHeM dataset. McCallum et al. [31] implemented a deep-learning model that employed the VGG1647 network pre-trained on the ImageNet48 dataset with an additional global average pooling layer, succeeded by two fully connected layers. The new model was validated on the dataset of 1064 sperm cell images to predict the DNA Fragmentation Index (DFI).
Similar research has also been performed on non-human sperm, such as cattle or canine semen. Velasco et al. [32] used InceptionResNetV2, Xception, DenseNet121, DenseNet169, MobileNetV1, InceptionV3, and DenseNet201 for the classification of cattle sperm, and they achieved an accuracy of 98.3% on a dataset of 602 images. In another paper, Velasco et al. [33] adopted Inception_ResNetV2 and transfer learning for canine semen evaluation, achieving the highest accuracy of 87%. Hidayatullah et al. [34] adopting a pre-trained DarkNet53 network for sperm-cell detection, achieving 86.91 mAP on the test dataset and a processing speed of 50.3 fps on bull-sperm observation videos. Nevertheless, the use of deep learning neural networks is not fully automatic, in the sense that they still require manual annotation of images used for training a neural network, which is a tedious task in itself [35].
This article proposes a novel deep learning-based algorithm to determine, from a video, whether a semen sample is suitable for an artificial insemination procedure. Our novelty is the application of a Faster R-CNN neural network for sperm head segmentation and identification, combined with a heuristic algorithm for sperm head viability (motility) evaluation.
The remaining parts of the article are as follows: Section 2 describes the dataset used and the methods proposed. Section 3 presents and describes the experimental results. Finally, Section 4 evaluates and discusses the results and presents conclusions.

2. Materials and Methods

2.1. Dataset

For our experiments, we used the videos and data variables available from the VISEM-dataset [36]. The Norwegian research laboratory Simula conducted a study that sought to establish a link between a man′s reproductive function and overweight and obesity factors [36]. The embryologists of the study analyzed the sperm samples submitted by the participants and assessed sperm motility, sperm head concentration, total sperm head count, ejaculate volume, sperm morphology, and viability. Samples were examined under a 400-fold magnification microscope with an integrated heated table (37 °C) and captured by a microscope camera (UEye UI-2210C, IDS Imaging Development Systems, Obersulm, Germany). Sample observation videos were recorded for the sperm motility study (see an example of a still image in Figure 2). The dataset contains more than 30 GB of videos, each lasting from two to seven minutes. The videos have a 640 × 480 px resolution and a 50 fps frame rate.
The dataset includes information collected from 85 participants over the age of 18. Sperm analysis was performed for each participant, videos of each sample recorded, sperm fatty acids and their composition were measured, demographic data were recorded, and the results of the spermatogram analysis were performed according to World Health Organization (WHO) requirements, which included single ejaculation volume, sperm color, biological fluid dilution time, and sperm viscosity.
A descriptive table of the main parameters of the dataset to show the median and range of the data is presented in Table 1.

2.2. Image Annotation

Image tagging (annotation) is the manual process of recognizing different image elements when tags are added to an image. A separate XML annotation file was created for each image using LabelImg, an image annotation tool. The LabelImg tool uses 2D rectangles to mark an object. An example of the annotation results can be seen in Figure 3. A database of 650 frames with sperm was created to identify the neural network for sperm. An average of 30 sperm heads were recorded in one frame. As only sperm head motility was examined, the sperm tails were not included in the tags, and only the head and neck were marked (Figure 3).

2.3. Neural Network

Faster R-CNN [37], a deep CNN, was applied to each region proposal defined in the previous step, which classified the region proposals into objects and background. The input frame was presented to the network, and a pre-trained convolutional neural network (Inception, Resnet, MobileNet, VGG-16) extracted the characteristics of the input frame. Features were sent to two different components of the Faster R-CNN architecture (see Figure 4).
The Faster R-CNN-Inception_V2-COCO model [37] was used to develop the sperm counting algorithm. Its advantage is that it is quite accurate—able to learn to distinguish objects even from a small database. The model takes about 50 ms to analyze the image. Because the model uses regions, it is easier to detect a larger number of objects in a single frame. Additionally, the R-CNN model can read information from frames of any size, so there is no need to crop frames before submitting them to the neural network (whereas the SSD_MobileNet neural network model only accepts 300 × 300 px frames).
The Inception neural network model [38] (see Figure 5) performs several different transformations on the same input frame, combining the results into a single output. For each layer, the 5 × 5, 3 × 3, and 1 × 1 convolutional transformations were performed (features of interest are extracted from the input frame), a 3 × 3 maximum value extraction operator (max pooling) was applied, and the transformation results were output to a single array. The array with the results was passed to another layer, and in the next layer the model decides how and what information to use next.
The parameters of the neural network were set as follows:
  • The number of classes is 1 because in the Tensorflow API (Python) the background is not considered as a class, and only objects themselves are counted as a single class;
  • The maximum number of first stage proposals—since there can be an average of 30 sperm per frame (the value ranges from several units to 100), the selected value is 100;
  • The maximum number of detections per class—the default value is 100. Since there can be on average 30 sperm per frame (value ranges from several units to 100), the selected value is 100;
  • The maximum number of total detections—the default value is 100. The project includes one class, depending on the maximum number of detections per class, the selected value is 100;
  • Score converter—the sigmoid function is used.

2.4. Spermatozoid Tracking Algorithm

The proposed active sperm count algorithm is based on tracking the central coordinates of a sperm head. When the program is started, the first frame is scanned (Figure 6). A sperm_track list type variable is created to store the sperm distances traveled. A trained neural network is loaded into the program. The network takes the scanned frame as an input tensor and outputs the output tensor—the coordinates (detection_boxes) of the rectangles bounding the objects (detection_scores), the detected object classes (detection_classes), and the number of detected objects (num_detections).
The proposed method (see Figure 7) reads the coordinates of all detected objects (Figure 7b). If the score assigned to an object is greater than the threshold value of 0.8, the coordinates of the bounding boxes are stored (Figure 7c).
The center coordinates of the bounding rectangles were calculated (Figure 7c), and the average of the start and end coordinates was derived. The center coordinates were appended to the input_centroids array. Then, the method was performed, checking the dictionary type list of objects; if it was empty, then the program did not track any objects. If the list of objects was not empty, the distance between the coordinates of the input objects and the previously recorded coordinates of objects was calculated (Figure 8) using the Euclidean formula to calculate the distance between two points (Equation (1)):
d ( p , q ) =   i = 1 n ( q i   p i ) 2 ;
here q i the coordinates of the first point, and p i the coordinates of the second point.
All obtained distances were checked using the following conditions: if the distance was greater than 80 pixels, then the minimum distances were calculated and applied to the corresponding objects according to their ID number. Within one second, a viable sperm must swim at 34.5 µm/s [39]. Since the size of one pixel is 0.0002 m and the videos were filmed under a magnifying microscope of 400 times, which means that the size of one pixel in the database frames is 0.66 µm, the sperm will be considered viable if it traveled a distance of 52 px/s. Sperm vitality is the ratio of the sum of viable sperm to the total distance traveled by sperm, expressed as a percentage. If the activity of the semen sample was greater than 58% [4], the sample was considered to be suitable for insemination.

3. Results

3.1. Results of Spermatozoid Viability Evaluation

From the test results of the active sperm counting algorithm shown in Table 2, the viability of only one video clip segment corresponded to that found in the database. The average deviation of the estimated viability of the ten videos from the experimental one was 5.3%. An algorithm that achieves this result is suitable for product development. This may be because only a small portion of the video was being explored. The entire video could not be explored because the location of the video changed several times in each video.

3.2. Ablation Study

Three learning rates were chosen for training in neural network sperm recognition: 0.00002, 0.0002, and 0.002 (Table 3). In all cases, the neural network learned 2000 steps. The neural network with the lowest learning rate did not have enough steps; it was able to recognize only 20% of the objects in the frame (Figure 9a), while the neural network with the highest learning rate did not absorb the input information, and only 87 objects were found in the same frame (Figure 9c). The neural network with the assigned learning rate of 0.0002 showed the best results; it detected 30 objects in the frame (Figure 9b), the average score offered by the network for the detected object was 0.85. This learning indicator is also applied to further training.
Four iterations were employed to test neural network sperm recognition training (Table 4). After selecting 50,000 iterations, the network detected 30 objects in the frame (Figure 9), the lowest of all calculations being accurate. After selecting 100,000 iterations, the algorithm detected 32 objects out of 33, and the average score for the detected object reached 0.86. After reaching 150,000 iterations, the number of detected objects decreased to 31 out of 33, but the average value of the score for the detected object increased to 0.96. After reaching 200,000 iterations, the algorithm detected all the objects in the frame.
After 200,000 iterations, the neural network loss dropped to 0.22 (Figure 10). Neural network training lasted 4 days. The loss curve decreased exponentially, which means that the algorithm had successfully learned and adapted to the input task.
The mean accuracy of sperm head detection on the testing dataset was 91.77% (95% CI, 91.11%–92.43%).
The analysis of the sperm vitality prediction results showed a very good correlation with actual sperm vitality established using laboratory analysis methods (Figure 11). The Pearson correlation coefficient was 0.969 (95% CI, 0.968–0.97). The confidence interval (CI) was established using the bootstrapping method. Furthermore, the statistical significance of the results was tested by a paired t-test, where a p-value of less than 0.05 was deemed to be significant. Usually, the t-test assumes that samples are independent. The difference between motile and non-motile sperm heads was statistically significant (p < 0.001), as indicated by the paired t-test. The mean speed of motile sperm heads was 162.5 px/s (95% CI, 157.8–167.2 px/s), while the mean speed of motile sperm heads was 33.35 px/s (95% CI, 31.48–35.22) (see a boxplot representation in Figure 12).
We also calculated the mean absolute error (MAE) of sperm vitality prediction. Following the suggestion presented in Hicks et al. [19], the results with a mean MAE value below 11 were considered as significant improvements when compared to the ZeroR baseline, which assumes that the predicted values are equal to the average value computed over the dataset. Our results show that in all sperm samples we have achieved an average MAE value of less than 11 (Figure 13), while the grand mean MAE was 2.92 (95% CI, 2.46–3.37).

4. Evaluation and Conclusions

For practical purposes, the classification of human sperm heads can be simplified to a two-class recognition problem, as normal and abnormal sperm heads. In this paper, the criterion for abnormality was the speed of moving sperm heads as an indicator of their motility. The developed deep learning method based on the R-CNN network architecture for sperm head detection was tested on the dataset of sperm videos. The responses of the Faster R-CNN network to different learning rates and number of iterations were investigated experimentally. The algorithm achieved the best results by choosing the learning rate of 0.0002, and the number of iterations of 200,000. These parameters are optimal for the sperm cell image segmentation to efficiently detect small objects such as sperm heads.
Our results show that deep learning neural networks can be used to assess sperm motility consistently and efficiently for the needs of the artificial insemination procedure. The motility of the semen sample calculated by the algorithm differs from the experimental result by only 2.92% on average, while the accuracy of sperm head detection was 91.77%. However, in some cases, our method has failed to correctly recognize sperm heads because of occlusion, similar artifacts, or the sperm head being connected to the video frame border, which made it look like a dark spot. The reason for this result may be that the video fragments used to test the algorithm were too small. The entire video could not be tested because the video area changed several times during the video, and the objects being tracked changed, but the algorithm did not capture the change.
Future work will include the validation of the proposed method on the Human Sperm Head Morphology dataset (HuSHeM) [40].

Author Contributions

Conceptualization, V.R.; methodology, V.R. and R.M.; software, V.V.; validation, V.R., R.M. and R.D.; formal analysis, V.R. and T.K.; investigation, V.V. and V.R.; resources, V.V. and V.R.; data curation, V.V.; writing—original draft preparation, V.V. and V.R.; writing—review and editing, R.M. and R.D.; visualization, V.V., R.M., and R.D.; supervision, V.R.; funding acquisition, T.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Publicly available dataset was analyzed in this study. This data can be found here: https://datasets.simula.no/visem/.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Agarwal, A.; Mulgund, A.; Hamada, A.; Chyatte, M.R. A unique view on male infertility around the globe. Reprod. Biol. Endocrinol. 2015, 13, 37. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Nachtigall, R.D.; Faure, N.; Glass, R.H. Artificial insemination of husband’s sperm. Fertil. Steril. 1979, 32, 141–147. [Google Scholar] [CrossRef] [PubMed]
  3. Kumar, N.; Singh, A.K. Trends of male factor infertility, an important cause of infertility: A review of literature. J. Hum. Reprod. Sci. 2015, 8, 191–196. [Google Scholar] [CrossRef]
  4. Cooper, T.G.; Noonan, E.; von Eckardstein, S.; Auger, J.; Baker, H.W.G.; Behre, H.M.; Haugen, T.B.; Kruger, T.; Wang, C.; Mbizvo, M.T.; et al. World Health Organization reference values for human semen characteristics. Hum. Reprod. Update 2009, 16, 231–245. [Google Scholar] [CrossRef] [PubMed]
  5. Guzick, D.S.; Overstreet, J.W.; Factor-Litvak, P.; Brazil, C.K.; Nakajima, S.T.; Coutifaris, C.; Vogel, D.L. Sperm morphology, motility, and concentration in fertile and infertile men. N. Engl. J. Med. 2001, 345, 1388–1393. [Google Scholar] [CrossRef]
  6. Wang, J.; Sauer, M.V. In vitro fertilization (IVF): A review of 3 decades of clinical innovation and technological advancement. Ther. Clin. Risk Manag. 2006, 2, 355–364. [Google Scholar] [CrossRef] [Green Version]
  7. Natali, I. Sperm Preparation Techniques for Artificial Insemination—Comparison of Sperm Washing, Swim Up, and Density Gradient Centrifugation Methods. In Artificial Insemination in Farm Animals; InTech: London, UK, 2011. [Google Scholar] [CrossRef] [Green Version]
  8. Koyun, E. The effect of intrauterine insemination time on semen parameters. J. Turk. Ger. Gynecol. Assoc. 2014, 15, 82–85. [Google Scholar] [CrossRef] [Green Version]
  9. Khan, M.A.; Ashraf, I.; Alhaisoni, M.; Damaševičius, R.; Scherer, R.; Rehman, A.; Bukhari, S.A.C. Multimodal brain tumor classification using deep learning and robust feature selection: A machine learning application for radiologists. Diagnostics 2020, 10, 565. [Google Scholar] [CrossRef]
  10. Sahlol, A.T.; Elaziz, M.A.; Jamal, A.T.; Damaševičius, R.; Hassan, O.F. A novel method for detection of tuberculosis in chest radiographs using artificial ecosystem-based optimisation of deep neural network features. Symmetry 2020, 12, 1146. [Google Scholar] [CrossRef]
  11. Orujov, F.; Maskeliūnas, R.; Damaševičius, R.; Wei, W. Fuzzy based image edge detection algorithm for blood vessel detection in retinal images. Appl. Soft Comput. J. 2020, 94, 106452. [Google Scholar] [CrossRef]
  12. Khan, M.A.; Khan, M.A.; Ahmed, F.; Mittal, M.; Goyal, L.M.; Jude Hemanth, D.; Satapathy, S.C. Gastrointestinal diseases segmentation and classification based on duo-deep architectures. Pattern Recognit. Lett. 2020, 131, 193–204. [Google Scholar] [CrossRef]
  13. Gao, Z.; Wang, X.; Sun, S.; Wu, D.; Bai, J.; Yin, Y.; de Albuquerque, V.H.C. Learning physical properties in complex visual scenes: An intelligent machine for perceiving blood flow dynamics from static CT angiography imaging. Neural Netw. 2020, 123, 82–93. [Google Scholar] [CrossRef] [PubMed]
  14. Wei, W.; Zhou, B.; Połap, D.; Woźniak, M. A regional adaptive variational PDE model for computed tomography image reconstruction. Pattern Recognit. 2019, 92, 64–81. [Google Scholar] [CrossRef]
  15. Woźniak, M.; Połap, D.; Kośmider, L.; Cłapa, T. Automated fluorescence microscopy image analysis of pseudomonas aeruginosa bacteria in alive and dead stadium. Eng. Appl. Artif. Intell. 2018, 67, 100–110. [Google Scholar] [CrossRef]
  16. Ke, Q.; Zhang, J.; Wei, W.; Damaševičius, R.; Woźniak, M. Adaptive independent subspace analysis of brain magnetic resonance imaging data. IEEE Access 2019, 7, 12252–12261. [Google Scholar] [CrossRef]
  17. Fernandez, E.I.; Ferreira, A.S.; Cecílio, M.H.M.; Chéles, D.S.; de Souza, R.C.M.; Nogueira, M.F.G.; Rocha, J.C. Artificial intelligence in the IVF laboratory: Overview through the application of different types of algorithms for the classification of reproductive data. J. Assist. Reprod. Genet. 2020, 37, 2359–2376. [Google Scholar] [CrossRef] [PubMed]
  18. Nissen, M.S.; Krause, O.; Almstrup, K.; Kjærulff, S.; Nielsen, T.T.; Nielsen, M. Convolutional Neural Networks for Segmentation and Object Detection of Human Semen. In Image Analysis. SCIA 2017. Lecture Notes in Computer Science; Sharma, P., Bianchi, F., Eds.; Springer: Cham, Switzerland, 2017; Volume 10269. [Google Scholar] [CrossRef] [Green Version]
  19. Hicks, S.A.; Andersen, J.M.; Witczak, O.; Thambawita, V.; Halvorsen, P.; Hammer, H.L.; Riegler, M.A. Machine learning-based analysis of sperm videos and participant data for male fertility prediction. Sci. Rep. 2019, 9, 16770. [Google Scholar] [CrossRef] [PubMed]
  20. Movahed, R.A.; Mohammadi, E.; Orooji, M. Automatic segmentation of sperm’s parts in microscopic images of human semen smears using concatenated learning approaches. Comput. Biol. Med. 2019, 109, 242–253. [Google Scholar] [CrossRef] [PubMed]
  21. Mohammed, K.K.; Afify, H.M.; Fouda, F.; Hassanien, A.E.; Bhattacharyya, S.; Vaclav, S. Classification of human sperm head in microscopic images using twin support vector machine and neural network. In Advances in Intelligent Systems and Computing; Springer: Singapore, 2020; pp. 857–871. [Google Scholar] [CrossRef]
  22. Thambawita, V.; Pål Halvorsen, P.; Hammer, H.; Riegler, M.; Haugen, T.B. Extracting temporal features into a spatial domain using autoencoders for sperm video analysis. arXiv 2019, arXiv:1911.03100. [Google Scholar]
  23. Butola, A.; Popova, D.; Prasad, D.K.; Ahmad, A.; Habib, A.; Tinguely, J.C.; Ahluwalia, B.S. High spatially sensitive quantitative phase imaging assisted with deep neural network for classification of human spermatozoa under stressed condition. Sci. Rep. 2020, 10, 13118. [Google Scholar] [CrossRef]
  24. Ilhan, H.O.; Serbes, G.; Aydin, N. Automated sperm morphology analysis approach using a directional masking technique. Comput. Biol. Med. 2020, 122, 103845. [Google Scholar] [CrossRef]
  25. Ilhan, H.O.; Sigirci, I.O.; Serbes, G.; Aydin, N. A fully automated hybrid human sperm detection and classification system based on mobile-net and the performance comparison with conventional methods. Med Biol. Eng. Comput. 2020, 58, 1047–1068. [Google Scholar] [CrossRef]
  26. Iqbal, I.; Mustafa, G.; Ma, J. Deep Learning-Based Morphological Classification of Human Sperm Heads. Diagnostics 2020, 10, 325. [Google Scholar] [CrossRef]
  27. Javadi, S.; Mirroshandel, S.A. A novel deep learning method for automatic assessment of human sperm images. Comput. Biol. Med. 2019, 109, 182–194. [Google Scholar] [CrossRef]
  28. Kandel, M.E.; Rubessa, M.; He, Y.R.; Schreiber, S.; Meyers, S.; Naves, L.M.; Popescu, G. Reproductive outcomes predicted by phase imaging with computational specificity of spermatozoon ultrastructure. Proc. Natl. Acad. Sci. USA 2020, 117, 18302–18309. [Google Scholar] [CrossRef]
  29. Riordon, J.; McCallum, C.; Sinton, D. Deep learning for the classification of human sperm. Comput. Biol. Med. 2019, 111, 103342. [Google Scholar] [CrossRef]
  30. Shaker, F.; Monadjemi, S.A.; Alirezaie, J.; Naghsh-Nilchi, A.R. A dictionary learning approach for human sperm heads classification. Comput. Biol. Med. 2017, 91, 181–190. [Google Scholar] [CrossRef]
  31. McCallum, C.; Riordon, J.; Wang, Y.; Kong, T.; You, J.B.; Sanner, S.; Lagunov, A.; Hannam, T.G.; Jarvi, K.; Sinton, S. Deep learning-based selection of human sperm with high DNA integrity. Commun. Biol. 2019, 2, 250. [Google Scholar] [CrossRef] [Green Version]
  32. Velasco, J.S.; Arago, N.M.; Mamba, R.M.; Padilla, M.V.C.; Ramos, J.P.M.; Virrey, G.C. Cattle sperm classification using transfer learning models. Int. J. Emerg. Trends Eng. Res. 2020, 8, 4325–4331. [Google Scholar] [CrossRef]
  33. Velasco, J.S.; Padilla, M.V.C.; Arago, N.M.; De Vera, E.P.E.; Domingo, F.E.M.; Ramos, R.E.R. Canine semen evaluation using transfer learning models. Int. J. Emerg. Trends Eng. Res. 2020, 8, 1–6. [Google Scholar]
  34. Hidayatullah, P.; Wang, X.; Yamasaki, T.; Mengko, T.L.E.R.; Munir, R.; Barlian, A.; Sukmawati, E.; Supraptono, S. DeepSperm: A robust and real-time bull sperm-cell detection in densely populated semen videos. arXiv 2020, arXiv:2003.01395. [Google Scholar]
  35. Hidayatullah, P.; Mengko, T.E.R.; Munir, R.; Barlian, A. A semiautomatic sperm cell data annotator for convolutional neural network. In Proceedings of the 5th International Conference on Science in Information Technology: Embracing Industry 4.0: Towards Innovation in Cyber Physical System, ICSITech 2019, Yogyakarta, Indonesia, 23–24 October 2019; pp. 211–216. [Google Scholar] [CrossRef]
  36. Haugen, T.B.; Hicks, S.A.; Andersen, J.M.; Witczak, O.; Hammer, H.L.; Borgli, R.; Halvorsen, P.; Riegler, M. VISEM: A multimodal video dataset of human spermatozoa. In Proceedings of the 10th ACM Multimedia Systems Conference, MMSys’19, Amherst, MA, USA, 18–21 June 2019; pp. 261–266. [Google Scholar] [CrossRef] [Green Version]
  37. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  38. Alom, Z.; Hasan, M.; Yakopcic, C.; Tarek, M.; Taha, T.M. Inception Recurrent Convolutional Neural Network for Object Recognition. arXiv 2017, arXiv:1704.07709. [Google Scholar]
  39. Holt, W.V.; Moore, H.D.M.; Hillier, S.G. Computer-assisted measurement of sperm swimming speed in human semen: Correlation of results with in vitro fertilization assays. Fertil. Steril. 1985, 44, 112–119. [Google Scholar] [CrossRef]
  40. Shaker, F. Human Sperm Head Morphology Dataset (HuSHeM) [Data set]; Mendeley: London, UK, 2018. [Google Scholar] [CrossRef]
Figure 1. Visual illustration of good and poor motility of a sperm head.
Figure 1. Visual illustration of good and poor motility of a sperm head.
Sensors 21 00072 g001
Figure 2. Still image from the sperm sample video.
Figure 2. Still image from the sperm sample video.
Sensors 21 00072 g002
Figure 3. Spermatozoid annotation using LabelImg application.
Figure 3. Spermatozoid annotation using LabelImg application.
Sensors 21 00072 g003
Figure 4. Faster Region Based Convolutional Neural Networks (R-CNN) network architecture.
Figure 4. Faster Region Based Convolutional Neural Networks (R-CNN) network architecture.
Sensors 21 00072 g004
Figure 5. Principle of operation of the Inception model.
Figure 5. Principle of operation of the Inception model.
Sensors 21 00072 g005
Figure 6. Proposed spermatozoid tracking algorithm.
Figure 6. Proposed spermatozoid tracking algorithm.
Sensors 21 00072 g006
Figure 7. Basic scheme of frame processing: (a) sperm sample image, (b) segmented images of spermatozoa, (c) coordinates of bounding boxes of spermatozoa.
Figure 7. Basic scheme of frame processing: (a) sperm sample image, (b) segmented images of spermatozoa, (c) coordinates of bounding boxes of spermatozoa.
Sensors 21 00072 g007
Figure 8. Illustration of the search for the least distant point in the neighborhood of a sperm head.
Figure 8. Illustration of the search for the least distant point in the neighborhood of a sperm head.
Sensors 21 00072 g008
Figure 9. Results of applying different learning speeds: (a) learning_rate = 0.00002, (b) learning_rate = 0.0002, (c) learning_rate = 0.002.
Figure 9. Results of applying different learning speeds: (a) learning_rate = 0.00002, (b) learning_rate = 0.0002, (c) learning_rate = 0.002.
Sensors 21 00072 g009
Figure 10. Neural network loss during training.
Figure 10. Neural network loss during training.
Sensors 21 00072 g010
Figure 11. Correlation study of predicted sperm vitality vs. actual sperm vitality. The regression line is plotted (R2 = 0.94).
Figure 11. Correlation study of predicted sperm vitality vs. actual sperm vitality. The regression line is plotted (R2 = 0.94).
Sensors 21 00072 g011
Figure 12. Difference between motile and non-motile sperm heads as determined by the proposed algorithm.
Figure 12. Difference between motile and non-motile sperm heads as determined by the proposed algorithm.
Sensors 21 00072 g012
Figure 13. The histogram of mean absolute error (MAE) value distribution.
Figure 13. The histogram of mean absolute error (MAE) value distribution.
Sensors 21 00072 g013
Table 1. Descriptive statistics of the dataset.
Table 1. Descriptive statistics of the dataset.
ParameterStatisticsValue
Sperm concentration (×106/mL)Median (range)
Mean ± std
68 (4–350)
82.33 ± 64.11
Average number of sperm heads in one frameMedian (range)34 (2–175)
Mean ± std41.16 ± 32.05
Table 2. Testing results of the active sperm counting algorithm.
Table 2. Testing results of the active sperm counting algorithm.
No.Frame IDObjects, PcsExperimental Viability, %Estimated Viability, %Deviation, %
FoundLiveViable
1180786885872
214654968412
32221151075723
423107685850
53021201191874
64255524385832
75190887786882
87992867697889
98180746481876
1082393424857213
Table 3. Neural network results of applying different learning speeds.
Table 3. Neural network results of applying different learning speeds.
Learning SpeedNumber of StepsObjectsAverage Score
MarkRecognizeWrong
0.00002200010730.71
0.00022000303000.85
0.0022000873000.78
Table 4. Results of applying a different number of iterations for neural network training.
Table 4. Results of applying a different number of iterations for neural network training.
Number of IterationsDetected ObjectsIdentified ObjectsUnrecognized ObjectsAccuracyAverage Score
50,000303030.90.88
100,000323210.970.86
150,000313120.940.96
200,0003333010.97
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Valiuškaitė, V.; Raudonis, V.; Maskeliūnas, R.; Damaševičius, R.; Krilavičius, T. Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination. Sensors 2021, 21, 72. https://doi.org/10.3390/s21010072

AMA Style

Valiuškaitė V, Raudonis V, Maskeliūnas R, Damaševičius R, Krilavičius T. Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination. Sensors. 2021; 21(1):72. https://doi.org/10.3390/s21010072

Chicago/Turabian Style

Valiuškaitė, Viktorija, Vidas Raudonis, Rytis Maskeliūnas, Robertas Damaševičius, and Tomas Krilavičius. 2021. "Deep Learning Based Evaluation of Spermatozoid Motility for Artificial Insemination" Sensors 21, no. 1: 72. https://doi.org/10.3390/s21010072

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop