Next Article in Journal
Association of Excessive Daytime Sleepiness with the Zung Self-Rated Depression Subscales in Adults with Coronary Artery Disease and Obstructive Sleep Apnea
Next Article in Special Issue
An End-to-End Pipeline for Early Diagnosis of Acute Promyelocytic Leukemia Based on a Compact CNN Model
Previous Article in Journal
Quality Improvements in Management of Children with Acute Diarrhea Using a Multiplex-PCR-Based Gastrointestinal Pathogen Panel
Previous Article in Special Issue
Method for Diagnosing the Bone Marrow Edema of Sacroiliac Joint in Patients with Axial Spondyloarthritis Using Magnetic Resonance Image Analysis Based on Deep Learning
 
 
Comment published on 18 July 2022, see Diagnostics 2022, 12(7), 1738.
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Accuracy of New Deep Learning Model-Based Segmentation and Key-Point Multi-Detection Method for Ultrasonographic Developmental Dysplasia of the Hip (DDH) Screening

1
Department of Orthopedic Surgery, Dongsan Medical Center, School of Medicine, Keimyung University, Daegu 42601, Korea
2
Department of Orthopedic Surgery, Korea University Anam Hospital, 73, Goryeodae-ro, Seongbuk-gu, Seoul 02841, Korea
3
Department of Biomedical Engineering, Keimyung University, Daegu 42601, Korea
4
Department of Anatomy, Dongsan Medical Center, School of Medicine, Keimyung University, Daegu 42601, Korea
*
Author to whom correspondence should be addressed.
Diagnostics 2021, 11(7), 1174; https://doi.org/10.3390/diagnostics11071174
Submission received: 28 May 2021 / Revised: 24 June 2021 / Accepted: 25 June 2021 / Published: 28 June 2021
(This article belongs to the Special Issue Clinical Diagnosis Using Deep Learning)

Abstract

:
Hip joint ultrasonographic (US) imaging is the golden standard for developmental dysplasia of the hip (DDH) screening. However, the effectiveness of this technique is subject to interoperator and intraobserver variability. Thus, a multi-detection deep learning artificial intelligence (AI)-based computer-aided diagnosis (CAD) system was developed and evaluated. The deep learning model used a two-stage training process to segment the four key anatomical structures and extract their respective key points. In addition, the check angle of the ilium body balancing level was set to evaluate the system’s cognitive ability. Hence, only images with visible key anatomical points and a check angle within ±5° were used in the analysis. Of the original 921 images, 320 (34.7%) were deemed appropriate for screening by both the system and human observer. Moderate agreement (80.9%) was seen in the check angles of the appropriate group (Cohen’s κ = 0.525). Similarly, there was excellent agreement in the intraclass correlation coefficient (ICC) value between the measurers of the alpha angle (ICC = 0.764) and a good agreement in beta angle (ICC = 0.743). The developed system performed similarly to experienced medical experts; thus, it could further aid the effectiveness and speed of DDH diagnosis.

1. Introduction

Developmental dysplasia of the hip (DDH) is a hip joint disease with various possible causes, including genetics, intrauterine factors, and cultural practices. Although the incidence rates vary by race and ethnicity, many developed countries report DDH rates ranging from 1.5 to 20 cases per 1000 births [1,2]. DDH causes multiple spectrum disorders, such as mild capsular laxity, acetabular deficiency, subluxation, and hip dislocation. This spectrum gradually progresses in severity, with treatment becoming more difficult as the patient ages and grows. Fortunately, DDH can be treated simply with a brace or splint when diagnosed early [3,4]. Otherwise, secondary anatomical changes caused by delayed diagnosis could require surgical treatment [5,6]. If the diagnosis is further delayed or DDH is not diagnosed, osteoarthritis of the hip joint may occur at an early age, requiring major surgical treatment (e.g., hip replacement) [7]. Therefore, DDH prognosis depends on early diagnosis and treatment. Recently, many countries (e.g., Germany, Israel, Korea) have conducted DDH screening tests on newborns at about six weeks of age because of the advantages of early diagnosis [8,9,10].
DDH diagnosis is done through radiography, physical examination, and US imaging. Radiography is an effective DDH diagnostic method when femoral head ossification occurs in children aged more than six months. Similarly, physical examinations can only diagnose complete hip dislocation. Thus, neither are adequate early DDH screening methods. The preferred means of early DDH screening is ultrasonographic (US) imaging because it can capture cartilage conditions before femoral head ossification occurs. In addition, it can be used to diagnose mild spectrums of DDH aside from dislocation. Moreover, there is no concern about ionizing radiation by US imaging.
The most common DDH US analysis is the Graf method [11], which measures a bony acetabular depth or coverage of cartilaginous with labrum in a coronal US image of the hip (Figure 1). However, this method has high interoperator variability [12,13,14] that could lead to misdiagnosis of DDH in half of the infant screenings and up to three-quarters in neonatal screenings [13]. The primary sources of image acquisition variability between operators arise from the manual selection of anatomical points for calculating angles (i.e., tri-radiate cartilage, the acetabular bone edge, the lower edge of the iliac bone, and the end of the acetabular labrum). An artificial intelligence (AI)-based computer-aided diagnosis (CAD) system can standardize image acquisition and reduce diagnosis time, thus improving the accuracy and objectivity of DDH diagnosis.
The concept of CAD emerged in the 1970s when scanned medical images were analyzed using a computer. In 1998, LeCun et al. [15] laid the foundations for today’s deep learning by outlining the convolutional neural network (CNN) framework—a neural network being a derivative of mathematical models that simulate the structure and function of biological neural networks. As graphics processing unit (GPU) images improved with advances in technology, deep learning, mainly through CNNs, has become a conventional approach to medical imaging and is even used for genetic analysis [16,17,18]. CNN is an effective tool in image recognition, especially in judging image borders and colors. In addition, it has shown excellent diagnostic performance based on endoscopic images and magnetic resonance imaging (MRI) [19,20]. Hence, multiple studies on DDH diagnosis using US images of the hip joint have been conducted [21,22].
Region-based CNNs (R-CNN) apply CNN for object detection in an image. Object detection consists of proposing a region in an image where an object is supposed to exist and analyzing the region. Despite R-CNN’s excellent object detection accuracy, its processing is slow because it performs CNN for too many regions in an image proposed by its selective search feature. Thus, Fast R-CNNs have been proposed to improve efficiency [23]. This type of R-CNN performs CNN for the entire image, proposes regions of interest (RoIs), and then uses the region of interest pooling (RoIPool) to make the objects of interest a feature of a fixed resolution. However, Fast R-CNN’s process is still time-consuming. Faster R-CNNs hasten the process using a region proposal network (RPN) to propose a region from a feature map. Mask R-CNN is based on Faster R-CNN and has the advantage of automatic image segmentation. Specifically, it can define and propose RoIs based on boundary contours of the anatomical structure in hip US images using the region of interest alignment (RoIAlign) rather than merely using boxes. Furthermore, it also can find specific points in an image [24].
Because US images are generally of low resolution and in grayscale, US imaging is at a disadvantage. Specifically, a large amount of image data are required to construct an artificial intelligence (AI) system that automatically detects anatomical structures and obtains the needed angles without human intervention. This disadvantage obstructs research and development in a single institution. Thus, this study segmented a relatively small number of US images using Mask R-CNN and developed a two-stage multi-detection method AI system in a single research institution to extract the needed data from the segmented images. To the best of the researchers’ knowledge, this is the first study to estimate automatically detected key points in US images using a deep neural network model.

2. Materials and Methods

The institutional review board approved the study (IRB No. DSMC 2019-10-003). The study’s dataset consisted of 1243 hip US images from 168 infants, using a 12.5 MHz linear probe with HD15 and HD7 XE ultrasound systems (Philips, Bothell, WA, USA) during a DDH neonatal screening and diagnosis program between 2002 and 2019, retrospectively. Of the total 1243 images, 289 images were randomly selected for AI training, 33 for validation, and the remaining 921 for testing.

2.1. Images

One radiologist and one pediatric orthopedic surgeon collected all images, both with more than 10 years of US DDH screening experience. These images were anonymized and documented in the vertically oriented standard plane as Digital Imaging and Communications in Medicine (DICOM) files.
Three lines were drawn to distinguish the localized anatomical structures that represent the alpha and beta Graf angles. However, alpha and beta angle measurements without parallelization to the vertical ilium are ineffective when evaluating hip dysplasia. Thus, a “check angle” was used to define the angle created by the iliac wing and baseline. Check angle values measured within ± 5° by both a pediatric orthopedic surgeon (referred to in this study as “doctor”) and the AI system were classified as the “appropriate image”. These were the only alpha and beta angles used to evaluate AI performance.

2.2. Multi-Detection Type Artificial Intelligence (AI)

In previous studies, the ilium, the acetabular roof, and the labrum in ultrasound images were segmented. Furthermore, the angles created by these three structures were calculated through an algorithm to obtain the alpha and beta Graf angles [21,22]. However, in practice, doctors calculate the same angles by extracting measurement points based on anatomical understanding. In this study, the ilium’s, the acetabulum’s, and labrum’s shapes were segmented using the Mask R-CNN method. Then, two check angle points and another three points were determined for calculating the alpha and beta angles. Of these points, one is an overlap of a check angle point and the middle tri-radiate cartilage point. Overall, four points were used to calculate the alpha and beta angles, as seen in Figure 2.

2.2.1. Segmentation

Segmentation was done using the Mask R-CNN technique, wherein a mask head is added to a Faster R-CNN [24]. Specifically, the Mask R-CNN adds a mask branch to determine whether a particular pixel corresponds to an object. Mask R-CNN is similar to Faster R-CNN, using region proposal network (RPN) to extract features and classify and tighten bounding boxes. Faster R-CNN uses RoIPool as a feature extraction method for quantifying each RoI region, thus solving the problem of differing RoI feature sizes at different scales by max pooling. However, this process causes spatial information loss, leading to the displacement of the original image RoI and extraction features. To solve this problem, Mask R-CNN replaces the RoI pooling of Faster R-CNN with RoI alignment (RoIAlign) to mark the object area. Figure 3 illustrates the image segmentation process of both Mask R-CNN and Faster R-CNN methods. The red arrows depict the Mask R-CNN method using region of interest alignment (RoIAlign) to draw a precise bounding box that matches an object and extracts specific key points from the object of interest. In this case, the areas of the ilium, acetabulum, and labrum were extracted using a mask head. The region of interest alignment (RoIAlign) and feature pyramid network structures were then used to extract features for the residual neural network (ResNet) algorithm consisting of 50 layers modified with the background Mask R-CNN algorithm. The black arrows show the Faster R-CNN method using the region of interest pooling (RoIPool), which creates a square-bound box containing an object.

2.2.2. File Conversion

Each patient’s data, including his/her age and sex, were documented as a DICOM file. All identifying information was removed, following ethical guidelines. The anonymized data were then converted into 256-bit grayscale portable network graphics (PNG) image files.

2.2.3. Layer Extraction

AI training started with extract image segmentation of the US images. The Computer Vision Annotation Tool (CVAT; Massachusetts Institute of Technology, MA, USA), a free online program that helps annotate videos and images for computer vision algorithms [25], was used to designate the area for the corresponding mask position (Figure 4).

2.2.4. First Training

Using CVAT’s point tool, the key anatomical structures in the US images were designated, while the extensible markup language (XML) data were obtained using the extraction tool. The first AI training used the initial modeling data composed of the generated XML file and image file. In this stage, the areas corresponding to the ilium, acetabulum, and labrum were converted to white (RGB 255, 255, 255) and the rest to black (RGB 0, 0, 0) as seen in Figure 5.

2.2.5. Second Training

Using CVAT, the points corresponding to the tri-radiate cartilage, the acetabular bone edge, the lower edge of the iliac bone, and the end of the acetabular labrum were marked with red dots. The secondary training was performed using the resulting image data of the first training and an XML file marked with the four key points (Figure 6).

2.2.6. Final Result

The multi-detection process was completed after the second training, wherein the alpha and beta angles were obtained. A diagram of the implementation process is shown in Figure 7.

2.3. Network Training

The training takes 40,000 steps in total, with validation at every 1000 steps to verify the best model selection. The learning rate and weight decay were 0.01 and 0.0001, respectively, while the stochastic gradient descent used a batch size of 10. The following data augmentation settings were applied: up/down-left/right random shift, −5–5%; random scale, 80–100%; random rotation, −15–15°; and random brightness, 75–100%. In addition, the Random Gaussian Noise setting was used for the robustness of various ultrasonic noises and brightness.

2.4. Statistical Analysis

The check angle and Graf angles of the same images calculated by a pediatric orthopedic surgeon and the AI were compared. The R language version 3.3.3 (R Foundation for Statistical Computing, Vienna, Austria) program was used for all statistical analyses. Percent agreement, positive percent agreement, and Cohen’s kappa were computed to assess inter-rater reliability (IRR) of the AI’s and doctor’s determined check angles. Furthermore, the intraclass correlation coefficient (ICC) was measured to evaluate the agreement of the AI’s and the doctor’s computed alpha and beta angles.

3. Results

3.1. Criteria for Data Classification

Before the performance evaluation, finding an image capable of measuring the alpha and beta angles from the 921 test data images was necessary. Two criteria were set: (1) all four key anatomical structures and their respective points (i.e., tri-radiate cartilage, the acetabular bone edge, the lower edge of the iliac bone, and the end of the acetabular labrum) were visible; and (2) the iliac wing is parallel to the baseline.

3.1.1. Data Classification Based on First Criteria

Hence, two measurers categorized the images based on detectability and as “detectable” if the key anatomical structures and their respective points are observable. The images were classified as “undetectable” if otherwise.
The doctor classified 542 images as “detectable” images; in contrast, the AI classified 555 images as “detectable” images. Overall, the 512 images were commonly categorized as “detectable” images. The 409 images that at least one doctor or AI could not detect were classified as “fail detection” (Table 1).

3.1.2. Data Classification Based on Second Criteria

Images categorized as “detectable” in the first classification were classified once again according to whether the correct alpha and beta angles could be measured. The image was defined to measure the correct alpha and beta angles when the baseline and the iliac wing are generally parallel. Thus, if images of the iliac wing parallel to the baseline measured a check angle value within ±5°, they were classified as “OK check angle” and meet the specified criteria. Otherwise, they were classified as “error check angle”.
The 512 commonly categorized “detectable” images from the performance evaluation were used to classify images according to check angle. Of these images, 320 were classified as “OK check angle” by both the doctor and AI (Table 2). The interobserver check angle agreement shows that the percent agreement at 80.9% and kappa coefficient of 0.5 (95% confidence interval 0.52; 0.5) is reasonably good. In addition, a positive agreement of 0.87 existed between the observers.

3.2. Statistical Result of Graf Angle Measurement

Only images classified as “detectable” with “OK check angle” (thus meeting both criteria) were defined as appropriate images and considered suitable for an appropriate DDH diagnosis. Figure 8 shows examples of the image classification. As a result, the alpha and beta Graf angles of the 320 appropriate images were compared. Similar to Cicchetti [26], the study considers an intraclass correlation coefficient (ICC) > 0.70 acceptable. The resulting ICC revealed excellent agreement for the alpha angle (ICC = 0.76) and good agreement for the beta angle (ICC = 0.74) (Table 3 and Figure 9). For the alpha-angle, 74.7% of images displayed a discrepancy of less than 5° between the observers. In addition, the observers agreed in the classification of 84.3% of cases as being normal or abnormal, as seen in Figure 10. The mean absolute deviation (MAD) of the alpha angle was 3.4 while it was 4.5 for the beta angle. The standard deviation of differences was 4.5°, similar to the interhuman observer variability reported in previous studies (Table 4) [12,27].

4. Discussion

US imaging is a noninvasive and safe diagnostic tool that offers excellent results for DDH screening [8], especially when using the Graf method. However, intraobserver and interobserver variances in the measurement of hip US imaging remain an issue [12,13,14]. The primary source of variability arises from the selection of anatomical points for calculating angles. In response, an automatic key point system for DDH diagnosis was developed based on a multi-detection deep learning system. Because it is a multi-training system, small data sizes can be analyzed.
The agreement between the AI system and the doctor was excellent in the alpha-angle (ICC = 0.76) and good for the beta-angle (ICC = 0.74) calculations. In addition, by analyzing 320 hip US images, the AI system differentiated normal and abnormal hip US images with a match rate of 84.37% to the doctor. These results show that the system improved from previous methods [12,28]. Simon et al. [28] evaluated concordance between orthopedic surgeons and pediatricians for the diagnosis of DDH, and the reported concordance rate was 82.91%, with class correlation coefficients for alpha angles (ICC = 0.72) and beta angles (ICC = 0.34). Roovers et al. [12] reported a match rate of 85% between doctors in the classification of 200 hip US images as normal or dysplasia. Thus, the proposed new deep learning system produced a higher interobserver agreement, and results with experienced orthopedic practitioners can be regarded as comparably good.
Our study had several limitations. First, although there were many images in the dataset, the number of images and image sources was still small. Thus, the results cannot be generalized. Second, the study sample of doctors was small, having two medical experts in data collection and one in measuring the Graf angles in images of the control group. Because of interoperator/interhuman variability, it is necessary to compare the images collected by more US imaging operators and angles calculated by more doctors. Finally, the study did not validate the data using external datasets.
DDH becomes challenging to treat when the diagnosis is delayed; hence, screening tests are essential. This new AI-based CAD system can be a helpful tool for inexperienced physicians to measure angles in hip images and lessen the interobserver variability of DDH diagnosis.

5. Conclusions

A CAD deep learning model based on segmentation and key-point multi-detection using Mask R-CNN was developed for DDH screening and diagnosing. The system’s performance was comparable to that of human experts. Hence, it can be used as an auxiliary method for DDH screening.

Author Contributions

Conceptualization, S.-W.L. & W.-Y.J.; Methodology, S.-W.L.; Software, J.-H.L.; Validation, S.-M.H.; Formal Analysis, Y.-R.H.; Investigation, H.-U.Y.; Resources, H.-U.Y.; Data Curation, K.-J.L.; Writing-Original Draft Preparation, H.-U.Y.; Writing-Review and Editing, H.-U.Y.; Visualization, S.-W.L.; Supervision, S.-W.L.; Project Administration, S.-W.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Ministry of Health and Welfare (South Korea), with the support of the Korea Health Information Service’s project for medical data-driven hospitals.

Institutional Review Board Statement

This study was conducted according to the guidelines of the Declaration of Helsinki, and approved by the Keimyung University Dongsan Hospital Institutional Review Board (IRB No. DSMC 2019-10-003).

Informed Consent Statement

As this study is a retrospective study, written informed consent was waived with the approval of the IRB.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to patient’s right to privacy.

Acknowledgments

We thank Young-Hee Son, Yoon-Jung Kang, and Chul-Woo Jegal for their support with data collection.

Conflicts of Interest

No benefits in any form have been received or will be received from a commercial party related directly or indirectly to the subject of this article.

References

  1. Patel, H. Canadian Task Force on Preventive Health Care. Preventive health care, 2001 update: Screening and management of developmental dysplasia of the hip in newborns. CMAJ 2001, 164, 1669–1677. [Google Scholar] [PubMed]
  2. Bialik, V.; Bialik, G.M.; Blazer, S.; Sujov, P.; Wiener, F.; Berant, M. Developmental dysplasia of the hip: A new approach to incidence. Pediatrics 1999, 103, 93–99. [Google Scholar] [CrossRef]
  3. Atalar, H.; Sayli, U.; Yavuz, O.Y.; Uraş, I.; Dogruel, H. Indicators of successful use of the Pavlik harness in infants with developmental dysplasia of the hip. Int. Orthop. 2007, 31, 145–150. [Google Scholar] [CrossRef] [Green Version]
  4. Pavone, V.; de Cristo, C.; Vescio, A.; Lucenti, L.; Sapienza, M.; Sessa, G.; Pavone, P.; Testa, G. Dynamic and static splinting for treatment of developmental dysplasia of the hip: A systematic review. Children 2021, 8, 104. [Google Scholar] [CrossRef] [PubMed]
  5. Ganz, R.; Klaue, K.; Vinh, T.S.; Mast, J.W. A new periacetabular osteotomy for the treatment of hip dysplasias: Technique and preliminary results. Clin. Orthop. Relat. Res. 1988, 232, 26–36. [Google Scholar] [CrossRef]
  6. Löchel, J.; Janz, V.; Perka, C.; Hofer, A.; Zimmerer, A.; Wassilew, G.I. A new rectus and sartorius sparing approach for periacetabular osteotomy in patients with developmental dysplasia of the hip. J. Clin. Med. 2021, 10, 601. [Google Scholar] [CrossRef]
  7. Furnes, O.; Lie, S.A.; Espehaug, B.; Vollset, S.E.; Engesaeter, L.B.; Havelin, L.I. Hip disease and the prognosis of total hip replacements: A review of 53,698 primary total hip replacements reported to the Norwegian Arthroplasty Register 1987–1999. J. Bone Jt. Surg. Br. 2001, 83, 579–586. [Google Scholar] [CrossRef]
  8. Tomà, P.; Valle, M.; Rossi, U.; Brunenghi, G.M. Paediatric hip—ultrasound screening for developmental dysplasia of the hip: A review. Eur. J. Ultrasound 2001, 14, 45–55. [Google Scholar] [CrossRef]
  9. Woolacott, N.F.; Puhan, M.A.; Steurer, J.; Kleijnen, J. Ultrasonography in screening for developmental dysplasia of the hip in newborns: Systematic review. BMJ 2005, 330, 1413. [Google Scholar] [CrossRef] [Green Version]
  10. Dogruel, H.; Atalar, H.; Yavuz, O.Y.; Sayli, U. Clinical examination versus ultrasonography in detecting developmental dysplasia of the hip. Int. Orthop. 2008, 32, 415–419. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  11. Graf, R. The diagnosis of congenital hip-joint dislocation by the ultrasonic combound treatment. Arch. Orthop. Trauma Surg. 1980, 97, 117–133. [Google Scholar] [CrossRef]
  12. Roovers, E.A.; Boere-Boonekamp, M.M.; Geertsma, T.S.A.; Zielhuis, G.A.; Kerkhoff, A.H. Ultrasonographic screening for developmental dysplasia of the hip in infants. Reproducibility of assessments made by radiographers. J. Bone Jt. Surg. Br. 2003, 85, 726–730. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Jaremko, J.L.; Mabee, M.; Swami, V.G.; Jamieson, L.; Chow, K.; Thompson, R.B. Potential for change in US diagnosis of hip dysplasia solely caused by changes in probe orientation: Patterns of alpha-angle variation revealed by using three-dimensional US. Radiology 2014, 273, 870–878. [Google Scholar] [CrossRef]
  14. Dias, J.J.; Thomas, I.H.; Lamont, A.C.; Mody, B.S.; Thompson, J.R. The reliability of ultrasonographic assessment of neonatal hips. J. Bone Jt. Surg. Br. 1993, 75, 479–482. [Google Scholar] [CrossRef]
  15. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  16. Awan, M.J.; Rahim, M.S.M.; Salim, N.; Mohammed, M.A.; Garcia-Zapirain, B.; Abdulkareem, K.H. Efficient detection of knee anterior cruciate ligament from magnetic resonance imaging using deep learning approach. Diagnostics 2021, 11, 105. [Google Scholar] [CrossRef] [PubMed]
  17. Azer, S.A. Challenges facing the detection of colonic polyps: What can deep learning do? Medicina 2019, 55, 473. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  18. Von Wrede, R.; Jeub, M.; Ariöz, I.; Elger, C.E.; von Voss, H.; Klein, H.-G.; Becker, A.J.; Schoch, S.; Surges, R.; Kunz, W.S. Novel KCNH1 Mutations Associated with Epilepsy: Broadening the Phenotypic Spectrum of KCNH1-Associated Diseases. Genes 2021, 12, 132. [Google Scholar] [CrossRef]
  19. Park, H.-C.; Kim, Y.-J.; Lee, S.-W. Adenocarcinoma recognition in endoscopy images using optimized convolutional neural networks. Appl. Sci. 2020, 10, 1650. [Google Scholar] [CrossRef] [Green Version]
  20. Dachena, C.; Casu, S.; Fanti, A.; Lodi, M.B.; Mazzarella, G.J. Combined Use of MRI, fMRIand Cognitive Data for Alzheimer’s Disease: Preliminary Results. Appl. Sci. 2019, 9, 3156. [Google Scholar] [CrossRef] [Green Version]
  21. Golan, D.; Donner, Y.; Mansi, C.; Jaremko, J.; Ramachandran, M. Fully automating Graf’s method for DDH diagnosis using deep convolutional neural networks. In Lecture Notes in Computer Science; Carneiro, G., Ed.; Springer: Cham, Switzerland, 2016; Volume 10008, pp. 130–141. [Google Scholar] [CrossRef]
  22. Hareendranathan, A.R.; Zonoobi, D.; Mabee, M.; Cobzas, D.; Punithakumar, K.; Noga, M.; Jaremko, J.L. Toward automatic diagnosis of hip dysplasia from 2D ultrasound. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, VIC, Australia, 18–21 April 2017; pp. 982–985. [Google Scholar] [CrossRef]
  23. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  24. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R.B. Mask R-CNN. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), 22–29 October 2017; pp. 2961–2969. [Google Scholar] [CrossRef]
  25. Sekachev, B. Computer Vision Annotation Tool: A Universal Approach to Data Annotation. Available online: https://software.intel.com/content/www/us/en/develop/articles/computer-vision-annotation-tool-a-universal-approach-to-data-annotation.html (accessed on 10 September 2020).
  26. Cicchetti, D.V. Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychol. Assess. 1994, 6, 284–290. [Google Scholar] [CrossRef]
  27. Zieger, M. Ultrasound of the infant hip. Part 2. Validity of the method. Pediatr. Radiol. 1986, 16, 488–492. [Google Scholar] [CrossRef] [PubMed]
  28. Simon, E.A.; Saur, F.; Buerge, M.; Glaab, R.; Roos, M.; Kohler, G. Inter-observer agreement of ultrasonographic measurement of alpha and beta angles and the final type classification based on the Graf method. Swiss Med. Wkly. 2004, 134, 671–677. [Google Scholar] [CrossRef]
Figure 1. (a) A standard hip ultrasonographic (US) image should show three key anatomical structures: 1. vertical ilium, 2. lower margin of Os ilium, and 3. labrum. (b) Alpha (α) and beta (β) Graf angles in a US image of the hip.
Figure 1. (a) A standard hip ultrasonographic (US) image should show three key anatomical structures: 1. vertical ilium, 2. lower margin of Os ilium, and 3. labrum. (b) Alpha (α) and beta (β) Graf angles in a US image of the hip.
Diagnostics 11 01174 g001
Figure 2. Calculating Graf angles using four points: (a) Two “check angle” points determine the appropriateness of the image. (b) Three points are used to calculate the alpha and beta angles, where one point on the tri-radiate cartilage overlaps with a check angle point.
Figure 2. Calculating Graf angles using four points: (a) Two “check angle” points determine the appropriateness of the image. (b) Three points are used to calculate the alpha and beta angles, where one point on the tri-radiate cartilage overlaps with a check angle point.
Diagnostics 11 01174 g002
Figure 3. Comparing the Mask region-based convolutional neural network (Mask R-CNN; red arrows) and Faster R-CNN (black arrows) methods for image segmentation.
Figure 3. Comparing the Mask region-based convolutional neural network (Mask R-CNN; red arrows) and Faster R-CNN (black arrows) methods for image segmentation.
Diagnostics 11 01174 g003
Figure 4. Sample model training using Computer Vision Annotation Tool (CVAT).
Figure 4. Sample model training using Computer Vision Annotation Tool (CVAT).
Diagnostics 11 01174 g004
Figure 5. (a) Original image ultrasonographic image of the hip. (b) The same image after the initial conversion.
Figure 5. (a) Original image ultrasonographic image of the hip. (b) The same image after the initial conversion.
Diagnostics 11 01174 g005
Figure 6. Point marks for the second training based on the first training’s resulting image.
Figure 6. Point marks for the second training based on the first training’s resulting image.
Diagnostics 11 01174 g006
Figure 7. Diagrammatic illustration of the multi-detection type artificial intelligence (AI) training courses: (a) area learning training using CVAT Program, (b) initial image file conversion, (c) the secondary point-marking training, and (d) the final result of a multi-detection-trained AI system.
Figure 7. Diagrammatic illustration of the multi-detection type artificial intelligence (AI) training courses: (a) area learning training using CVAT Program, (b) initial image file conversion, (c) the secondary point-marking training, and (d) the final result of a multi-detection-trained AI system.
Diagnostics 11 01174 g007
Figure 8. Examples of ultrasonographic image classifications in the study: (a) Case of the appropriate image where detection of all key points was possible, the measured check angle was less than 5° for both observers, and the alpha and beta angles were evaluated (n = 320). (b) Case of the inappropriate image where detection of all key points was possible, but the check angle was more than 5° for at least one observer (n = 192). (c) Case of the fail detection where at least one observer could not detect all key points (n = 409).
Figure 8. Examples of ultrasonographic image classifications in the study: (a) Case of the appropriate image where detection of all key points was possible, the measured check angle was less than 5° for both observers, and the alpha and beta angles were evaluated (n = 320). (b) Case of the inappropriate image where detection of all key points was possible, but the check angle was more than 5° for at least one observer (n = 192). (c) Case of the fail detection where at least one observer could not detect all key points (n = 409).
Diagnostics 11 01174 g008
Figure 9. (a) Correlation scatter plot of the alpha angles measured by the artificial intelligence (AI) system and the doctor (Human). (b) Correlation scatter plot of the beta angles measured by the artificial intelligence (AI) system and the doctor (Human).
Figure 9. (a) Correlation scatter plot of the alpha angles measured by the artificial intelligence (AI) system and the doctor (Human). (b) Correlation scatter plot of the beta angles measured by the artificial intelligence (AI) system and the doctor (Human).
Diagnostics 11 01174 g009
Figure 10. Scatter plot of the alpha angles measured by the artificial intelligence (AI) system and the doctor (Human) on the y-axis and x-axis, respectively. The black points between the dotted lines represent a correct diagnosis wherein the difference in the observers’ measurements is less than 5°. These specific points account for 74% of the correct diagnoses made. Furthermore, the black points on the first and third quadrants represent correct diagnoses. Following the Graf method, the observers had the same normal and abnormal classifications based on a 60° alpha angle. These specific points account for 84% of all the measurements made. In contrast, values expressed as red crosses represent the observers’ different normal and abnormal classification results. These crosses represent 16% of all the measurements made.
Figure 10. Scatter plot of the alpha angles measured by the artificial intelligence (AI) system and the doctor (Human) on the y-axis and x-axis, respectively. The black points between the dotted lines represent a correct diagnosis wherein the difference in the observers’ measurements is less than 5°. These specific points account for 74% of the correct diagnoses made. Furthermore, the black points on the first and third quadrants represent correct diagnoses. Following the Graf method, the observers had the same normal and abnormal classifications based on a 60° alpha angle. These specific points account for 84% of all the measurements made. In contrast, values expressed as red crosses represent the observers’ different normal and abnormal classification results. These crosses represent 16% of all the measurements made.
Diagnostics 11 01174 g010
Table 1. Data classification based on first criteria.
Table 1. Data classification based on first criteria.
Subgroupn (%)
Doctor 921 (100)
Detectable542 (58.9)
Undetectable379 (41.1)
Artificial Intelligence 921 (100)
Detectable555 (60.3)
Undetectable366 (39.7)
Doctor and Artificial Intelligence 921 (100)
Commonly detectable512 (55.6)
Fail detection409 (44.4)
Table 2. Data classification according to the check angles of detectable images.
Table 2. Data classification according to the check angles of detectable images.
n (%)Check Angle (Doctor)
OK
Check Angle (Doctor)
Error
All detectable image512 (100)
Check angle (AI): OK 320 (86.7)49 (34.3)
Check angle (AI): Error 49 (13.3)94 (65.7)
Table 3. Intraclass correlation coefficient (ICC) scatter plot between artificial intelligence (AI) and doctor using alpha and beta angle as measurement.
Table 3. Intraclass correlation coefficient (ICC) scatter plot between artificial intelligence (AI) and doctor using alpha and beta angle as measurement.
ICC95% Confidence IntervalAgreement
Alpha angle0.7640.699–0.815Excellent
Beta angle07430.689–0.788Good
Table 4. Mean and standard deviation of angles measured from AI and doctor.
Table 4. Mean and standard deviation of angles measured from AI and doctor.
n (%)Artificial
Intelligence (AI)
HumanAI–HumanMean Absolute Deviation (MAD)
Alpha angle320 (100)62.843 ± 6.51464.117 ± 7.12−1.274 ± 4.5553.470
Beta angle320 (100)40.785 ± 7.6940.679 ± 8.710.106 ± 5.94.501
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lee, S.-W.; Ye, H.-U.; Lee, K.-J.; Jang, W.-Y.; Lee, J.-H.; Hwang, S.-M.; Heo, Y.-R. Accuracy of New Deep Learning Model-Based Segmentation and Key-Point Multi-Detection Method for Ultrasonographic Developmental Dysplasia of the Hip (DDH) Screening. Diagnostics 2021, 11, 1174. https://doi.org/10.3390/diagnostics11071174

AMA Style

Lee S-W, Ye H-U, Lee K-J, Jang W-Y, Lee J-H, Hwang S-M, Heo Y-R. Accuracy of New Deep Learning Model-Based Segmentation and Key-Point Multi-Detection Method for Ultrasonographic Developmental Dysplasia of the Hip (DDH) Screening. Diagnostics. 2021; 11(7):1174. https://doi.org/10.3390/diagnostics11071174

Chicago/Turabian Style

Lee, Si-Wook, Hee-Uk Ye, Kyung-Jae Lee, Woo-Young Jang, Jong-Ha Lee, Seok-Min Hwang, and Yu-Ran Heo. 2021. "Accuracy of New Deep Learning Model-Based Segmentation and Key-Point Multi-Detection Method for Ultrasonographic Developmental Dysplasia of the Hip (DDH) Screening" Diagnostics 11, no. 7: 1174. https://doi.org/10.3390/diagnostics11071174

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop