Next Article in Journal
Intervertebral Canals and Intracanal Ligaments as New Terms in Terminologia anatomica
Previous Article in Journal
Study of Root Transparency in Different Postmortem Intervals Using Scanning Electron Microscopy
Previous Article in Special Issue
Machine Learning Assisting the Prediction of Clinical Outcomes following Nucleoplasty for Lumbar Degenerative Disc Disease
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Automatic Refractive Error Estimation Using Deep Learning-Based Analysis of Red Reflex Images

1
oDocs Eye Care, Dunedin 9013, New Zealand
2
University of Sydney, Sydney, NSW 2050, Australia
3
Choithram Netralaya, Indore 453112, India
4
Dargaville Medical Centre, Dargaville 0310, New Zealand
5
Public Health Unit, Dunedin Hospital, Te Whatu Ora Southern, Dunedin 9016, New Zealand
*
Author to whom correspondence should be addressed.
Diagnostics 2023, 13(17), 2810; https://doi.org/10.3390/diagnostics13172810
Submission received: 17 July 2023 / Revised: 23 August 2023 / Accepted: 26 August 2023 / Published: 30 August 2023
(This article belongs to the Special Issue Medical Image Processing and Analysis)

Abstract

:
Purpose: We evaluate how a deep learning model can be applied to extract refractive error metrics from pupillary red reflex images taken by a low-cost handheld fundus camera. This could potentially provide a rapid and economical vision-screening method, allowing for early intervention to prevent myopic progression and reduce the socioeconomic burden associated with vision impairment in the later stages of life. Methods: Infrared and color images of pupillary crescents were extracted from eccentric photorefraction images of participants from Choithram Hospital in India and Dargaville Medical Center in New Zealand. The pre-processed images were then used to train different convolutional neural networks to predict refractive error in terms of spherical power and cylindrical power metrics. Results: The best-performing trained model achieved an overall accuracy of 75% for predicting spherical power using infrared images and a multiclass classifier. Conclusions: Even though the model’s performance is not superior, the proposed method showed good usability of using red reflex images in estimating refractive error. Such an approach has never been experimented with before and can help guide researchers, especially when the future of eye care is moving towards highly portable and smartphone-based devices.

1. Introduction

The most common cause of distance vision impairment is an uncorrected refractive error, of which myopia is the leading disorder [1]. The global prevalence of myopia and high myopia (≤−5.00 D) is rising rapidly and they are expected to affect 50% (4.7 billion) and 10% (1 billion) of the global population by 2050, respectively [2]. Myopia is an emergent public health issue recognized by the World Health Organization (WHO) as one of the leading causes of preventable blindness [3].
High myopia is associated with significant ocular co-morbidities, many of which cause irreversible vision loss, including myopic macular degeneration, macular neovascularization, glaucoma, and retinal detachment [4,5,6]. Even low myopes (−1.00 to −3.00 D) experience a two-fold increase in developing myopia-associated ocular morbidities [7,8]. Without effective intervention, vision loss from myopia-associated ocular pathology is expected to increase sevenfold [2].
Myopia and associated complications cause a significant individual public burden with an estimated annual cost of USD 202 billion worldwide [9]. The socioeconomic burden of myopia is expected to be further exacerbated by the declining age of onset and faster progression in children [10,11,12]. Effective intervention to control myopic progression is critical to reducing the disease burden for individuals and wider society [12]. To address the growing myopic burden, there is an urgent need for a cost-effective myopia screening program to provide efficacious treatment.
Vision screening programs, such as the New Zealand B4 School Check (B4SC) and Australian Statewide Eyesight Screening Program (StEPS), rely on visual acuity testing to detect refractive error [13]. Visual acuity-based screening programs are screener-dependent and produce high false positives with poor positive predictive values [14,15]. Autorefractor-based screening, although more straightforward and cost-effective, is limited by its necessity for maintaining position and visual fixation for a sufficient and substantial amount of time, which can be difficult in uncooperative children [16].

1.1. Red Reflex Test

The red reflex test, also known as the Brückner test, was pioneered by Brückner in 1962 [17] and has been postulated as a potential screening test to detect refractive errors. The test is performed in a darkened room using a coaxial light source, typically a direct ophthalmoscope, with the examiner positioned at an arm’s length distance from the subject. The examiner then looks through the ophthalmoscope and focuses on both corneas simultaneously, noting a red reflex in each pupil [18]. A normal test consists of symmetrically bright and round red reflexes in both eyes [19]. Asymmetry of the reflexes may be associated with strabismus, anisometropia, media opacities or cataracts, retinoblastoma, colobomas, and pigment abnormalities of the posterior pole [18,20]. The location and size of the pupillary crescent can also indicate the presence of ametropia [21,22,23]. Kothari [20] used the pupillary crescent from the red reflex test to determine and classify a patient’s refractive state as myopic (inferior crescent > 1 mm), hyperopic (superior crescent ≥ 2 mm), or astigmatic (crescent at any location decentred by >1 clock hour). The test was determined to have a sensitivity of 91% and specificity of 72.8%, with overall data supporting the use of the Brückner test in screening for refractive errors. A recent study based on smartphone photography reported by Srivastava et al. [24] has used Brückner’s reflex and has demonstrated reliability in identifying amblyogenic conditions in school children.

1.2. Photoscreeners

Several studies [20,23,25,26] have also looked into photo screening, a technique that works on the principle of the Brückner test. Photorefraction devices capture images of the light crescent generated on the pupillary red reflex. Results have been comparable to the Brückner test [23,27]. However, unlike the Brückner test, which relies on a live, moving subject, photoscreener images permit an unlimited amount of time for interpretation [20]. Images can be either manually graded or automatically graded by prediction algorithms [28]. A wide variety of photoscreeners are now available, including the PlusoptiX, Welch Allyn Spot Vision Screener, and the smartphone-based GoCheckKids app [29,30,31].

1.3. Artificial Intelligence in Ophthalmology

In the past few years, there has been a tremendous expansion of research into artificial intelligence (AI) applications in health care [32]. Much of the progress in AI research is achieved through the use of deep learning (DL), a subset of AI that utilizes convolutional neural networks (CNNs), comprised multiple layers of algorithms, to perform high-level feature extraction [33,34]. DL allows a machine to automatically process and learn from raw datasets and analyze complex non-linear relationships [35,36]. One of the key benefits of using DL algorithms in medicine has been in medical imaging analysis and screening.
In ophthalmology, DL systems have been applied in the classification of retinal fundus images, visual field results, and optical coherence tomography (OCT), to aid in the detection of ocular pathologies such as refractive error [16,37], diabetic retinopathy [38,39,40], glaucoma [41,42], retinopathy of prematurity [43], macular oedema [44], and age-related macular degeneration [42,45,46,47]. DL has also been shown to predict clinical parameters such as high coronary artery calcium scores by extracting the hidden features of retinal fundus images, which human clinicians may not have picked up [48,49].
Varadarajan et al. [37] developed a deep learning-based prediction model with high accuracy in estimating refractive error from retinal fundus images. More recently, Chun et al. [16] trained a CNN with eccentric photorefraction photographs to categorize refractive errors into mild, moderate, and high myopia or hyperopia. Although the deep learning model demonstrated an overall accuracy of 81.6% in estimating refractive error compared to cycloplegic examination [16], the generalizability of the results is limited unless the same smartphone model and camera settings can be replicated.
This paper proposes an experimental and novel method to estimate refractive error using color and infrared (IR) red reflex images acquired by an economical and portable smartphone-based fundus camera device known as nun IR [50].

2. Methods

2.1. Data Preparation

The proposed study was a pooled analysis of two prospective cohort studies conducted at Dargaville Medical Center in Northland, New Zealand, and Choithram Hospital in India.
The study at Dargaville Medical Center aimed to evaluate the usability of retinal images taken by a smartphone-based fundus camera in a rural general practice setting. Patients aged 16 years and older who presented to Dargaville Medical Center with visual disturbance, headache, hypertensive urgency, transient ischaemic attack, or stroke were invited to participate in the study, which took place over one year from 15 November 2021 to 23 November 2022. Patients who could not consent, were non-ambulatory, in need of resuscitation, or deemed too unwell by the triage nurse or attending physician were excluded. Written informed consent was obtained from all 152 participants.
At Choithram Hospital, 360 patients aged 20–60 years were recruited. Exclusion criteria included any condition or disease that could potentially affect photorefraction and the ability to obtain a reliable photo image, such as media opacity (corneal scarring, cataracts), posterior segment pathology (retinal detachment, age-related macular degeneration, diabetic retinopathy, vitreous haemorrhage), nystagmus, neurological impairment, and poor cooperation. All patients provided written informed consent before participation.
Two types of red reflex images, IR and color, were captured from each patient using the nun IR fundus camera. An optometrist took photographs obtained from Choithram Hospital, while a clinician captured those from Dargaville Medical Center. Refractive error metrics, including spherical and cylindrical power, were obtained using an autorefractor.
Images were captured in a dark room with a nun IR fundus camera attached to an Android smartphone, as shown in Figure 1. The patient was instructed to look straight ahead at a green fixation light in the distance and not blink. This was to reduce accommodation, where changes in focus can dynamically affect the shape of the lens. Thus, if accommodation is not effectively controlled, this would potentially alter the refractive power of the lens and, consequently, the accuracy of the red reflex images.
Two images were taken for each patient: one of the right eye at a distance of 8cm and one of the left at a distance of 8 cm. Each image was then saved as a full-color red reflex image and an IR red reflex image, giving each patient a total of four red reflex images. The images were then cropped to include only the pupil, as shown in Figure 2.

2.2. IR-Based Imaging with nun IR

The nun IR handheld fundus camera is non-mydriatic, using IR light to allow the retina to be viewed without causing pupillary constriction. It sends out paraxial rays into the subject’s eye as shown in Figure 3.

2.3. Observation of Crescents

When viewing the anterior segment of the eye with the nun IR, two crescents typically appear (Figure 2). The observed crescents are far apart in patients with myopia, while in those with hyperopia, they are merged. A clear differentiation is seen when examining a model eye (Heine Ophthalmoscope Trainer) with the nun IR, as shown in Figure 4.
In the model eye, it was noted that when the crescents are far apart, the spherical power is negative, and when the crescents overlap, the spherical power is more positive. This phenomenon was also observed in actual patients. The top row of Figure 5 shows two visible crescents far apart in a myopic eye (−8.00 D). The second row of Figure 5 shows crescents merged in an emmetropic eye (0.00 D). These figures indicate that the positions of crescents in red reflex images could help predict myopia.

2.4. Grading Classification with CNN

After data preparation, the DL-based AI platform MedicMind [51] (Figure 6) was used to train a grading classifier to predict spherical and cylindrical power.
MedicMind uses the open-source framework TensorFlow with the CNNs Inception-v3 (Figure 7) and EfficientNet (Figure 8). Inception-v3 and EfficientNet are pre-trained models, although EfficientNet has a more advanced architecture. We implemented the transfer learning technique in both models to improve training accuracy and speed. MedicMind resized images to 299 × 299 for Inception-v3 and 384 × 384 for EfficientNet, as these are the required sizes for the pre-trained models. MedicMind also applied data augmentation to the red reflex images, changing the contrast, aspect ratio, flipping, and brightness to reduce overfitting. The last classification layer of each CNN was modified to become a grading classifier with one output, trained with Euclidean distance as the loss function, as opposed to softmax. The RMSProp optimizer was used for back-propagation.

2.5. Training, Validation, and Evaluation

We used the data from Choithram Netralaya as the training set. There were 357 patients from Choithram Hospital who were primarily of Indian ethnicity. The training set included 288 IR and 528 color images, as some photographs were excluded due to poor quality. The Choithram data had a median spherical power of −2.00 D, meaning that the patients were largely myopic. Accuracy was measured using the red reflex images of 152 patients from Dargaville Medical Centre as a validation set, which included 143 IR images and color images. Photographs that needed to be of better quality were excluded. Autorefraction results were provided as the ground truth. Thus, approximately 80% of images were used for training and 20% for validation, which were then evaluated with the MedicMind DL model to predict refractive error versus ground truth. All training data were scaled so that the median was 0.5, and predictions were considered correct if the predicted and the ground truth values were over 0.5 or if the predicted and ground truth values were less than 0.5.

3. Results

3.1. CNN Grading Results

The EfficientNet grading classifier gave 63% accuracy for spherical power when training with IR images and 70% for cylindrical power. For color red reflex images, 64% accuracy was obtained for spherical power and 57% for the cylinder.
The Inception-v3 grading classifier was of consistently lower accuracy than EfficientNet, with 55% accuracy for spherical power and 49% for cylinders with IR images. For color red reflex images, Inception-v3 achieved 54% accuracy for spherical power and 50% for the cylinder (Table 1).
For the grading of IR red reflex images with EfficientNet, 84% sensitivity and 47% specificity were obtained, being more pronounced in color with 94% sensitivity and 29% specificity. The grading classifier is good at predicting true positives but not as good at predicting true negatives.
Results were validated with the dataset from Dargaville Medical Centre. This dataset included predominantly patients of Māori (indigenous people of New Zealand) and European ethnicity. The data were cropped and processed using the same technique as Choithram, leaving 143 IR images. The distribution of spherical power is shown in Figure 9.
Figure 9 shows that patients were more hyperopic in the Dargaville medical center dataset (median spherical power of +0.50 D) compared to the Choithram hospital dataset (median spherical power of −2.00 D).
An accuracy of 52% was obtained when inference was performed on the 143 images from the Dargaville dataset using the DL models trained with the Choithram dataset. The training was also directly performed on the Dargaville dataset, giving 52% accuracy.

3.2. Categorizing into Crescent Types

The accuracy of both grading classifiers was consistently 70% or lower. A higher accuracy is necessary to be useful for myopia screening. IR images were manually categorized based on crescent types to determine how accuracy could be improved, as shown in Figure 10.
The spherical power of each image in the dataset was divided into crescent-type categories. Figure 11 shows that images belonging to category A, with crescents furthest apart, are universally myopic (negative spherical power), but those in category D can be either myopic or non-myopic. This is also observed with high sensitivity for both IR and color images (84% and 94%, respectively) and low specificity (47% and 29%, respectively), showing that the DL model can predict myopia considerably accurately but is not as accurate in detecting non-myopia.
Spherical power overlap can be seen in each category. The degree of overlap is more pronounced for cylinder, where there appears to be no correlation between crescent type and cylinder power, as shown in Figure 12.
Power overlap between different crescent types may be why the grading classifier does not give a higher accuracy.

3.3. Multiclass Classification with CNN

A multiclass classifier was trialed to improve the grading classifier results. The IR images were put into two bins, depending on whether the spherical power was greater or less than the median spherical power of −2.00 D. A multiclass classifier was then trained on the two classes.
Diagnostics 13 02810 i001
The following graph (Figure 13) shows the distribution of spherical powers for the 288 IR images:
After training a multiclass classifier for both Inception-v3 and EfficientNet, we obtained 70% accuracy for spherical power and 55% for the cylinder with EfficientNet for IR images, and 64% accuracy for spherical power and 60% for the cylinder with EfficientNet for color images (Table 2). Inception-v3 was consistently of lower accuracy for both IR and color images. Overall, color red reflex images gave lower accuracy than IR red reflex images when training with a multiclass classifier.
To improve the accuracy of the multiclass classifier, we trialed removing a certain percentage of images bordering on either side of the median so that the difference in spherical power between the two bins was more pronounced. Figure 14 shows the spherical power distribution when the middle 30% of images from the median was removed.
By removing 20% of the IR images bordering on either side of the median, the accuracy of a multiclass classifier was approximately 75% for spherical power and 50% for cylinder, with similar results for both EfficientNet and Inception-v3. After removing 20% of the color images bordering on either side of the median, accuracy was lower in terms of spherical power and cylinder for both EfficientNet and Inception-v3.
The removal of 1%, 5%, 10%, 20%, 30%, and 40% of images was trialed using Inception-v3 spherical power values, with the removal of 20% giving the best results (Table 3). However, removing the images only marginally improved accuracy.

3.4. Increasing Contrast

Increasing the contrast of images was also trialed to enhance the visibility of the crescents (Figure 15). This was tried for spherical power with EfficientNet with an accuracy of 69%. This was lower than the accuracy of 75% that was obtained before increasing contrast, indicating that increased contrast does not help improve accuracy.

3.5. Exclusion of Small Images

For EfficientNet, we trialed removing IR images that were small in size and less than 180 pixels wide. This resulted in a lower accuracy of 68% with the EfficientNet multiclass classifier, compared to the previous 74% (Table 4). The lower accuracy may have been due to fewer numbers (80 images) in each bin.
We also tried removing small-sized color red reflex images that were less than 450 pixels wide, as color images are much wider than IR images. We then trained spherical power with an EfficientNet multiclass classifier. We obtained 60% accuracy compared to 68% before removal. These results suggest that removing smaller images from the color red reflex images also does not improve accuracy.

3.6. Combining IR and Color

Rather than training on IR and color red reflex images separately, each IR and color image was placed side by side and combined into one picture in the hope that it would improve overall accuracy (Figure 16). The left side of the picture contained the IR image, and the right side contained the color image. Again, this did not improve accuracy, giving similar results to when trained separately (Table 5).

3.7. Summary of Results

The red reflex technique helps predict myopia with high accuracy. When crescents are type A (far apart), they are universally myopic, but if they are type D (merged), they can be either myopic or non-myopic (Figure 17).
The red reflex images above are all type A, where the crescents are apart and all myopic. This is also indicated by the high sensitivity values for predicting myopia with a grading classifier (84% and 94% for both IR and color).
However, when the red reflex images are all type D, only the image on the left is not myopic, with the other two images being myopic (Figure 18). This is also demonstrated in the low specificity values for IR and color image-based systems (47% and 29%, respectively).
The difficulty in classifying type D crescents as myopic or non-myopic was verified by the evaluation of the Dargaville dataset against Choithram-trained models. In the Dargaville dataset, there were far fewer type A images and more type D (median spherical power of +0.50 vs. −2.00 for Choithram). The low number of type A IR images and the high number of type D IR images in the Dargaville dataset is consistent with the low accuracy of 52% obtained with training the Dargaville dataset. Figure 19 shows a graphical summary of various training techniques used for the EfficientNet-based classifier and their performances.

4. Discussion

This study experimented with the possibility of using an AI-based model to predict refractive error from red reflex images taken using a smartphone-based portable fundus camera. The successful implementation of such a model would allow myopic individuals at risk of progression to be triaged and monitored remotely and digitally. This could reduce the number of clinical visits and improve treatment compliance, convenience, and patient satisfaction, thereby translating to saved time and costs for healthcare providers and consumers.
An effective screening test should be precise, economical, efficient, portable, and simple to administer. Although photoscreeners have gained considerable traction in research for their potential use in large-scale vision screening, the significant costs associated with many of these devices limit their practicality, particularly in developing and low-income countries. Nowadays, smartphones represent an integral part of modern life. Mobile applications among healthcare professionals in routine clinical practice are also becoming increasingly widespread, allowing direct access to medical knowledge, journals, calculators, and diagnostic tools [52]. As smartphone usage is so ubiquitous, photographs can be easily obtained virtually from anywhere in the world, overcoming the geographical and financial barriers of many photoscreeners.
Several studies have investigated the accuracy of smartphones for photorefractive screening [30,53,54]. Gupta et al. [53] utilized smartphone photography to screen for amblyogenic factors and found that all high refractive errors ≥ 5 D were successfully detected, although moderate refractive errors (3.25–5.00 D) revealed false negatives. However, the authors excluded low refractive errors due to the red reflex appearing normal, and the total number of subjects with a moderate or high refractive error was only 22, thereby limiting the applicability of their results. Another study [54] assessed the detection of amblyogenic risk factors using the GoCheckKids smartphone app. As both of these studies [53,54] evaluated the generalized prediction of amblyogenic risk factors and did not look at refractive error specifically, we could not directly compare the performance of our method to theirs. Arnold et al. [30] used the GoCheckKids app to identify refractive amblyopia risk factors, reporting a sensitivity of 76% and a specificity of 85%. However, these figures were achieved with manual grading of images by trained ophthalmic imaging specialists, which would involve substantial costs if this method were used for mass vision screening. An AI-based screening system could alleviate the burden of limited human resources, particularly in developing countries, and help economize this process.
In recent years, there has been a surge in the research and development of DL models to predict refractive error. In 2018, Varadarajan et al. [37] showed that contrary to existing beliefs, DL models could be trained to predict refractive error from retinal fundus images. Similarly, DL algorithms by Tan et al. [45] outperformed six human expert graders in detecting myopic macular degeneration and high myopia from retinal photographs. Another novel study [55] applied a DL algorithm to estimate uncorrected refractive error using posterior segment OCT images.
In 2020, Chua et al. [11] trained a deep CNN to predict categories of refractive error using 305 smartphone-captured eccentric photorefraction images, achieving an overall accuracy of 81.6%. However, their participants were all of one homogenous ethnicity (Korean), whereas our study included patients of Indian, European, and Māori descent. Furthermore, their study comprised only patients aged six months to 8 years, while our cohort had a much greater age range of participants. Our study also had a larger sample size of 512 participants compared to 164. Images used in their study were taken with a 16-megapixel LG smartphone camera, and the other aforementioned smartphone-based photorefractive studies have all used distinct models, such as the Nokia Lumia 1020 [54], the iPhone 7 [30], and the OPPOA37f [53]. Unless the smartphone model and/or camera settings are somewhat standardized, the consistency of results may be affected. The nun IR camera used in our study has been specifically designed to be compatible with Android smartphones and comes together with a Samsung A03 pre-installed with the nun IR mobile app. The nun IR camera used in our study has inbuilt optics for imaging the eye and only uses smartphone as a user interface. Hence, the images acquired are not dependent on the camera specifications of the smartphone used. As nun IR is also a fundus camera, unlike regular smartphone cameras, it can complete a more comprehensive screening assessment, examining pupillary red reflexes and the retinal fundus.
Our study has several limitations. Firstly, a considerable number of photographs could not be included for reasons such as excessive brightness and low quality, thus rendering the pupillary crescents undetectable. This may have been due to poor technique in image acquisition, as this is user-dependent and could be improved by a more rigorous and meticulous data collection system. Secondly, our DL model produced better results for true positives compared to true negatives. As the nun IR camera was primarily designed to be a fundus camera, the optics could be adjusted to enhance the visibility of the pupillary crescents in IR mode if the camera were to be repurposed for myopia detection. Thirdly, we used autorefraction to measure refractive error, despite cycloplegic refraction being the gold standard [30]; thus, accommodation may have needed to be sufficiently controlled. Finally, although our study had a wide age range of participants, we did not include any children under 16 years. As many vision screening programs are focused on the early detection of refractive error, future studies should also encompass a greater pediatric cohort. Further, since the proposed study is based on comparatively fewer data samples, the generalizability of the model is limited.
Future research that could address the above limitations can definitely improve the overall performance of such red reflex image-based refractive error estimation systems.

5. Conclusions

In summary, we developed a DL-based model to estimate uncorrected refractive error using red reflex images captured by the nun IR fundus camera. The proposed approach achieved 75% accuracy in predicting spherical power with a multiclass classifier.
The results are inferior and do not achieve the target we aimed for as a primary screening tool. However, it demonstrates a definite potential for future research in vision screening with better data collection procedures in place, a larger sample size, better targeted CNN architectures, and additional collection of pediatric data.

Author Contributions

Conceptualization, G.L., R.C., B.O. and S.C.H.; Methodology, G.L., R.C. and S.C.H.; Software, G.L. and R.C.; Validation, G.L. and R.C.; Investigation, G.L. and R.C.; Resources, D.S. and S.D.; Data curation, D.S. and S.D.; Writing—original draft, G.L., R.C., L.Z. and J.L.H.; Writing—review & editing, R.C., L.Z., J.L.H., B.O., S.D. and S.C.H.; Supervision, R.C.; Project administration, R.C.; Funding acquisition, R.C. and S.C.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by Precision Driven Health New Zealand Project Grant (1390) for the period 2020–23. The grant funded the first and second authors. The grant also covered the data collection expenses. The authors would like to thank the leadership team from Precision Driven Health New Zealand, especially Kevin Ross, Kelly Atkinson, and Fleur Armstrong, for their valuable feedback and guidance on the project. We would also like to thank L T Pagrani, Ashwini Varma, Kirti Malhotra, Srishti Sharma, Anjum Khan, and Abhiroop Sen from Choithram Netralaya, India, for their support towards the project.

Institutional Review Board Statement

The present study was approved by the Health and Disability Ethics Committee (HDEC), New Zealand, for the research conducted at Dargaville Medical Centre, New Zealand (21/NTB/213) and by the lntegrity Ethics Committee (the institutional Ethics Committee) IEC, CARE CHL-Hospitals, Unit of Convenient Hospitals Limited, India, for the research conducted at Choithram Netralaya, India (EC/JAN/21/3).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study. Written informed consent has been obtained from the patient(s) to publish this paper.

Data Availability Statement

The datasets used and/or analyzed during the current study are available from the corresponding author upon reasonable request and upon the permission from the concerned authorities involved.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. GBD 2019 Blindness and Vision Impairment Collaborators; Vision Loss Expert Group of the Global Burden of Disease Study. Causes of blindness and vision impairment in 2020 and trends over 30 years, and prevalence of avoidable blindness in relation to VISION 2020: The Right to Sight: An analysis for the Global Burden of Disease Study. Lancet Glob. Health 2021, 9, e144–e160. [Google Scholar] [CrossRef] [PubMed]
  2. Holden, B.A.; Fricke, T.R.; Wilson, D.A.; Jong, M.; Naidoo, K.S.; Sankaridurg, P.; Wong, T.Y.; Naduvilath, T.J.; Resnikoff, S. Global prevalence of myopia and high myopia and temporal trends from 2000 through 2050. Ophthalmology 2016, 123, 1036–1042. [Google Scholar] [CrossRef] [PubMed]
  3. Pizzarello, L.; Abiose, A.; Ffytche, T.; Duerksen, R.; Thulasiraj, R.; Taylor, H.; Faal, H.; Rao, G.; Kocur, I.; Resnikoff, S. VISION 2020: The Right to Sight: A global initiative to eliminate avoidable blindness. Arch. Ophthalmol. 2004, 122, 615–620. [Google Scholar] [CrossRef]
  4. Haarman, A.E.; Enthoven, C.A.; Tideman, J.W.L.; Tedja, M.S.; Verhoeven, V.J.; Klaver, C.C. The complications of myopia: A review and meta-analysis. Investig. Ophthalmol. Vis. Sci. 2020, 61, 49. [Google Scholar] [CrossRef] [PubMed]
  5. Verhoeven, V.J.; Wong, K.T.; Buitendijk, G.H.; Hofman, A.; Vingerling, J.R.; Klaver, C.C. Visual consequences of refractive errors in the general population. Ophthalmology 2015, 122, 101–109. [Google Scholar] [CrossRef] [PubMed]
  6. Fricke, T.R.; Jong, M.; Naidoo, K.S.; Sankaridurg, P.; Naduvilath, T.J.; Ho, S.M.; Wong, T.Y.; Resnikoff, S. Global prevalence of visual impairment associated with myopic macular degeneration and temporal trends from 2000 through 2050: Systematic review, meta-analysis and modelling. Br. J. Ophthalmol. 2018, 102, 855–862. [Google Scholar] [CrossRef]
  7. Flitcroft, D. The complex interactions of retinal, optical and environmental factors in myopia aetiology. Prog. Retin. Eye Res. 2012, 31, 622–660. [Google Scholar] [CrossRef] [PubMed]
  8. Ha, A.; Kim, C.Y.; Shim, S.R.; Chang, I.B.; Kim, Y.K. Degree of myopia and glaucoma risk: A dose-response meta-analysis. Am. J. Ophthalmol. 2022, 236, 107–119. [Google Scholar] [CrossRef]
  9. Fricke, T.; Holden, B.; Wilson, D.; Schlenther, G.; Naidoo, K.; Resnikoff, S.; Frick, K. Global cost of correcting vision impairment from uncorrected refractive error. Bull. World Health Organ. 2012, 90, 728–738. [Google Scholar] [CrossRef] [PubMed]
  10. Yekta, A.; Hooshmand, E.; Saatchi, M.; Ostadimoghaddam, H.; Asharlous, A.; Taheri, A.; Khabazkhoob, M. Global prevalence and causes of visual impairment and blindness in children: A systematic review and meta-analysis. J. Curr. Ophthalmol. 2022, 34, 1–15. [Google Scholar] [PubMed]
  11. Chua, S.Y.; Sabanayagam, C.; Cheung, Y.B.; Chia, A.; Valenzuela, R.K.; Tan, D.; Wong, T.Y.; Cheng, C.Y.; Saw, S.M. Age of onset of myopia predicts risk of high myopia in later childhood in myopic Singapore children. Ophthalmic Physiol. Opt. 2016, 36, 388–394. [Google Scholar] [CrossRef] [PubMed]
  12. Wong, K.; Dahlmann-Noor, A. Myopia and its progression in children in London, UK: A retrospective evaluation. J. Optom. 2020, 13, 146–154. [Google Scholar] [CrossRef] [PubMed]
  13. French, A.N.; Murphy, E.; Martin, F.; de Mello, N.; Rose, K.A. Vision Screening in Children: The New South Wales Statewide Eyesight Preschooler Screening Program. Asia-Pac. J. Ophthalmol. 2022, 11, 425–433. [Google Scholar] [CrossRef]
  14. Langeslag-Smith, M.A.; Vandal, A.C.; Briane, V.; Thompson, B.; Anstice, N.S. Preschool children’s vision screening in New Zealand: A retrospective evaluation of referral accuracy. BMJ Open 2015, 5, e009207. [Google Scholar] [CrossRef]
  15. Vision in Preschoolers Study Group. Sensitivity of screening tests for detecting vision in preschoolers-targeted vision disorders when specificity is 94%. Optom. Vis. Sci. 2005, 82, 432–438. [Google Scholar] [CrossRef] [PubMed]
  16. Chun, J.; Kim, Y.; Shin, K.Y.; Han, S.H.; Oh, S.Y.; Chung, T.Y.; Park, K.A.; Lim, D.H. Deep learning–based prediction of refractive error using photorefraction images captured by a smartphone: Model development and validation study. JMIR Med. Inform. 2020, 8, e16225. [Google Scholar] [CrossRef]
  17. Covenant, A.C.; Circumcision, R. The red reflex examination in neonates: An efficient tool for early diagnosis of congenital ocular diseases. Isr. Med. Assoc. J. 2010, 12, 259–261. [Google Scholar]
  18. Tongue, A.C. Refractive errors in children. Pediatr. Clin. N. Am. 1987, 34, 1425–1437. [Google Scholar] [CrossRef]
  19. Toli, A.; Perente, A.; Labiris, G. Evaluation of the red reflex: An overview for the pediatrician. World J. Methodol. 2021, 11, 263. [Google Scholar] [CrossRef] [PubMed]
  20. Paysse, E.A.; Williams, G.C.; Coats, D.K.; Williams, E.A. Detection of red reflex asymmetry by pediatric residents using the Bruckner reflex versus the MTI photoscreener. Pediatrics 2001, 108, e74. [Google Scholar] [CrossRef]
  21. Jain, P.; Kothari, M.T.; Gode, V. The opportunistic screening of refractive errors in school-going children by pediatrician using enhanced Brückner test. Indian J. Ophthalmol. 2016, 64, 733. [Google Scholar] [PubMed]
  22. Kothari, M.T. Can the Brückner test be used as a rapid screening test to detect significant refractive errors in children? Indian J. Ophthalmol. 2007, 55, 213–215. [Google Scholar] [CrossRef] [PubMed]
  23. Bani, S.A.; Amitava, A.K.; Sharma, R.; Danish, A. Beyond photography: Evaluation of the consumer digital camera to identify strabismus and anisometropia by analyzing the Bruckner’s reflex. Indian J. Ophthalmol. 2013, 61, 608. [Google Scholar] [PubMed]
  24. Srivastava, R.M.; Verma, S.; Gupta, S.; Kaur, A.; Awasthi, S.; Agrawal, S. Reliability of Smart Phone Photographs for School Eye Screening. Children 2022, 9, 1519. [Google Scholar] [CrossRef] [PubMed]
  25. Simons, B.D.; Siatkowski, R.M.; Schiffman, J.C.; Berry, B.E.; Flynn, J.T. Pediatric photoscreening for strabismus and refractive errors in a high-risk population. Ophthalmology 1999, 106, 1073–1080. [Google Scholar] [CrossRef] [PubMed]
  26. Ma, S.; Guan, Y.; Yuan, Y.; Tai, Y.; Wang, T. A one-step, streamlined children’s vision screening solution based on smartphone imaging for resource-limited areas: Design and preliminary field evaluation. JMIR MHealth UHealth 2020, 8, e18226. [Google Scholar] [CrossRef]
  27. Kothari, M.T.; Turakhia, J.K.; Vijayalakshmi, P.; Karthika, A.; Nirmalan, P.K. Can the Brückner Test Be Used as a Rapid Screening Test to Detect Amblyogenic Factors in Developing Countries? Am. Orthopt. J. 2003, 53, 121–126. [Google Scholar] [CrossRef]
  28. Molteno, A.; Hoare-Nairne, J.; Parr, J.; Simpson, A.; Hodgkinson, I.; O’Brien, N.; Watts, S. The Otago photoscreener, a method for the mass screening of infants to detect squint and refractive errors. Trans. Ophthalmol. Soc. N. Z. 1983, 35, 43–49. [Google Scholar] [PubMed]
  29. Peterseim, M.M.W.; Trivedi, R.H.; Monahan, S.R.; Smith, S.M.; Bowsher, J.D.; Alex, A.; Wilson, M.E.; Wolf, B.J. Effectiveness of the Spot Vision Screener using updated 2021 AAPOS guidelines. J. Am. Assoc. Pediatr. Ophthalmol. Strabismus 2023, 27, 24.e1–24.e7. [Google Scholar] [CrossRef] [PubMed]
  30. Arnold, R.W.; O’Neil, J.W.; Cooper, K.L.; Silbert, D.I.; Donahue, S.P. Evaluation of a smartphone photoscreening app to detect refractive amblyopia risk factors in children aged 1–6 years. Clin. Ophthalmol. 2018, 12, 1533–1537. [Google Scholar] [CrossRef]
  31. Zhang, X.; Wang, J.; Li, Y.; Jiang, B. Diagnostic test accuracy of Spot and Plusoptix photoscreeners in detecting amblyogenic risk factors in children: A systemic review and meta-analysis. Ophthalmic Physiol. Opt. 2019, 39, 260–271. [Google Scholar] [CrossRef]
  32. Redd, T.K.; Campbell, J.P.; Chiang, M.F. Artificial intelligence for refractive surgery screening: Finding the balance between myopia and hype-ropia. JAMA Ophthalmol. 2020, 138, 526–527. [Google Scholar] [CrossRef] [PubMed]
  33. Foo, L.L.; Ng, W.Y.; San Lim, G.Y.; Tan, T.E.; Ang, M.; Ting, D.S.W. Artificial intelligence in myopia: Current and future trends. Curr. Opin. Ophthalmol. 2021, 32, 413–424. [Google Scholar] [CrossRef] [PubMed]
  34. Ng, W.Y.; Zhang, S.; Wang, Z.; Ong, C.J.T.; Gunasekeran, D.V.; Lim, G.Y.S.; Zheng, F.; Tan, S.C.Y.; Tan, G.S.W.; Rim, T.H.; et al. Updates in deep learning research in ophthalmology. Clin. Sci. 2021, 135, 2357–2376. [Google Scholar] [CrossRef] [PubMed]
  35. Yang, J.; Wu, S.; Zhang, C.; Yu, W.; Dai, R.; Chen, Y. Global trends and frontiers of research on pathologic myopia since the millennium: A bibliometric analysis. Front. Public Health 2022, 10, 1047787. [Google Scholar] [CrossRef]
  36. Lim, J.S.; Hong, M.; Lam, W.S.; Zhang, Z.; Teo, Z.L.; Liu, Y.; Ng, W.Y.; Foo, L.L.; Ting, D.S. Novel technical and privacy-preserving technology for artificial intelligence in ophthalmology. Curr. Opin. Ophthalmol. 2022, 33, 174–187. [Google Scholar] [CrossRef]
  37. Varadarajan, A.V.; Poplin, R.; Blumer, K.; Angermueller, C.; Ledsam, J.; Chopra, R.; Keane, P.A.; Corrado, G.S.; Peng, L.; Webster, D.R. Deep learning for predicting refractive error from retinal fundus images. Investig. Ophthalmol. Vis. Sci. 2018, 59, 2861–2868. [Google Scholar] [CrossRef]
  38. Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
  39. Abràmoff, M.D.; Lou, Y.; Erginay, A.; Clarida, W.; Amelon, R.; Folk, J.C.; Niemeijer, M. Improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning. Investig. Ophthalmol. Vis. Sci. 2016, 57, 5200–5206. [Google Scholar] [CrossRef]
  40. Chalakkal, R.J.; Abdulla, W.H.; Hong, S.C. Fundus retinal image analyses for screening and diagnosing diabetic retinopathy, macular edema, and glaucoma disorders. In Diabetes and Fundus OCT; Elsevier: Amsterdam, The Netherlands, 2020; pp. 59–111. [Google Scholar]
  41. Li, F.; Wang, Z.; Qu, G.; Song, D.; Yuan, Y.; Xu, Y.; Gao, K.; Luo, G.; Xiao, Z.; Lam, D.S.; et al. Automatic differentiation of Glaucoma visual field from non-glaucoma visual filed using deep convolutional neural network. BMC Med. Imaging 2018, 18, 35. [Google Scholar] [CrossRef]
  42. Ting, D.S.W.; Cheung, C.Y.L.; Lim, G.; Tan, G.S.W.; Quang, N.D.; Gan, A.; Hamzah, H.; Garcia-Franco, R.; San Yeo, I.Y.; Lee, S.Y.; et al. Development and validation of a deep learning system for diabetic retinopathy and related eye diseases using retinal images from multiethnic populations with diabetes. JAMA 2017, 318, 2211–2223. [Google Scholar] [CrossRef]
  43. Brown, J.M.; Campbell, J.P.; Beers, A.; Chang, K.; Ostmo, S.; Chan, R.P.; Dy, J.; Erdogmus, D.; Ioannidis, S.; Kalpathy-Cramer, J.; et al. Automated diagnosis of plus disease in retinopathy of prematurity using deep convolutional neural networks. JAMA Ophthalmol. 2018, 136, 803–810. [Google Scholar] [CrossRef]
  44. Lee, C.S.; Tyring, A.J.; Deruyter, N.P.; Wu, Y.; Rokem, A.; Lee, A.Y. Deep-learning based, automated segmentation of macular edema in optical coherence tomography. Biomed. Opt. Express 2017, 8, 3440–3448. [Google Scholar] [CrossRef] [PubMed]
  45. Tan, T.E.; Anees, A.; Chen, C.; Li, S.; Xu, X.; Li, Z.; Xiao, Z.; Yang, Y.; Lei, X.; Ang, M.; et al. Retinal photograph-based deep learning algorithms for myopia and a blockchain platform to facilitate artificial intelligence medical research: A retrospective multicohort study. Lancet Digit. Health 2021, 3, e317–e329. [Google Scholar] [CrossRef] [PubMed]
  46. Burlina, P.M.; Joshi, N.; Pekala, M.; Pacheco, K.D.; Freund, D.E.; Bressler, N.M. Automated grading of age-related macular degeneration from color fundus images using deep convolutional neural networks. JAMA Ophthalmol. 2017, 135, 1170–1176. [Google Scholar] [CrossRef] [PubMed]
  47. Grassmann, F.; Mengelkamp, J.; Brandl, C.; Harsch, S.; Zimmermann, M.E.; Linkohr, B.; Peters, A.; Heid, I.M.; Palm, C.; Weber, B.H. A deep learning algorithm for prediction of age-related eye disease study severity scale for age-related macular degeneration from color fundus photography. Ophthalmology 2018, 125, 1410–1420. [Google Scholar] [CrossRef]
  48. Son, J.; Shin, J.Y.; Chun, E.J.; Jung, K.H.; Park, K.H.; Park, S.J. Predicting high coronary artery calcium score from retinal fundus images with deep learning algorithms. Transl. Vis. Sci. Technol. 2020, 9, 28. [Google Scholar] [CrossRef]
  49. Munjral, S.; Maindarkar, M.; Ahluwalia, P.; Puvvula, A.; Jamthikar, A.; Jujaray, T.; Suri, N.; Paul, S.; Pathak, R.; Saba, L.; et al. Cardiovascular risk stratification in diabetic retinopathy via atherosclerotic pathway in COVID-19/non-COVID-19 frameworks using artificial intelligence paradigm: A narrative review. Diagnostics 2022, 12, 1234. [Google Scholar] [CrossRef] [PubMed]
  50. Hafiz, F.; Chalakkal, R.J.; Hong, S.C.; Linde, G.; Hu, R.; O’Keeffe, B.; Boobin, Y. A new approach to non-mydriatic portable fundus imaging. Expert Rev. Med. Devices 2022, 19, 303–314. [Google Scholar] [CrossRef]
  51. MedicMind, Dunedin, Otago, 9013, NZ. 2022. Available online: https://www.medicmind.tech/ (accessed on 1 January 2017).
  52. Lee, M.; Bin Mahmood, A.B.S.; Lee, E.S.; Smith, H.E.; Tudor Car, L. Smartphone and Mobile App Use Among Physicians in Clinical Practice: Scoping Review. JMIR MHealth UHealth 2023, 11, e44765. [Google Scholar] [CrossRef]
  53. Gupta, R.; Agrawal, S.; Srivastava, R.M.; Singh, V.; Katiyar, V. Smartphone photography for screening amblyogenic conditions in children. Indian J. Ophthalmol. 2019, 67, 1560. [Google Scholar] [PubMed]
  54. Peterseim, M.M.W.; Rhodes, R.S.; Patel, R.N.; Wilson, M.E.; Edmondson, L.E.; Logan, S.A.; Cheeseman, E.W.; Shortridge, E.; Trivedi, R.H. Effectiveness of the GoCheck Kids vision screener in detecting amblyopia risk factors. Am. J. Ophthalmol. 2018, 187, 87–91. [Google Scholar] [CrossRef] [PubMed]
  55. Yoo, T.K.; Ryu, I.H.; Kim, J.K.; Lee, I.S. Deep learning for predicting uncorrected refractive error using posterior segment optical coherence tomography images. Eye 2022, 36, 1959–1965. [Google Scholar] [CrossRef] [PubMed]
Figure 1. IR and colour red reflex image capture using nun IR.
Figure 1. IR and colour red reflex image capture using nun IR.
Diagnostics 13 02810 g001
Figure 2. Images of pupils cropped from colour and IR images. Column 1 is the original image. Column 2 is cropped version.
Figure 2. Images of pupils cropped from colour and IR images. Column 1 is the original image. Column 2 is cropped version.
Diagnostics 13 02810 g002
Figure 3. nun IR illumination path [50].
Figure 3. nun IR illumination path [50].
Diagnostics 13 02810 g003
Figure 4. Red reflex color images taken by nun IR using a fake eye, for which spherical power can be adjusted.
Figure 4. Red reflex color images taken by nun IR using a fake eye, for which spherical power can be adjusted.
Diagnostics 13 02810 g004
Figure 5. IR images with corresponding crescent types.
Figure 5. IR images with corresponding crescent types.
Diagnostics 13 02810 g005
Figure 6. Pupils cropped automatically using MedicMind-AI portal.
Figure 6. Pupils cropped automatically using MedicMind-AI portal.
Diagnostics 13 02810 g006
Figure 7. Inception-v3 architecture.
Figure 7. Inception-v3 architecture.
Diagnostics 13 02810 g007
Figure 8. EfficientNet architecture.
Figure 8. EfficientNet architecture.
Diagnostics 13 02810 g008
Figure 9. Spherical power distribution in the Dargaville dataset.
Figure 9. Spherical power distribution in the Dargaville dataset.
Diagnostics 13 02810 g009
Figure 10. Crescent type categories (AD).
Figure 10. Crescent type categories (AD).
Diagnostics 13 02810 g010
Figure 11. Sphere vs. crescent type in the Choithram dataset.
Figure 11. Sphere vs. crescent type in the Choithram dataset.
Diagnostics 13 02810 g011
Figure 12. Cylinder vs. crescent type.
Figure 12. Cylinder vs. crescent type.
Diagnostics 13 02810 g012
Figure 13. Distribution of spherical power for 288 IR red reflex images.
Figure 13. Distribution of spherical power for 288 IR red reflex images.
Diagnostics 13 02810 g013
Figure 14. Distribution after removing median 30% of IR images.
Figure 14. Distribution after removing median 30% of IR images.
Diagnostics 13 02810 g014
Figure 15. IR image before (left) and after (right) increasing contrast.
Figure 15. IR image before (left) and after (right) increasing contrast.
Diagnostics 13 02810 g015
Figure 16. Combined color and IR images.
Figure 16. Combined color and IR images.
Diagnostics 13 02810 g016
Figure 17. Myopic IR images.
Figure 17. Myopic IR images.
Diagnostics 13 02810 g017
Figure 18. Normal (left image) and myopic (second and third image) IR images.
Figure 18. Normal (left image) and myopic (second and third image) IR images.
Diagnostics 13 02810 g018
Figure 19. Accuracy of techniques for IR images using EfficientNet (≥70% dark, ≥60% medium otherwise light).
Figure 19. Accuracy of techniques for IR images using EfficientNet (≥70% dark, ≥60% medium otherwise light).
Diagnostics 13 02810 g019
Table 1. Performance of different architectures.
Table 1. Performance of different architectures.
IR Inception-v3IR EfficientNetColour Inception-v3Colour EfficientNet
Sphere55%63%54%64%
Cylinder49%70%50%57%
Table 2. Performance of the multiclass classifier.
Table 2. Performance of the multiclass classifier.
IR Inception-v3IR EfficientNetColour Inception-v3Colour EfficientNet
Sphere55%70%50%64%
Cylinder48%55%56%60%
Table 3. Accuracy when removing median 20% of images.
Table 3. Accuracy when removing median 20% of images.
IR Inception-v3IR EfficientNetColour Inception-v3Colour EfficientNet
Sphere75%74%63%68%
Cylinder49%49%49%60%
Table 4. Effect of only training with red reflex images greater than 180 pixels in width for IR images and greater than 450 pixels in width for colour images.
Table 4. Effect of only training with red reflex images greater than 180 pixels in width for IR images and greater than 450 pixels in width for colour images.
IR EfficientNetColour EfficientNet
20% spherical power median removed74%68%
Small-sized images and 20% spherical power median removed64%60%
Table 5. Accuracy when training with combined IR and color images.
Table 5. Accuracy when training with combined IR and color images.
Inception-v3EfficientNet
Sphere74%75%
Cylinder60%56%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Linde, G.; Chalakkal, R.; Zhou, L.; Huang, J.L.; O’Keeffe, B.; Shah, D.; Davidson, S.; Hong, S.C. Automatic Refractive Error Estimation Using Deep Learning-Based Analysis of Red Reflex Images. Diagnostics 2023, 13, 2810. https://doi.org/10.3390/diagnostics13172810

AMA Style

Linde G, Chalakkal R, Zhou L, Huang JL, O’Keeffe B, Shah D, Davidson S, Hong SC. Automatic Refractive Error Estimation Using Deep Learning-Based Analysis of Red Reflex Images. Diagnostics. 2023; 13(17):2810. https://doi.org/10.3390/diagnostics13172810

Chicago/Turabian Style

Linde, Glenn, Renoh Chalakkal, Lydia Zhou, Joanna Lou Huang, Ben O’Keeffe, Dhaivat Shah, Scott Davidson, and Sheng Chiong Hong. 2023. "Automatic Refractive Error Estimation Using Deep Learning-Based Analysis of Red Reflex Images" Diagnostics 13, no. 17: 2810. https://doi.org/10.3390/diagnostics13172810

APA Style

Linde, G., Chalakkal, R., Zhou, L., Huang, J. L., O’Keeffe, B., Shah, D., Davidson, S., & Hong, S. C. (2023). Automatic Refractive Error Estimation Using Deep Learning-Based Analysis of Red Reflex Images. Diagnostics, 13(17), 2810. https://doi.org/10.3390/diagnostics13172810

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop