1. Introduction
Accurately diagnosing ocular disorders is essential in ophthalmology. Advances in AI and mobile health are transforming diagnostic methods, addressing the rise in visual disorders linked to increased time and smartphone use [
1,
2,
3]. New self-screening tools, including visual acuity tests, automated strabismus detection, and virtual reality (VR)-based assessments, are driving these advancements [
4,
5,
6,
7].
Tele-ophthalmology enables remote screening and consultation, improving diagnostic accuracy for conditions like diabetic retinopathy, age-related macular degeneration, and glaucoma while enhancing accessibility to care [
8,
9,
10].
1.1. Convergence Insufficiency (CI)
Convergence is crucial for binocular vision and depth perception. Convergence Insufficiency (CI) affects near vision, causing eye strain, headaches, diplopia, difficulty reading, and blurred vision [
11,
12]. Its prevalence varies from 1.35% to 33%, depending on measurement methods [
13,
14]. Among fifth and sixth graders, prevalence reaches 13% [
15], while in adults and military personnel with mild traumatic brain injuries, it ranges from 40% to 46% [
16].
1.2. Measuring near Point of Convergence (NPC) and Convergence Insufficiency (CI) Diagnosis
Traditional NPC tests, such as the RAF Ruler test (RulerT) [
17] and Pencil test (PencilT) [
18], involve moving a target closer to the patient and recording when fixation is lost [
19]. These methods are effective but require professional administration, rely on subjective patient feedback, and may be inaccessible in remote areas, leading to undiagnosed CI cases.
This study introduces MobileS, an AI-based system for automatic NPC detection and CI diagnosis using a smartphone’s front camera and the MediaPipe framework (
Figure 1). MobileS eliminates the need for specialized equipment and professional supervision, making it accessible for patients with communication difficulties and enhancing tele-ophthalmology practices.
1.3. Purpose
This study compares NPC measurements from MobileS to those from RulerT and PencilT to assess its effectiveness in CI detection. RulerT remains the gold standard in CI diagnosis.
2. Materials and Methods
2.1. Measurements by Novel Algorithm
Our smartphone application integrates an AI algorithm to detect the NPC using the front camera. The algorithm extracts key features, including pupillary distance (PD), eye-phone distance, and eye movement relative to the bridge of the nose. These metrics are processed frame by frame, applying a mathematical technique known as Gaussian Smoothing to reduce noise while preserving data trends [
21]. The final output includes the NPC, showcasing the algorithm’s efficacy in NPC detection and CI diagnosis. The MediaPipe system is designed to function across a wide range of smartphone models and camera qualities, ensuring consistent performance on different devices.
2.1.1. Step 1—Measurements of Static Eye Features and Distances
The algorithm uses the MediaPipe Eye Tracking Model, which employs an RGB camera to generate a 3D mesh of the face, focusing on the eye region [
22,
23] (
Figure 1A). It detects 478 key points and processes them through a neural network to determine their 3D coordinates [
24].
Key static parameters include pupillary distance (PD) (A in
Figure 1C), white-to-white (WTW) measurements (B and C in
Figure 1C), and distances from the eyes to the smartphone. The system tracks eye movements relative to each other and to the nose bridge (
Figure 1C; the distance between B and D, and the distance between C and D).
The iris's diameter (WTW) is assumed constant at approximately 11.7 ± 0.5 mm [
25,
26,
27], ensuring consistency in calculations.
The formula for PD (mm) is as follows:
where PD denotes the pupillary distance, representing the distance between the centre of the eyes in millimetres. RightEyeCenter.X refers to the X component of the landmark representing the centre of the right iris in pixels, LeftEyeCenter.X represents the X component of the landmark denoting the centre of the left iris in pixels, and IrisDiameter is the diameter of the iris in pixels. The distance from the pupils to the bridge of the nose can be calculated using the same method as in Equation (1).
The smartphone-to-eye distance (
Figure 1B) is calculated using
where DistanceFromCamera (d) signifies the distance from the iris to the smartphone camera in millimeters, FocalLength (f) denotes the focal length of the smartphone camera in pixels, and IrisDiameter is the diameter of the iris in pixels.
As shown in Equation (2), this method, tested on various smartphone models, has shown an error margin of less than 10% [
20].
2.1.2. Step 2—Measurement of Convergence
NPC is determined by analyzing changes in PD and smartphone distance. Normally, both decrease smoothly until fixation is lost (
Figure 2A). If the PD stops decreasing or increases while the phone continues moving closer, the system detects a Break Point (
Figure 2B). A threshold-based evaluation determines CI.
The ExoCounter, a novel feature, identifies simultaneous outward eye deviations linked to blurred vision [
28]. By integrating ExoCounter with PD shifts (
Figure 2C), the system enhances NPC detection and provides valuable insights into the patient’s convergence potential and mechanism.
2.2. Step 2—Clinical Trial
The clinical trial involves the validation of our MobileS compared to the established standard methods for measuring NPC and diagnosing CI and checking the reliability and accuracy of our algorithm.
2.2.1. Subjects and Facility
A prospective controlled study was conducted in accordance with the Helsinki Declaration and approved by the Ethics Committee.
Inclusion criteria: age 18 years or more who can sign the consent form, with best corrected visual acuity of 6/12 or more in each eye. Exclusion criteria: patients with a history of intraocular or strabismus surgery.
Participants used prescribed optical correction for accurate NPC measurements. Additionally, the MediaPipe framework used in our system supports eye tracking both with and without glasses, ensuring measurement accuracy.
2.2.2. Procedure
The NPC of each participant was assessed using three methods in a specific sequence: PencilT, RulerT conducted by a single researcher, and MobileS performed independently by the patient following a brief explanation. Between each test, participants were given a 3 min break, during which they were instructed to focus on a distant object. Each test was performed twice. Participants ranked usability post-assessment. The clinical trial involved validating the MobileS system by statistically comparing its NPC measurements to those obtained using the gold standard method (RulerT) for the same participants. This analysis assessed the reliability and accuracy of our algorithm. To ensure standardized execution, including movement speed, all NPC measurements in this study were conducted under professional supervision.
The RAF Ruler test (RulerT): An ophthalmologist moves a target closer until fixation is lost (due to blurred or double vision).
The Pencil test (PencilT): A manually held pencil replaces the ruler.
The Automatic System Test (MobileS): Participants follow a target on the smartphone until it touches their nose (
Figure 3).
Historically and in conventional practice, an NPC of less than 10 cm (4 inches) is considered normal, as young people can maintain fusion at close distances [
29,
30]. If the NPC exceeds 10 cm, professional evaluation for CI should be conducted.
A population-based study found that the mean NPC was 8.59 ± 4.82 cm, with younger individuals (10–19 years) averaging 6.95 ± 3.87 cm. The NPC increased with age, reaching 13.06 ± 5.2 cm in those over 70 years [
31].
In this study, we set 13.41 cm (8.59 + 4.82 cm [
31]) as the threshold for CI while also considering the widely accepted 10 cm cutoff.
2.3. Statistical Analysis
The sample size (n = 86) justified the use of Pearson’s correlation coefficients to compare the measurements from our system and the traditional methods. Intraclass correlation coefficients (ICCs) were also calculated to evaluate the agreement between these methods. An ICC value close to 1 would indicate strong agreement.
3. Results
This study included 86 participants, consisting of 43 men and 43 women, with a mean age of 32 ± 11 years (ranging from 18 to 59 years). For calculating specificity and sensitivity, the RulerT was used as the gold standard.
Comparison of NPC Measurements: The mean NPC values measured by the PencilT, RulerT, and our MobileS were 7.57 ± 4.49 cm, 9.87 ± 3.04 cm, and 9.69 ± 3.59 cm, respectively.
Statistical Correlations: The statistical analysis revealed significant correlations between the methods. The Pearson correlation coefficient between RulerT and MobileS was 0.74, and between RulerT and PencilT was 0.81 (all
p values < 0.001), suggesting a strong positive correlation in both comparisons (
Table 1). The intraclass correlation coefficient (ICC) between RulerT and MobileS was 0.73, and between RulerT and PencilT was 0.70 (all
p values < 0.001), confirming a good level of agreement (
Table 2).
Diagnostic Capability for CI: When the diagnosis of CI is based on an NPC threshold of >10 cm, the MobileS could identify CI in 26 out of 86 participants (30.23%), demonstrating a sensitivity of 76% and a specificity of 85% (
Figure 4A). When CI was based on NPC > 13.41 cm, CI was found in nine patients (10.46%), corresponding to a sensitivity of 64% and a specificity of 94% (
Figure 4B). A percentage of 83.72% of the NPC values obtained through the MobileS were within ±3 cm compared to the NPC values acquired by the RulerT (
Figure 5).
Participant Preference: In terms of user experience, approximately 71% of participants favoured the MobileS, 21% preferred the RulerT, and 8% chose the PencilT. This indicates the acceptability and ease of use of our system in a clinical setting.
4. Discussion
Our study has successfully demonstrated the feasibility and reliability of the innovative MobileS for measuring the NPC and diagnosing CI. Additionally, it introduced a novel distinct feature (ExoCounter) alongside the conventional PD, allowing for the determination of NPC without supplementary instrumentation. Furthermore, the system facilitates self-testing by the patient. MobileS demonstrates high correlation and agreement with the RulerT and has received positive feedback from users for its ease of use. Notably, the questionnaire completed by participants at the end of the test revealed that approximately 71% of respondents preferred the MobileS system over the other two methods. We expect this percentage to increase as the prototype evolves with further development and as users gain more experience with home-based testing.
4.1. Unique Aspects and Clinical Significance of MobileS
MobileS uses AI and a smartphone camera for remote NPC assessment, eliminating the need for specialized equipment or in-clinic supervision. This enhances accessibility and allows for broader population screening. The system works with any cellular or Wi-Fi network, making it effective even in remote areas where VR-based CI tools may be impractical. Additionally, the growing integration of VR and LTE-enabled devices opens future opportunities for compatibility with CI treatment platforms [
32,
33]. MobileS also provides immediate diagnostic feedback, enabling users to track progress and share results with healthcare providers—an important advantage in tele-ophthalmology and rehabilitation.
This study has some limitations. In unsupervised settings, the application currently lacks real-time monitoring to correct potential testing errors, such as inappropriate testing speed, gaze deviation, or poor lighting conditions. Additionally, children, a key clinical target group, were not included in this study. Lastly, while the sample size is reasonable, further expansion would enhance generalizability and statistical robustness. Future improvements may involve algorithmic modifications to address occasional inaccuracies, such as test performance or gaze deviations.
Regarding sensitivity, it is important to note that our system is compared to the RulerT, which is the gold standard with 100% sensitivity. Therefore, our objective is to achieve sensitivity as close as possible to that of the RulerT. When compared to the PencilT, MobileS exhibits higher sensitivity. Specifically, at a threshold of 13.41 cm, MobileS achieves 64% sensitivity compared to 36% for the Pencil Test. At a threshold of 10 cm, both methods show 76% sensitivity. Thus, MobileS demonstrates superior sensitivity as a diagnostic tool compared to the Pencil Test.
We initially established normative data thresholds based on literature but have collected demographic data (such as age and gender) to refine these thresholds if needed. Future improvements may involve adjusting algorithms to address occasional inaccuracies, such as variations in test performance or gaze deviations.
While this study focuses on diagnosing CI using MobileS, it does not include post-treatment data. However, a follow-up study is underway to compare app-based convergence exercises with traditional pencil push-ups, aiming to assess clinical benefits. Although MobileS provides an accessible screening tool, clinical referral remains essential for confirmation and treatment management. The system is designed to recommend follow-up evaluations for those requiring further assessment.
4.2. Future Work
We continue to refine NPC detection by identifying true convergence points and distinguishing them from “NPC Mimickers” based on previous research [
34]. Incorporating machine learning, we aim to improve accuracy by analyzing eye behaviour near NPC. While this study does not yet utilize training and test sets, future versions will integrate machine learning to enhance precision.
Planned expansions include detecting binocular vision anomalies such as strabismus and nystagmus, assessing accommodation, and providing vision therapy exercises. These enhancements align with our current technology, which supports both diagnostics and rehabilitation. Key challenges involve adapting the algorithm for diverse eye movement patterns and collecting high-quality datasets for validation.
Currently, MobileS is available for Android, with plans to extend support to iOS.
5. Conclusions
MobileS stands as a transformative tool in ophthalmic diagnostics, heralding the era of telemedicine and home-based care. Its AI-driven approach and smartphone integration provide accessible, efficient, and innovative eye health solutions. Given its accuracy and user-friendly design, MobileS has the potential to expand its role in diagnosing a range of ocular disorders, marking a shift in modern ophthalmology.
6. Patents
This work is protected by U.S. Patent Application CRML-P-047-USP, titled “System and Method for Diagnosis and Treatment of Various Movement Disorders and Diseases of the Eye”. Please note that the patent application is currently pending and has not yet been granted.
Author Contributions
Conceptualization, A.K., S.R., H.N., H.J.-H., and I.S.; methodology, A.K., S.R., H.N., H.J.-H., and I.S.; software, A.K. and I.S.; validation, A.K., S.R., and H.N.; formal analysis, A.K., S.R., and I.S.; investigation, A.K., H.N., and H.J.-H.; resources, A.K., S.R., H.N., H.J.-H., and I.S.; data curation, A.K., H.J.-H., and H.N.; writing—original draft preparation, A.K.; writing—review and editing, A.K., S.R., H.N., H.J.-H., and I.S.; visualization, A.K., S.R., H.J.-H., and I.S.; supervision, I.S.; project administration, A.K., S.R., H.J.-H., and I.S. All authors have read and agreed to the published version of the manuscript.
Funding
This research was funded by the Israel Innovation Authority, grant number 79054.
Institutional Review Board Statement
The study was conducted in accordance with the Declaration of Helsinki (0099-22-POR, Date: 6 December 2022), approved by the Ethics Committee of the Faculty of Social Science at the University of Haifa (087/22, Date: 16 March 2022), and was carried out at the ophthalmology clinic of Tzafon Medical Center.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
The data presented in this study are available on request from the corresponding author due to privacy and ethical restrictions. Specifically, video recordings and patient images cannot be shared. However, other anonymized data may be provided upon reasonable request and subject to ethical approval.
Acknowledgments
We would like to thank Nitza Barkan for her assistance with statistical analysis and Mohammad Mahamid for his contributions to the development of the mobile application.
Conflicts of Interest
The authors declare no conflicts of interest.
Abbreviations
The following abbreviations are used in this manuscript:
AI | Artificial Intelligence |
NPC | Near Point of Convergence |
CI | Convergence Insufficiency |
MobileS | Mobile application System |
PD | Pupillary Distance |
ExoCounter | Exodeviation episode’s Counter |
RulerT | RAF Ruler Test |
ICC | Intraclass Correlation Coefficient |
VR | Virtual Reality |
PencilT | Pencil Test |
WTW | White-to-white |
References
- Khushdeep Kaur, K.K.; Harpinder Kaur, H.K.; Sidhu, M.K. Computer vision syndrome: A major concern for VDT users. Asian J. Home Sci. 2015, 10, 250–253. [Google Scholar] [CrossRef]
- Abuallut, I.; Ajeebi, R.E.; Bahari, A.Y.; Abudeyah, M.A.; Alyamani, A.A.; Zurayyir, A.J.; Alharbi, A.H.; Al Faqih, A.A.; Suwaydi, A.Z.; Alqasemi, M.I.; et al. Prevalence of computer vision syndrome among school-age children during the COVID-19 pandemic, Saudi Arabia: A cross-sectional survey. Children 2022, 9, 1718. [Google Scholar] [CrossRef] [PubMed]
- Song, M.; Li, L.; Guo, J.; Liu, T.; Li, S.; Wang, Y.; Wang, J. A new method for muscular visual fatigue detection using electrooculogram. Biomed. Signal Process. Control 2020, 58, 101865. [Google Scholar] [CrossRef]
- Xian, Y.; Ye, Y.; Liu, F.; Zhou, X.; Zhao, J. Agreement between a mobile applet-based visual acuity test program and the conventional method for distance and near visual acuity tests. Clin. Exp. Ophthalmol. 2023, 51, 112–121. [Google Scholar] [CrossRef]
- Zheng, C.; Yao, Q.; Lu, J.; Xie, X.; Lin, S.; Wang, Z.; Wang, S.; Fan, Z.; Qiao, T. Detection of referable horizontal strabismus in children’s primary gaze photographs using deep learning. Transl. Vis. Sci. Technol. 2021, 10, 33. [Google Scholar] [CrossRef]
- Chen, Z.; Fu, H.; Lo, W.L.; Chi, Z. Strabismus recognition using eye-tracking data and convolutional neural networks. J. Healthc. Eng. 2018, 2018, 814958. [Google Scholar] [CrossRef]
- Kerasidis, H.; Simmons, G.H.; Watkins, C.M. Systems and Methods for Neuro-Ophthalmology Assessments in Virtual Reality. Google Patents. US Patent App. 16/274,233, 15 August 2019. [Google Scholar]
- Caffery, L.J.; Taylor, M.; Gole, G.; Smith, A.C. Models of care in tele-ophthalmology: A scoping review. J. Telemed. Telecare 2019, 25, 106–122. [Google Scholar] [CrossRef]
- Far, P.M.; Tai, F.; Ogunbameru, A.; Pechlivanoglou, P.; Sander, B.; Wong, D.T.; Brent, M.H.; Felfeli, T. Diagnostic accuracy of teleretinal screening for detection of diabetic retinopathy and age-related macular degeneration: A systematic review and meta-analysis. BMJ Open Ophthalmol. 2022, 7, e000915. [Google Scholar]
- Gulshan, V.; Peng, L.; Coram, M.; Stumpe, M.C.; Wu, D.; Narayanaswamy, A.; Venugopalan, S.; Widner, K.; Madams, T.; Cuadros, J.; et al. Development and validation of a deep learning algorithm for detection of diabetic retinopathy in retinal fundus photographs. JAMA 2016, 316, 2402–2410. [Google Scholar] [CrossRef]
- Eyewiki. Convergence Insufficiency. American Academy of Ophthalmology. November 2019. Available online: http://eyewiki.aao.org/Convergence_Insufficiency (accessed on 21 December 2024).
- Simon, J.W.; Buckley, E.G.; Drack, A.V.; Hutchinson, A.K.; Plager, D.A.; Raab, E.L. Basic and Clinical Science Course. Section 6: Pediatric Ophthalmology and Strabismus, 2005–2006; American Academy of Ophthalmology: San Francisco, CA, USA, 2006. [Google Scholar]
- Gantz, L.; Stiebel-Kalish, H. Convergence insufficiency: Review of clinical diagnostic signs. J. Optom. 2022, 15, 256–270. [Google Scholar] [CrossRef]
- Nunes, A.F.; Monteiro, P.M.; Ferreira, F.B.; Nunes, A.S. Convergence insufficiency and accommodative insufficiency in children. BMC Ophthalmol. 2019, 19, 1–8. [Google Scholar] [CrossRef] [PubMed]
- Rouse, M.W.; Borsting, E.; Hyman, L.; Hussein, M.; Cotter, S.A.; Flynn, M.; Scheiman, M.; Gallaway, M.; de Land, P.N. Frequency of convergence insufficiency among fifth and sixth graders. Optom. Vis. Sci. 1999, 76, 643–649. [Google Scholar] [CrossRef] [PubMed]
- Mishra, S.; Kim, Y.S.; Intarasirisawat, J.; Kwon, Y.T.; Lee, Y.; Mahmood, M.; Lim, H.R.; Herbert, R.; Yu, K.J.; Ang, C.S.; et al. Soft, wireless periocular wearable electronics for real-time detection of eye vergence in a virtual reality toward mobile eye therapies. Sci. Adv. 2020, 6, eaay1729. [Google Scholar] [CrossRef] [PubMed]
- Neely, J. The RAF near-point rule. Br. J. Ophthalmol. 1956, 40, 636. [Google Scholar] [CrossRef]
- Siderov, J.; Chiu, S.C.; Waugh, S.J. Differences in the near point of convergence with target type. Ophthalmic Physiol. Opt. 2001, 21, 356–360. [Google Scholar] [CrossRef]
- Eyewiki. Convergence Ability. American Academy of Ophthalmology. September 2019. Available online: http://eyewiki.aao.org/Convergence_ability (accessed on 21 December 2024).
- Vakunov, A.; Lagun, D. MediaPipe Iris: Real-Time Iris Tracking and Depth Estimation. Google AI Blog. 2020. Available online: https://research.google/blog/mediapipe-iris-real-time-iris-tracking-depth-estimation/ (accessed on 21 December 2024).
- Babaud, J.; Witkin, A.P.; Baudin, M.; Duda, R.O. Uniqueness of the Gaussian kernel for scale-space filtering. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 8, 26–33. [Google Scholar] [CrossRef]
- Grishchenko, I.; Ablavatski, A.; Kartynnik, Y.; Raveendran, K.; Grundmann, M. Attention Mesh: High-fidelity face mesh prediction in real-time. arXiv 2020, arXiv:2006.10962. [Google Scholar]
- Ablavatski, A.; Vakunov, A.; Grishchenko, I.; Raveendran, K.; Zhdanovich, M. Real-time pupil tracking from monocular video for digital puppetry. arXiv 2020, arXiv:2006.11341. [Google Scholar]
- Kartynnik, Y.; Ablavatski, A.; Grishchenko, I.; Grundmann, M. Real-time facial surface geometry from monocular video on mobile GPUs. arXiv Preprint 2019, arXiv:1907.06724. [Google Scholar]
- Hashemi, H.; Khabazkhoob, M.; Emamian, M.H.; Shariati, M.; Yekta, A.; Fotouhi, A. White-to-white corneal diameter distribution in an adult population. J. Curr. Ophthalmol. 2015, 27, 21–24. [Google Scholar] [CrossRef]
- Bergmanson, J.P.; Martinez, J.G. Size does matter: What is the corneo-limbal diameter? Clin. Exp. Optom. 2017, 100, 522–528. [Google Scholar] [CrossRef] [PubMed]
- Gharaee, H.; Abrishami, M.; Shafiee, M.; Ehsaei, A. White-to-white corneal diameter: Normal values in a healthy Iranian population obtained with the Orbscan II. Int. J. Ophthalmol. 2014, 7, 309. [Google Scholar] [PubMed]
- Wright, K.W.; Spiegel, P.H.; Hengst, T. Pediatric Ophthalmology and Strabismus; Springer Science & Business Media: New York, NY, USA, 2013. [Google Scholar]
- Pickwell, L.; Hampshire, R. The significance of inadequate convergence. Ophthalmic Physiol. Opt. 1981, 1, 13–18. [Google Scholar] [CrossRef]
- Mahto, R. Eye strain from convergence insufficiency. Br. Med. J. 1972, 2, 564–565. [Google Scholar] [CrossRef]
- Ostadimoghaddam, H.; Hashemi, H.; Nabovati, P.; Yekta, A.; Khabazkhoob, M. The distribution of near point of convergence and its association with age, gender, and refractive error: A population-based study. Clin. Exp. Optom. 2017, 100, 255–259. [Google Scholar] [CrossRef]
- Tan, Z.; Li, Y.; Li, Q.; Zhang, Z.; Li, Z.; Lu, S. Supporting mobile VR in LTE networks: How close are we? Proc. ACM Meas. Anal. Comput. Syst. 2018, 2, 1–31. [Google Scholar]
- Hu, F.; Deng, Y.; Saad, W.; Bennis, M.; Aghvami, A.H. Cellular-connected wireless virtual reality: Requirements, challenges, and solutions. IEEE Commun. Mag. 2020, 58, 105–111. [Google Scholar] [CrossRef]
- Arnoldi, K.; Reynolds, J.D. A review of convergence insufficiency: What are we really accomplishing with exercises? Am. Orthopt. J. 2007, 57, 123–130. [Google Scholar] [CrossRef]
| Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).