Next Article in Journal
Extended Reality Applications for CNC Machine Training: A Systematic Review
Previous Article in Journal
VRChances: An Immersive Virtual Reality Experience to Support Teenagers in Their Career Decisions
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Brief Report

Can Generative AI Contribute to Health Literacy? A Study in the Field of Ophthalmology

by
Carlos Ruiz-Núñez
1,†,
Javier Gismero Rodríguez
2,†,
Antonio J. Garcia Ruiz
3,4,
Saturnino Manuel Gismero Moreno
4,
María Sonia Cañizal Santos
5 and
Iván Herrera-Peco
6,7,8,*,†
1
Programa de Doctorado en Biomedicina, Investigación Traslacional y Nuevas Tecnologías en Salud, Facultad de Medicina, Universidad de Málaga, Blvr. Louis Pasteur, 29010 Málaga, Spain
2
HU Virgen de la Victoria, Servicio Andaluz de Salud, Residente Oftalmología, Campus de Teatinos, 29010 Málaga, Spain
3
Instituto de Investigación Biomédica de Málaga, IBIMA—Bionand, Severo Ochoa 35, 29590 Málaga, Spain
4
Departamento de Farmacología y Pediatría, Facultad de Medicina, Universidad de Málaga, Blvr. Louis Pasteur, 29010 Málaga, Spain
5
Servicio Andaluz de Salud, Administración, Calle Severo Ochoa, 29590 Málaga, Spain
6
Facultad de Ciencias de la Salud, Universidad Alfonso X el Sabio, Villanueva de la Cañada, 28691 Madrid, Spain
7
Faculty of Health Sciences-HM Hospitals, University Camilo José Cela, Urb. Villafranca del Castillo, 49. Villanueva de la Cañada, 28692 Madrid, Spain
8
Instituto de Investigación Sanitaria HM Hospitales, 28015 Madrid, Spain
*
Author to whom correspondence should be addressed.
These authors have contributed equally to this work.
Multimodal Technol. Interact. 2024, 8(9), 79; https://doi.org/10.3390/mti8090079
Submission received: 31 July 2024 / Revised: 23 August 2024 / Accepted: 2 September 2024 / Published: 4 September 2024

Abstract

:
ChatGPT, a generative artificial intelligence model, can provide useful and reliable responses in the field of ophthalmology, comparable to those of medical professionals. Twelve frequently asked questions from ophthalmology patients were selected, and responses were generated both in the role of an expert user and a non-expert user. The responses were evaluated by ophthalmologists using three scales: Global Quality Score (GQS), Reliability Score (RS), and Usefulness Score (US), and analyzed statistically through descriptive study, association, and comparison. The results indicate that there are no significant differences between the responses of expert and non-expert users, although the responses from the expert user tend to be slightly better rated. ChatGPT’s responses proved to be reliable and useful, suggesting its potential as a complementary tool to enhance health literacy and alleviate the informational burden on healthcare professionals.

1. Introduction

Artificial intelligence (AI) has demonstrated remarkable potential in the healthcare field since its inception in the 1970s with pioneering projects such as Mycin [1], which used statistical inference rules to emulate the decision-making process of medical experts through predictive reasoning. However, it is in recent years that there has been a surge in innovative solutions in the healthcare sector, with generative artificial intelligence (AI) focused on creating original content being among the most prominent.
It is essential to understand that artificial intelligence is an umbrella term encompassing various primary techniques, the most used being machine learning (ML), deep learning (DL), and natural language processing (NLP).
Initially launched in 2022 with the ChatGPT-3.5 version and updated to ChatGPT-4 in 2023, this generative AI model falls under the category of chatbots, designed to simulate human conversations using natural language processing (NLP) and machine learning (ML) [2], proving to be useful in various fields, including healthcare, thanks to its iterative learning capability and human feedback. Currently, there are several versions available, ranging from a free basic ChatGPT-4 to subscription versions like ChatGPT-4 Pro. To avoid confusion about the different versions, we will refer to ChatGPT-4 whenever discussing this generative artificial intelligence tool.
ChatGPT is characterized by its iterative capability, meaning it can expand its knowledge through reinforced learning from human feedback. The uses of this chatbot are diverse, including generating conversations, personal assistants, search engines, human resource planning, and even writing scientific articles. It has been asserted that there are no significant differences between a cover letter written by a human and one written by ChatGPT [3].
Numerous articles have been written about the ability to simulate human writing and reasoning, driven by concerns that AI might replace humans. These studies conclude that it is difficult to distinguish the origin of a text [4], whether robotic or human, to the extent that its use for writing scientific articles is acknowledged due to its excellent performance [5,6]. As far as is known, at least four scientific articles have listed ChatGPT as a co-author since its launch in 2022. However, there are also critical voices. In the context of scientific publishing, many publishers oppose the use of AI, some for the entire process and others only allowing it to improve the writing process, but not the interpretation of data [7], as explicitly stated in their publication conditions.
ChatGPT’s ability to integrate multiple types of input (such as patient questions) and generate outputs (detailed and comprehensible responses) offers a unique opportunity to enhance the user experience in the healthcare field. Although this tool does not replace direct interaction with healthcare professionals, it can better prepare patients for their consultations by reducing the barrier of technical knowledge and facilitating a more informed dialogue between the patient and the professional. The importance of online information search tools cannot be overlooked either [8]. The success of traditional search engines is now joined by chatbots, and the habit of online health inquiries is no exception.
There are still few definitive studies on the reliability of the information provided by GPT in the healthcare setting. However, the existing studies offer revealing data on its potential. Madrid-García et al. (2023) established that GPT is an excellent medium for answering medical questions with high precision and that its clinical reasoning is consistent [9].
A concept gaining importance in these times, driven by patient empowerment and active participation, is health literacy. This is defined as the health information an individual needs to know to make good decisions about their health [10]. It involves possessing the necessary knowledge, motivation, and skills to access, understand, and apply information in the healthcare domain.
This literacy is nourished by various sources, one of the most important being mass media, which includes social networks, search engines, and chatbots. All these sources contribute to informing or misinforming about health, including health promotion or public health programs. They even have the capacity to modify public actions and perceptions [11], especially when they come from groups the population identifies as reliable sources of information, such as healthcare professionals [12].
The Ottawa Charter for Health Promotion [13] provides an essential framework for the development of health literacy. According to this framework, as individuals increase their ability to control their health, their overall well-being tends to improve. For this reason, health literacy is recognized as a determinant of health.
The use of digital tools, such as chatbots, provides the population with a set of skills that allow them to access, understand, and use health-related information. If fed with reliable data, these tools can become great allies in the fight against misinformation and will be highly useful in enhancing patient independence and decision-making capacity. This can help avoid productivity losses among professionals by automating repetitive information delivery, where generative AI can be of great assistance.
Few studies have verified the capabilities of information created by generative AI, making it necessary to validate the reliability and usefulness of the information provided by ChatGPT. There is scepticism about whether the information obtained may be incorrect or dangerous for users relying on chatbots for medical information.
This study focuses on evaluating how multimodal technologies, specifically generative AI, can support and enhance health literacy, a fundamental aspect for informed decision-making by patients. It assesses ChatGPT’s ability to match the quality of responses provided by medical professionals in ophthalmology and determines whether its responses contribute to improving health literacy.
Specifically, we aim to (i) determine whether generative AI can provide responses of comparable quality to those of expert professionals, (ii) assess whether the information generated by ChatGPT-3.5 can be effectively understood and used by non-expert users in ophthalmology, and (iii) discuss how the information provided by ChatGPT-3.5 can be integrated into a broader framework of ophthalmological health literacy, which includes the critical interpretation and application of this information under the supervision of healthcare professionals.

2. Materials and Methods

Regarding the use of ChatGPT-3.5 in our study, we would like to clarify that this version was active at the time the research was conducted, and as it was the free version, it was the most accessible and widely used by the public. This decision allowed us to evaluate the potential of the tool in a more realistic and representative context of widespread use among users.
The selection process for the panel of ophthalmologists was designed to ensure the participation of experts with extensive clinical experience and deep knowledge in ophthalmology, thus ensuring the quality and relevance of the questions selected for the study. The only inclusion criterion was that the ophthalmologists be active in clinical practice and have experience in managing a wide variety of ophthalmological conditions.
Ophthalmologists from various leading medical institutions in Spain were invited to participate, forming a diverse panel in terms of both subspecialties within ophthalmology (retina, cornea, glaucoma, among others) and geographical distribution. These experts were asked to provide a list of the most common questions they receive from their patients during consultations, covering a wide range of topics, from diagnostic aspects to treatment and prognosis considerations.
A total of 36 different questions were received, which were reviewed by a member of the research team, who analyzed their relevance, clarity, and representativeness. From these, the 12 most recurrent and representative questions were selected, ensuring they covered a diversity of common topics in ophthalmological practice and were understandable to a general audience.
The selected questions were adapted to improve their clarity and were then used to generate responses via ChatGPT-3.5 in two different scenarios: in the first, without a specific prompt (simulating the interaction of a non-expert user), and in the second, with a prompt that simulated the response of a medical specialist (simulating the interaction of an expert user). It is important to note that “expert” users, such as healthcare professionals, and “non-expert” users, such as patients, have different levels of understanding and expectations regarding the information provided. These differences are reflected in how they interact with the chatbot to obtain information, so the aim is to determine whether the responses generated by ChatGPT-3.5 are equally useful and reliable for users with different levels of health literacy, regardless of their level of knowledge.
The ophthalmologists evaluated ChatGPT-3.5’s responses to the 12 selected questions using three validated scales. The Global Quality Scale (GQS) [14,15], on a five-point scale (where 1 indicates the worst information and 5 the best), was used to assess the quality and ease of understanding of the information. The Reliability Score (RS) scale evaluated the reliability of the medical sources, and the Usefulness Score (US) scale assessed the usefulness of the response for the patient, both on a scale of 0 to 7, where 7 indicates the highest reliability and usefulness, respectively.
A total of 42 questionnaires were received evaluating the questions, 21 for each type of prompt (customized or not). However, one of the questionnaires from the non-expert user group was discarded due to being incomplete. After an initial exploratory analysis, the sample was expanded to 400 questionnaires through synthetic data generation using the bootstrapping technique, which allows for the creation of random samples from an existing dataset through resampling. Python libraries pandas, numpy, and os were used.
The list of questions provided by the panel of experts was as follows:
  • Q1. Will my cataracts be operated on with a laser?
  • Q2. Is eye pressure related to blood pressure?
  • Q3. How many drops of eye drops should I use each time?
  • Q4. If I have surgery, will I stop being myopic?
  • Q5. Why don’t ophthalmologists get surgery to get rid of their glasses?
  • Q6. What is considered normal eye pressure?
  • Q7. Until when will my myopia increase?
  • Q8. Is vitreous detachment the same as retinal detachment?
  • Q9. When will I regain my vision after retinal surgery?
  • Q10. Do I have macula, doctor?
  • Q11. Will it hurt a lot if I get an injection in my eye?
  • Q12. Is astigmatism for distance or near vision?
Descriptive, association, and comparison statistical analyses were performed using the open-source software R version 4.4.1 and the graphical interface RStudio, now called Posit. The Shapiro–Wilk and Levene tests were used to verify the assumptions of normality and homoscedasticity, respectively. The level of statistical significance was set at p < 0.05.

3. Results

The results obtained using ChatGPT-3.5 demonstrated that the generated responses were rated as useful and reliable by the ophthalmologists in both user categories (expert and non-expert). Representative examples of the responses are included to illustrate its performance in the context of ophthalmological questions, highlighting the usefulness and reliability of the answers.
Example 1—Question about surgical procedures For the question “Will my cataract surgery be done with a laser?”, ChatGPT 3.5 generated the following response, as a non-expert user: “Cataract treatment is commonly performed using a technique called phacoemulsification, which uses ultrasound to dissolve the cloudy lens. In some cases, a laser may be used for certain parts of the procedure, but it is not the most common approach”. The ophthalmologists rated this response highly for usefulness (5.5/7) and reliability (5/7) for non-expert users, noting that the answer is accurate and understandable, although more details could be added for expert users.
Example 2—Question about ophthalmological conditions: For the question “What is considered a normal eye pressure?”, ChatGPT 3.5 generated the following response, as an expert user: “Normal intraocular pressure generally ranges between 10 and 21 mmHg. However, higher values may be indicative of glaucoma, especially if other risk factors are present. It is important to have regular check-ups to monitor eye pressure”. This response was rated 6/7 for usefulness and 6.5/7 for reliability by the ophthalmologists, who highlighted that the response is appropriate for an expert context, providing accurate information with a level of detail that supports its validity.
The results show that there are no statistically significant differences between the responses of expert and non-expert users, although the expert user responses received slightly higher ratings. On the GQS scale, both types of responses exceeded the midpoint of the scale, with an average score of 7.66 for the expert user and 7.38 for the non-expert user. Regarding reliability (RS) and usefulness (US), the expert user responses were also rated higher, although some specific questions received better scores when answered without prompt customization.
Regarding the reliability of the responses, on the Global Quality Score (GQS) scale, no statistically significant differences were found between the responses of the expert user and the non-expert user for most questions. Both types of responses exceeded the midpoint of the scale, with a mean of 7.66 for the expert user and 7.38 for the non-expert user. Individually, in questions Q3, Q5, Q8, Q9, and Q10, the non-expert user responses were rated higher than those of the expert user, although these differences were not statistically significant (Table 1).
On the Reliability Score (RS) scale, the responses were generally considered reliable, with a mean of 4.99 for the expert user and 4.75 for the non-expert user. In questions Q3, Q8, Q9, and Q10, the non-expert user responses slightly exceeded those of the expert user, without statistically significant differences (Table 2).
Analyzing the usefulness for the patient, the Usefulness Score (US) scale showed a mean of 5.12 for the expert user and 4.73 for the non-expert user. Only in question Q1 were statistically significant differences found (p = 0.008) in favor of the expert user (M = 4.75) compared to the non-expert user (M = 2). In questions Q3, Q9, Q10, and Q12, the non-expert user responses were rated higher than those of the expert user, although these differences did not reach statistical significance (Table 3).
The results indicate that, in general, there are no statistically significant differences between the responses generated by ChatGPT under the expert user role and those generated under the non-expert user role. However, the expert user responses tend to be slightly better rated in terms of reliability and usefulness.

4. Discussion

This study has demonstrated the ability of generative AI to generate responses with above-average reliability and usefulness, correctly answering all the questions posed, according to the consulted ophthalmologists. This opens an interesting field of application in patient education and even for future professionals [16].
The responses generated by ChatGPT-3.5 were consistently rated as useful and reliable by the ophthalmologists, both in contexts involving expert and non-expert users. However, it is important to clarify that, while the responses generated by ChatGPT-3.5 exceeded the midpoint of the reliability and usefulness scales, not all observed differences reached a level of statistical significance. This underscores the potential of AI as a complementary tool in education and health literacy, while also highlighting the importance of personalizing and adapting the model to different user profiles.
Differences were observed in the evaluation of responses between expert and non-expert users, with responses in expert user contexts being rated slightly higher. It is noteworthy that 4 questions received better ratings from non-expert users than from expert users, accounting for 33.3% of the questions asked. This may be influenced by the chatbot’s training or the availability of more consistent information. In digital health, this means that AI-generated interfaces and responses might need adjustments depending on the user’s prior knowledge, which could maximize the tool’s usefulness and effectiveness. For the future development of AI in health, this differentiated approach could improve user satisfaction and precision, enhancing the adoption of these technologies in clinical settings.
In the context of health literacy, the model demonstrated its ability to provide accurate and accessible information, which is essential for empowering patients in managing their ophthalmological health. Moreover, the evaluations conducted by the ophthalmologists indicate that, although the responses are consistent and of high quality, there is room to further personalize the prompts to better meet the specific needs of expert users. This suggests the importance of correctly training the model with personalized prompts, guided by medical specialists, to avoid unrepresentative answers.
The responses provided by generative AI proved to be particularly useful for non-expert users, offering clear and accessible explanations on complex ophthalmological topics. This highlights AI’s potential to democratize access to healthcare information, especially in specialized fields like ophthalmology. By providing answers that are easily understood by patients without medical training, ChatGPT can play a crucial role in improving health literacy, helping patients make informed decisions about their eye health. This accessibility is an important step towards health equity, ensuring that even those with fewer resources or knowledge can access accurate and useful information.
The expansion of the study through bootstrapping allowed us to observe specific patterns in the questions where ChatGPT-3.5’s responses were more useful for non-expert users. Identifying these patterns provides valuable insights for designing digital interventions aimed at improving health literacy. Developers of digital health technologies can use these findings to optimize how AI generates educational content, focusing on areas where non-expert users need more support. Additionally, this approach can guide the creation of personalized AI-based educational modules that dynamically adapt to the user’s needs in real-time.
The reliability, accuracy, and usefulness of the information provided can ensure its use as an autonomous information source, with specialists defining its scope of action. The ability of ChatGPT-3.5 to generate responses that meet the expectations of users with different levels of knowledge reinforces its applicability in health education and improving health literacy. This can reduce the informational burden on healthcare professionals, allowing them to focus on more complex cases. Thus, a potential use as a virtual assistant for certain frequently asked questions by patients could relieve specialists from informational bureaucracy or provide a foundational knowledge base to make the consultation with the specialist more understandable. Furthermore, its rapid advancement and implementation benefit all stakeholders in the healthcare sector, including patients [17], so we are likely to witness a true explosion of interventions aimed at them.
There are several limitations associated with the use of ChatGPT, the most notable being the uncertainty regarding data updates and, consequently, the training updates of the models, since it does not perform web searches and rejects information from the past two years.
The study was based on 12 ophthalmological questions selected by a panel of experts. While these questions were chosen to represent a diverse range of common inquiries in clinical practice, the limited number of questions may not capture the full spectrum of possible interactions and concerns that patients may have in a real-world setting. This limitation implies that the study’s results may not be fully representative of all potential ophthalmological consultations. Selecting a greater number of questions could have provided a broader and more detailed view of how ChatGPT-3.5 handles different types of ophthalmological inquiries. Additionally, with only 12 questions, it is possible that some key areas of ophthalmology were not covered, limiting the applicability of the results to more general contexts or specific ophthalmological subspecialties.
Related to the above, the limited number of questions could increase the variance in the responses generated by ChatGPT-3.5, as some questions might be more easily handled by the AI than others, which could bias the results. This aspect may restrict the study’s ability to generalize its findings to other areas of medicine or even to other specialties within ophthalmology. Although the use of techniques like bootstrapping helped increase the sample size and reduce variance, a larger initial sample of questions could have provided a more robust database for analysis.
Another limitation related to ethics is the tendency of these models to amplify biases [18,19], such as automation bias or attributing greater importance to information generated by a computer program. However, improvements and continuous training, due to the iterative process carried out daily by millions of users, are enhancing this aspect.
Finally, we cannot overlook ChatGPT’s so-called hallucinations, which occur due to the need to always provide an answer, even when not all the information is available, creating strange and disconnected compositions from reality.
Artificial intelligence is integrating into healthcare information systems, coexisting and potentially becoming a fundamental pillar of the medicine of the near future [20]. From this statement, new research should begin, exploring different concepts of the relationship between technology and health.

5. Conclusions

This study demonstrates that generative artificial intelligence, such as ChatGPT, has significant potential to complement patient education in the healthcare field by providing responses with above-average reliability and usefulness, as assessed by the consulted ophthalmologists. In this context, we propose that generative AI, specifically ChatGPT, should be considered a tool within the broader process of ophthalmological health literacy, rather than a comprehensive solution. By providing accurate and well-founded information, it can facilitate the first stage of health literacy: access to information. However, we acknowledge that this is only the first step in a more complex process that includes understanding, critical evaluation, and the application of information by the patient—tasks that require the ongoing guidance of healthcare professionals.
The use of ChatGPT in healthcare could alleviate the informational burden on healthcare professionals, allowing them to focus on more critical aspects of patient care. By delivering precise and useful information, ChatGPT can better prepare patients for their consultations, improve their understanding of medical procedures, and increase their independence in managing their health.
To maximize the potential of ChatGPT and ensure its effective integration into healthcare, it is essential to personalize the prompts used to generate responses, ensuring they reflect the accuracy and depth required in the medical field. This will lead to close collaboration with medical specialists to design and oversee these prompts. Additionally, it is necessary to keep the data constantly updated and implement strategies to identify and mitigate biases in the generated responses, ensuring fairness and representativeness in the information provided. Lastly, mechanisms of transparency must be established in the use of generative AI, along with ongoing supervision by experts to validate the quality of the responses and their suitability for patient needs.
Finally, this research not only provides new insights into the applicability of generative AI in the medical field but also offers an interdisciplinary perspective that combines data analysis, artificial intelligence, and communication sciences.

Author Contributions

Conceptualization: C.R.-N. and I.H.-P. Formal analysis: A.J.G.R. Data curation: C.R.-N. Investigation: C.R.-N., A.J.G.R. and I.H.-P. Writing—original draft: J.G.R. and S.M.G.M. Methodology: C.R.-N., A.J.G.R. and I.H.-P. Writing—review and editing: M.S.C.S. Supervision: C.R.-N. and J.G.R. Validation: I.H.-P. Project administration: S.M.G.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data are available upon request from the corresponding author.

Acknowledgments

We express our gratitude to the ophthalmology specialists who developed the questions and evaluated the responses for their collaboration and assistance.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shortliffe, E.H. MYCIN: A Rule-Based Computer Program for Advising Physicians Regarding Antimicrobial Therapy Selection. Doctoral Dissertation, Stanford University, Stanford, CA, USA, 1974. [Google Scholar]
  2. Adamopoulou, E.; Moussiades, L. An overview of chatbot technology. In En IFIP Advances in Information and Communication Technology; Springer International Publishing: Berlin/Heidelberg, Germany, 2020; pp. 373–383. [Google Scholar]
  3. Deveci, C.D.; Baker, J.J.; Sikander, B.; Rosenberg, J. A comparison of cover letters written by ChatGPT-4 or humans. Dan. Med. J. 2023, 70, A06230412. Available online: https://pubmed.ncbi.nlm.nih.gov/38018708/ (accessed on 15 March 2024).
  4. Else, H. Abstracts written by ChatGPT fool scientists. Nature 2023, 613, 423. [Google Scholar] [CrossRef]
  5. Thorp, H.H. ChatGPT is fun, but not an author. Science 2023, 379, 313. [Google Scholar] [CrossRef] [PubMed]
  6. Editorials, N. Tools such as ChatGPT threaten transparent science; here are our ground rules for their use. Nature 2023, 613, 612. [Google Scholar] [CrossRef]
  7. Sample, I. Science Journals Ban Listing of ChatGPT as Co-Author on Papers; Ian Sample Science, Ed.; The Guardian: London, UK, 2023. [Google Scholar]
  8. Eurostat. One in Two EU Citizens Look for Health Information Online. (6 April 2021). Available online: https://ec.europa.eu/eurostat/web/products-eurostat-news/-/edn-20210406-1 (accessed on 15 March 2024).
  9. Madrid-García, A.; Rosales-Rosado, Z.; Freites-Nuñez, D.; Pérez-Sancristóbal, I.; Pato-Cour, E.; Plasencia-Rodríguez, C.; Cabeza-Osorio, L.; Abasolo-Alcázar, L.; León-Mateos, L.; Fernández-Gutiérrez, B.; et al. Harnessing ChatGPT and GPT-4 for evaluating the rheumatology questions of the Spanish access exam to specialized medical training. Sci. Rep. 2023, 13, 22129. [Google Scholar] [CrossRef] [PubMed]
  10. Canela-Soler, J.; Elvira-Martínez, D.; Labordena-Barceló, M.J.; Loyola-Elizondo, E. Information systems in health and health indicators: An integrating perspective. Med. Clin. 2010, 134 (Suppl. S1), 3–9. [Google Scholar] [CrossRef] [PubMed]
  11. Mheidly, N.; Fares, J. Leveraging media and health communication strategies to overcome the COVID-19 infodemic. J. Public Health Policy 2020, 41, 410–420. [Google Scholar] [CrossRef] [PubMed]
  12. Herrera-Peco, I.; Jiménez-Gómez, B.; Peña Deudero, J.J.; Benitez De Gracia, E.; Ruiz-Núñez, C. Healthcare Professionals’ Role in Social Media Public Health Campaigns: Analysis of Spanish Pro Vaccination Campaign on Twitter. Healthcare 2021, 9, 662. [Google Scholar] [CrossRef] [PubMed]
  13. World Health Organization. Ottawa Charter for Health Promotion; World Health Organization: Ottawa, AN, Canada, 1986. [Google Scholar]
  14. Barlas, T.; Ecem Avci, D.; Cinici, B.; Ozkilicaslan, H.; Muhittin Yalcin, M.; Eroglu Altinova, A. The quality and reliability analysis of YouTube videos about insulin resistance. Int. J. Med. Inform. 2023, 170, 104960. [Google Scholar] [CrossRef] [PubMed]
  15. Memioglu, T.; Ozyasar, M. Analysis of YouTube videos as a source of information for myocarditis during the COVID-19 pandemic. Clin. Res. Cardiol. 2022, 111, 1113–1120. [Google Scholar] [CrossRef] [PubMed]
  16. Jin, J.Q.; Dobry, A.S. ChatGPT for healthcare providers and patients: Practical implications within dermatology. J. Am. Acad. Dermatol. 2023, 89, 870–871. [Google Scholar] [CrossRef] [PubMed]
  17. Johnson, K.B.; Wei, W.-Q.; Weeraratne, D.; Frisse, M.E.; Misulis, K.; Rhee, K.; Zhao, J.; Snowdon, J.L. Precision medicine, AI, and the future of personalized health care. Clin. Transl. Sci. 2021, 14, 86–93. [Google Scholar] [CrossRef] [PubMed]
  18. Salvagno, M.; Taccone, F.S.; Gerli, A.G. Can artificial intelligence help for scientific writing? Crit. Care 2023, 27, 75. [Google Scholar] [CrossRef] [PubMed]
  19. Beltrami, E.J.; Grant-Kels, J.M. Consulting ChatGPT: Ethical dilemmas in language model artificial intelligence. J. Am. Acad. Dermatol. 2023, 90, 879–880. [Google Scholar] [CrossRef] [PubMed]
  20. Kantor, J. ChatGPT, large language models, and artificial intelligence in medicine and health care: A primer for clinicians and researchers. JAAD Int. 2023, 13, 168–169. [Google Scholar] [CrossRef] [PubMed]
Table 1. Results according to the Global Quality Score (GQS).
Table 1. Results according to the Global Quality Score (GQS).
Non-Expert UserExpert UserX2UMW
95% CI 95% CI
MeanMedianLower UpperMeanMedianLower Upperpp
Q1220.933.073.6342.634.620.2020.024
Q22.7121.443.993.633.534.250.1490.12
Q33.7142.694.743.53.52.54.50.9660.72
Q43.1432.154.133.3832.494.260.550.658
Q54.4343.934.924.254.53.514.990.2520.8
Q63.8643.034.694.3843.944.810.1090.211
Q7443.084.9244.5350.7330.95
Q84.2953.265.314.1343.34.950.6890.569
Q94.4353.75.163.753.52.784.720.490.24
Q103.5742.075.073.383.52.044.710.9920.81
Q11443.084.924.254.53.514.990.7360.617
Q124.1453.025.273.753.52.784.720.7820.497
Table 2. Results according to the Reliability Scale (RS).
Table 2. Results according to the Reliability Scale (RS).
Non-Expert UserExpert UserX2UMW
95% CI 95% CI
MeanMedianLower UpperMeanMedianLower Upperpp
Q1320.935.074.54.53.165.840.3240.196
Q23.2932.014.564.3843.045.710.5090.214
Q35.4364.386.484.634.53.226.030.1380.289
Q44.1443.025.274.543.095.910.460.763
Q55.2954.266.315.385.54.26.550.8080.812
Q65.4354.256.616.1366.36.950.2320.277
Q75.1454.026.275.1353.836.420.1570.953
Q85.4363.757.15.3863.976.780.9190.856
Q95.5764.396.7554.53.456.550.5010.588
Q104.7162.287.144.254.51.946.560.920.718
Q114.7143.236.25.6364.376.880.6250.259
Q124.8653.226.5553.336.670.3820.769
Table 3. Results according to the Usefulness Scale (US).
Table 3. Results according to the Usefulness Scale (US).
Non-Expert UserExpert UserX2UMW
95% CI 95% CI
MeanMedianLower UpperMeanMedianLowerUpperpp
Q1210.313.964.7553.356.150.0350.008
Q2331.494.514.7553.515.990.4820.064
Q3563.46.64.7553.156.350.1990.814
Q44.2942.624.954.7543.515.990.4910.513
Q56.1475.027.275.756.54.437.070.9190.659
Q65.2954.016.566.1365.36.950.2670.249
Q75.1453.96.395.255.53.936.570.8080.812
Q85.5763.987.165.564.166.840.9660.808
Q95.7154.556.87553.276.730.6880.501
Q104.5752.516.634.384.52.196.560.8290.905
Q11559.936.075.564.326.680.460.406
Q125.1453.56.78553.336.670.6260.906
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ruiz-Núñez, C.; Gismero Rodríguez, J.; Garcia Ruiz, A.J.; Gismero Moreno, S.M.; Cañizal Santos, M.S.; Herrera-Peco, I. Can Generative AI Contribute to Health Literacy? A Study in the Field of Ophthalmology. Multimodal Technol. Interact. 2024, 8, 79. https://doi.org/10.3390/mti8090079

AMA Style

Ruiz-Núñez C, Gismero Rodríguez J, Garcia Ruiz AJ, Gismero Moreno SM, Cañizal Santos MS, Herrera-Peco I. Can Generative AI Contribute to Health Literacy? A Study in the Field of Ophthalmology. Multimodal Technologies and Interaction. 2024; 8(9):79. https://doi.org/10.3390/mti8090079

Chicago/Turabian Style

Ruiz-Núñez, Carlos, Javier Gismero Rodríguez, Antonio J. Garcia Ruiz, Saturnino Manuel Gismero Moreno, María Sonia Cañizal Santos, and Iván Herrera-Peco. 2024. "Can Generative AI Contribute to Health Literacy? A Study in the Field of Ophthalmology" Multimodal Technologies and Interaction 8, no. 9: 79. https://doi.org/10.3390/mti8090079

Article Metrics

Back to TopTop