Next Article in Journal
Lawsonia inermis as an Active Corrosion Inhibitor for Mild Steel in Hydrochloric Acid
Previous Article in Journal
The Validation of the Defensive Reactive Agility Test in Top-Level Volleyball Male Players: A New Approach to Evaluating Slide Speed Using Witty SEM
Previous Article in Special Issue
Chemoembolization for Hepatocellular Carcinoma Including Contrast Agent-Enhanced CT: Response Assessment Model on Radiomics and Artificial Intelligence
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Google Gemini’s Performance in Endodontics: A Study on Answer Precision and Reliability

1
Department of Pre-Clinic Dentistry, School of Biomedical Sciences, Universidad Europea de Madrid, Calle Tajo s/n, Villaviciosa de Odón, 28670 Madrid, Spain
2
Department of Clinical Dentistry, School of Biomedical Sciences, Universidad Europea de Madrid, Calle Tajo s/n, Villaviciosa de Odón, 28670 Madrid, Spain
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(15), 6390; https://doi.org/10.3390/app14156390
Submission received: 26 June 2024 / Revised: 14 July 2024 / Accepted: 20 July 2024 / Published: 23 July 2024
(This article belongs to the Special Issue Advances in AI-Powered Medical Applications)

Abstract

:
(1) Background: Large language models (LLMs) are revolutionising various scientific fields by providing advanced support tools. However, the effectiveness of these applications depends on extensive, up-to-date databases to ensure certainty and predictive power. Transparency about information sources in Medicine remains a significant issue. (2) Methods: To evaluate Google Gemini’s accuracy and reproducibility in endodontic diagnosis and treatment, 60 questions were designed based on the European Society of Endodontology Position Statements. Thirty questions were randomly selected and answered using Gemini during April 2023. Two endodontic experts independently scored the answers using a 3-point Likert scale. Discrepancies were resolved by a third expert. The relative frequency and absolute percentage of responses were detailed. Accuracy was assessed using the Wald binomial method, and repeatability was assessed using percentage agreement, Brennan and Prediger’s coefficient, Conger’s generalised kappa, Fleiss’ kappa, Gwet’s AC, and Krippendorff’s alpha, all with 95% confidence intervals. Statistical analysis was performed using STATA software. (3) Results: A total of 900 answers were generated. The percentage of correct answers varied from 0% to 100% per question. Overall accuracy was 37.11% with a 95% confidence interval of 34.02–40.32%; (4) Conclusions: Gemini is not currently designed for medical use and therefore needs to be used with caution when considered for this purpose.

1. Introduction

Access to large language models (LLMs) is transforming several scientific disciplines by providing a wide range of support tools. These models, an advanced form of artificial intelligence (AI), allow interaction with information available on the Internet or any database to which they are applied, with the ability to analyse complex algorithms and self-learning. We are entering a new era in medicine. AI can be applied to clinical practice through risk assessment models, improving diagnostic accuracy and workflow efficiency [1].
These models can be used in medical diagnostics as well as in symptom checking applications to provide potential diagnoses and assist users in self-classification [2]. They can also support clinical decision making and patient monitoring, even suggesting inter-professional consultations based on signs and symptoms [3].
However, for these applications to be truly effective, it is essential to have a large number of up-to-date databases. This increases the level of certainty, credibility, and predictive power of the models [4]. One of the main problems with these models is the lack of transparency about the sources of information they use. This can raise concerns about the veracity and reliability of the information provided, as users have no way of verifying the accuracy of the data or the context in which they originated [5].
There are numerous tools on the market. One of the most popular is Chat Generative Pre-trained Transformer (ChatGPT, OpenAI, San Francisco, CA, USA) [6], the latest version of which was released in February 2023. In the same year were also introduced Gemini from Google DeepMind (London, UK) and Google Bard chatbot (Alphabet, Mountain View, CA, USA), although their development dates back to 2020 [7].
Gemini is trained on a multimodal and multilingual dataset that includes web documents, books, and code, as well as image, audio, and video data [8].
The use of chatbots in healthcare could be of significant benefit, as they can assist both professionals and patients [9]. Despite a positive reception from users and early promising signs, further studies are needed to assess the full extent of the utility of these chatbots in healthcare settings [10]. In dentistry, although the scientific literature has already demonstrated the effectiveness of chatbot applications [11,12], the use of Gemini remains untested in terms of determining its reliability and reproducibility in a clinical setting.
Therefore, the aim of this study was to evaluate the accuracy and reproducibility of Gemini-generated answers to diagnostic and treatment questions in endodontics, to test its effectiveness as a diagnostic tool for dentists.

2. Materials and Methods

2.1. Ethical Approval

This research did not require ethical approval since no human participants were involved.

2.2. Question Design

To evaluate the accuracy and repeatability of the Gemini Pro software (Google, Mountain View, CA, USA) in generating answers to questions about diagnosis and treatment in endodontics, two endodontic specialists (M.T. and V.D.F.) designed 60 questions based on the European Society of Endodontology (ESE) Position Statements [13]. These documents were chosen because they contain clinical practice guidelines developed by a group of recognised experts in endodontics.
Each question was preceded by the prompt “To answer this question, imagine that you are an endodontist and I am a general dentist”, thus directing the answers to a specific situation.
Thirty questions were then randomly selected using a random number generator within a range using JavaScript ECMAScript 2021 (Mozilla Corporation, San Francisco, CA, USA). The selected questions are listed in Table 1.

2.3. Generating Answers in Gemini

Thirty answers were generated for each question using Gemini with the ‘new chat’ function on each occasion. The answers were obtained at different times of the day, namely morning, afternoon, and evening, during the month of April 2023. All answers were stored in a spreadsheet (Excel version 16; (Microsoft, Redmond, WA, USA)).

2.4. Evaluation of Gemini Answers by Human Experts

Two experts in endodontics (B.T. and A.S.) independently scored the 900 generated answers using the guidelines used to design the questions, with a 3-point Likert scale (0 = incorrect; 1 = partial or incomplete; 2 = correct). Discrepancies in the ratings were resolved by a third expert (YF).

2.5. Statistical Analysis

The 900 ratings of each answer were analysed using the statistical software STATA, version BE 14 (StataCorp, Texas, TX, USA).
For each of the 30 questions, the relative frequency (n) and absolute percentage (%) of answers that received different ratings from the experts were described. To analyse the accuracy of the answers generated by Gemini, the proportion of answers scoring 2 (correct) was calculated for both the total set of questions and for each individual question, together with its 95% confidence interval using the Wald binomial method.
To assess repeatability, the consistency of ratings across replicates was analysed using weighted concordance analysis for ordinal categories and multiple replicates. Measures included percent agreement, Brennan and Prediger’s coefficient, Conger’s generalised kappa, Fleiss’ kappa, Gwet’s AC, and Krippendorff’s alpha with their corresponding 95% confidence intervals. The 900 evaluations of each response were analysed using STATA statistical software, version BE 14 (StataCorp, Texas, TX, USA).

3. Results

Gemini generated 900 answers, providing 30 answers for each of the 30 questions asked. Table 2 shows the frequency of expert grading for Gemini answers. The percentage of correct repetitions for the questions ranged from 0% to 100%, depending on the question. The experts considered that eight questions showed accuracy and repeatability; five questions showed incomplete or partial answers and repeatability, and two questions (12 and 22) showed both inaccuracy and repeatability. A total of 16 questions showed varying degrees of repeatability. The overall accuracy of the answers generated by Gemini was 37.11%, with a 95% confidence interval ranging from 34.02% to 40.32%, as shown in Table 3. The repeatability of expert grading of Gemini generated answers ranged from substantial to almost perfect.
It was observed that Gemini could not solve one of the questions (“What are the recommended antibiotics in endodontics?”) in any of the repetitions given, obtaining as an answer “I can’t assist you with that, as I’m only a language model and don’t have the capacity to understand and respond” (Figure 1).

4. Discussion

The introduction of new technologies in the healthcare sector has had a transformative effect on the diagnosis and treatment of diseases as well as facilitating the work of healthcare professionals.
In particular, the use of LLMs by professionals is proving to be a promising tool, not only from a clinical perspective [14], but also in the training of future healthcare professionals [15]. These models can analyse large amounts of medical data, provide preliminary diagnoses and offer treatment recommendations, which can optimise clinical decision making.
However, despite the considerable potential of these applications, the most well-known programmes are currently not specifically designed for medical use or use in medical settings. This means that they need to be used with sufficient caution and under professional supervision [16].
In addition, there are concerns about the transparency and provenance of the data used to train these models, which may affect the reliability of the results.
Although ChatGPT versions 3.5 and 4 (OpenAI, San Francisco, CA, USA) have become the standard reference for studies on the use of chatbots in the medical sciences [17,18], particularly in dentistry [19,20], there is limited evidence on the performance of other LLMs, such as Gemini.
The Gemini software is presented as a new tool that, unlike ChatGPT, is characterised by its multimodal capability [7]. This means that Gemini can understand and process photos, audio, text, and other types of input natively, without the need to implement additional plug-ins.
Its use has been documented in educational activities for dental students [21,22], demonstrating its potential as a didactic tool. However, its utility as a diagnostic tool has not received the same level of scientific development and validation as its competitor, ChatGPT.
The results of this study indicate that the accuracy of the answers provided by the Gemini model is between 34.02% and 40.32% with 95% confidence. These results suggest that although Gemini has the potential to be a useful tool for diagnostic support, its current performance may not be sufficiently reliable for critical clinical applications, as approximately one third of the answers provided by Gemini are correct, which may not be sufficient for applications requiring high precision, such as diagnosis or treatment recommendations.
Although the high repeatability of Gemini’s answers is a positive aspect, providing results considered to be between substantial and near perfect, it is important to note that the consistency of incorrect answers remains an issue, highlighting the need for improvements in the model’s accuracy.
In this context, Ozden et al. [23] analysed the ability of Gemini to answer 25 dichotomous (yes/no) questions about dental trauma based on the International Association of Dental Traumatology guidelines, [24], and obtained 64% correct answers with Gemini, compared to 37.11% in the present study. However, Suárez et al. obtained slightly lower accuracy results than the present study when evaluating the performance of ChatGPT in generating answers in the field of endodontics. In another study [25], ChatGPT-3.5 provided more credible information compared to Bing and Google Bard. Therefore, different factors could influence the different results observed in the literature. The type of question (dichotomous versus open-ended) could influence the results of the tool. However, similar to the present study, several studies [11,23] also used guidelines published by international scientific societies, providing scientific certainty both in the formulation of the questions and in the analysis of the answers obtained.
In the present study, the variability in the percentages of correct answers for diagnosis suggests that the accuracy of the answers depends significantly on how the questions are phrased. This may be due to factors such as the clarity of the question, the relevance of the content, or the inherent difficulty of the question, which is a limitation of the study. Analysis of this variability may help to improve the wording of questions in future surveys or exams, in order to obtain a more accurate measure of Gemini’s knowledge or skills.
On the other hand, during the answer acquisition phase, it was observed that the language used by Gemini differed from that used by ChatGPT [26], even using expressions such as “This is my recommendation” (Figure 2). This use of language suggests a significant level of personalisation and familiarity in the answers. However, this apparent personalisation in Gemini’s answers could be problematic and ethically questionable, especially in sensitive contexts such as health. Such expressions may mislead the user into believing that they are receiving advice from a human professional. Therefore, the structure of Gemini’s answers should ensure that the user understands that they are interacting with an AI model and not a human professional.
The use of chatbots is a promising future tool that can not only assist healthcare professionals, but also make a significant contribution to the sustainability of the healthcare sector. A platform that can access scientific literature databases without the limitations of a manual search, or with the immediacy that current technologies allow, would enable clinicians to make more reliable diagnoses, benefiting both the patient and society as a whole, and reducing the environmental impact associated with traditional medical care processes.
At present, there are places in the world where access to quality scientific information is very complex, both because of the cost involved and the quality of the computer systems However, with just one of these tools it would be possible to have access to millions of pieces of scientific evidence more efficiently and sustainably.
In this sense, the systems that power these applications need to be trained by clinicians to prevent the errors that chatbots can make, as this study shows, which currently limit their use in dentistry. In addition, optimising these tools to increase their sustainability could lead to more efficient resource management and a reduction in waste, thereby contributing to global sustainable development goals.

5. Conclusions

It is important to recognise that Google’s Gemini application is not designed for clinical use, as is clearly stated in its documentation. This limitation underlines the need for extreme caution when using it in medical contexts. Although the use of chatbots promises to be one of the most transformative developments in healthcare, the use of Gemini in diagnostic processes still lacks the necessary scientific robustness. It is therefore essential that its use in medical practice is always monitored by professionals to ensure its effectiveness and compliance with the ethical standards required for such applications. Further studies are needed to determine the performance of chatbots in medical applications, as well as comparisons of the effectiveness of the different LLMs currently available on the market.

Author Contributions

Conceptualization, methodology, and writing, V.D.-F.G., Y.F. and A.S.; investigation, M.T., A.S., B.T., R.E. and V.D.-F.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy reasons.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kaul, V.; Enslin, S.; Gross, S.A. History of artificial intelligence in medicine. Gastrointest. Endosc. 2020, 92, 807–812. [Google Scholar] [CrossRef]
  2. You, Y.; Gui, X. Self-Diagnosis through AI-enabled Chatbot-based Symptom Checkers: User Experiences and Design Considerations. AMIA Annu. Symp. Proc. 2021, 25, 1354–1363. [Google Scholar]
  3. Dave, T.; Athaluri, S.A.; Singh, S. ChatGPT in medicine: An overview of its applications, advantages, limitations, future prospects, and ethical considerations. Front. Artif. Intell. 2023, 4, 1169595. [Google Scholar] [CrossRef]
  4. Wang, Y.; Li, N.; Chen, L.; Wu, M.; Meng, S.; Dai, Z.; Zhang, Y.; Clarke, M. Guidelines, Consensus Statements, and Standards for the Use of Artificial Intelligence in Medicine: Systematic Review. J. Med. Internet Res. 2023, 22, e46089. [Google Scholar] [CrossRef]
  5. ChatGPT is a black box: How AI research can break it open. Nature 2023, 619, 671–672. [CrossRef] [PubMed]
  6. Kuroiwa, T.; Sarcon, A.; Ibara, T.; Yamada, E.; Yamamoto, A.; Tsukamoto, K.; Fujita, K. The Potential of ChatGPT as a Self-Diagnostic Tool in Common Orthopedic Diseases: Exploratory Study. J. Med. Internet Res. 2023, 25, e47621. [Google Scholar] [CrossRef]
  7. Gemini Team, Google. Gemini: A Family of Highly Capable Multimodal Models. Available online: https://storage.googleapis.com/deepmind-media/gemini/gemini_1_report.pdf (accessed on 12 June 2024).
  8. Saab, K.; Tu, T.; Weng, W.-H.; Tanno, R.; Stutz, D.; Wulczyn, E.; Zhang, F.; Strother, T.; Park, C.; Vedadi, E.; et al. Capabilities of Gemini Models in Medicine. Available online: https://arxiv.org/abs/2404.18416 (accessed on 12 June 2024).
  9. Erren, T.C. Patients, Doctors, and Chatbots. JMIR Med. Educ. 2024, 4, e50869. [Google Scholar] [CrossRef] [PubMed]
  10. Webster, P. Medical AI chatbots: Are they safe to talk to patients? Nat. Med. 2023, 29, 2677–2679. [Google Scholar] [CrossRef]
  11. Suárez, A.; Díaz-Flores García, V.; Algar, J.; Sánchez, M.G.; de Pedro, M.L.; Freire, Y. Unveiling the ChatGPT phenomenon: Evaluating the consistency and accuracy of endodontic question answers. Int. Endod. J. 2024, 57, 108–113. [Google Scholar] [CrossRef]
  12. Rao, A.; Pang, M.; Kim, J.; Kamineni, M.; Lie, W.; Prasad, A.K.; Landman, A.; Dreyer, K.; Succi, M.D. Assessing the Utility of ChatGPT Throughout the Entire Clinical Workflow: Development and Usability Study. J. Med. Internet Res. 2023, 22, e48659. [Google Scholar] [CrossRef]
  13. European Society of Endodontology. Resources for Clinicians. 2023. [WWW Document]. Available online: https://www.e-s-e.eu/for-professionals/resources-for-clinicians/ (accessed on 12 June 2024).
  14. Wu, J.; Ma, Y.; Wang, J.; Xiao, M.J. The Application of ChatGPT in Medicine: A Scoping Review and Bibliometric Analysis. Multidiscip. Healthc. 2024, 18, 1681–1692. [Google Scholar] [CrossRef] [PubMed]
  15. Shorey, S.; Mattar, C.; Pereira, T.L.; Choolani, M. A scoping review of ChatGPT’s role in healthcare education and research. Nurse Educ. Today 2024, 135, 106121. [Google Scholar] [CrossRef] [PubMed]
  16. Biswas, S.; Davies, L.N.; Sheppard, A.L.; Logan, N.S.; Wolffsohn, J.S. Utility of artificial intelligence-based large language models in ophthalmic care. Ophthalmic Physiol. Opt. 2024, 44, 641–671. [Google Scholar] [CrossRef] [PubMed]
  17. Gilson, A.; Safranek, C.W.; Huang, T.; Socrates, V.; Chi, L.; Taylor, R.A.; Chartash, D. How Does ChatGPT Perform on the United States Medical Licensing Examination (USMLE)? The Implications of Large Language Models for Medical Education and Knowledge Assessment. JMIR Med. Educ. 2023, 8, e45312. [Google Scholar] [CrossRef] [PubMed]
  18. Wang, L.; Wan, Z.; Ni, C.; Song, Q.; Li, Y.; Clayton, E.W.; Malin, B.A.; Yin, Z. A Systematic Review of ChatGPT and Other Conversational Large Language Models in Healthcare. medRxiv 2024, 27, 24306390. [Google Scholar]
  19. Huang, H.; Zheng, O.; Wang, D.; Yin, J.; Wang, Z.; Ding, S.; Yin, H.; Xu, C.; Yang, R.; Zheng, Q.; et al. ChatGPT for shaping the future of dentistry: The potential of multi-modal large language model. Int. J. Oral Sci. 2023, 28, 29. [Google Scholar] [CrossRef] [PubMed]
  20. Buzayan, M.M.; Sivakumar, I.; Mohd, N.R. Artificial intelligence in dentistry: A review of ChatGPT’s role and potential. Quintessence Int. 2023, 17, 526–527. [Google Scholar]
  21. Uribe, S.E.; Maldupa, I.; Kavadella, A.; El Tantawi, M.; Chaurasia, A.; Fontana, M.; Marino, R.; Innes, N.; Schwendicke, F. Artificial intelligence chatbots and large language models in dental education: Worldwide survey of educators. Eur. J. Dent. Educ. 2024, 1–12. [Google Scholar] [CrossRef]
  22. Ahmed, W.M.; Azhari, A.A.; Alfaraj, A.; Alhamadani, A.; Zhang, M.; Lu, C.T. The Quality of AI-Generated Dental Caries Multiple Choice Questions: A Comparative Analysis of ChatGPT and Google Bard Language Models. Heliyon 2024, 19, e28198. [Google Scholar] [CrossRef]
  23. Ozden, I.; Gokyar, M.; Ozden, M.E.; Sazak Ovecoglu, H. Assessment of artificial intelligence applications in responding to dental trauma. Dent. Traumatol. 2024, 1–8. [Google Scholar] [CrossRef]
  24. Bourguignon, C.; Cohenca, N.; Lauridsen, E.; Flores, M.T.; O’Connell, A.C.; Day, P.F.; Tsilingaridis, G.; Abbott, P.V.; Fouad, A.F.; Hicks, L.; et al. International Association of Dental Traumatology guidelines for the management of traumatic dental injuries: 1. Fractures and luxations. Dent. Traumatol. 2020, 36, 314–330. [Google Scholar] [CrossRef] [PubMed]
  25. Mohammad-Rahimi, H.; Ourang, S.A.; Pourhoseingholi, M.A.; Dianat, O.; Dummer, P.M.H.; Nosrat, A. Validity and reliability of artificial intelligence chatbots as public sources of information on endodontics. Int. Endod. J. 2024, 57, 305–314. [Google Scholar] [CrossRef] [PubMed]
  26. Antaki, F.; Touma, S.; Milad, D.; El-Khoury, J.; Duval, R. Evaluating the performance of ChatGPT in ophthalmology: An analysis of its successes and shortcomings. Ophthalmol. Sci. 2023, 3, 100324. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Answer obtained to question 29 in all its repetitions.
Figure 1. Answer obtained to question 29 in all its repetitions.
Applsci 14 06390 g001
Figure 2. Answer with a significant level of personalisation and familiarity.
Figure 2. Answer with a significant level of personalisation and familiarity.
Applsci 14 06390 g002
Table 1. Selected questions used in the answer generation.
Table 1. Selected questions used in the answer generation.
Guide Used (DOI)Question
10.1111/iej.13916Why does resorption appear in a tooth?
10.1111/iej.13916Which are the types of root resorption in dentistry?
10.1111/iej.13916What is the treatment for internal resorption in a tooth?
10.1111/iej.13916What is the treatment for external resorption in a tooth?
10.1111/iej.12606What is the objective of regenerative endodontics?
10.1111/iej.12606When is it contraindicated to perform regenerative endodontics?
10.1111/iej.12606What are the success criteria for regenerative endodontics?
10.1111/iej.12606What is the percentage, amount and time of sodium hypochlorite that we should use to irrigate during regenerative endodontics at the first appointment?
10.1111/iej.12606What is the irrigation protocol during the first appointment in regenerative endodontics?
10.1111/iej.12606After how much time can we perform the second session in regenerative endodontics?
10.1111/iej.12606What is the irrigation protocol in the second session of regenerative endodontics?
10.1111/iej.12606When should we do follow up on regenerative endodontics?
10.1111/iej.12606What type of material should be placed on top of the blood clot when we do regenerative root canal?
10.1111/iej.13456What is a surgical extrusion of a tooth?
10.1111/iej.13456What is an intentional reimplantation of a tooth?
10.1111/iej.13456What is dental autotransplantation?
10.1111/iej.13456What is the total extra-alveolar time during an intentional reimplantation?
10.1111/iej.13456What should splinting be like after intentional reimplantation in dentistry?
10.1111/iej.13456How long should splinting be maintained after intentional reimplantation?
10.1111/iej.13456How long is the maximum time to do a root canal after surgical procedures?
10.1111/iej.13456Should the socket be curetted in a surgical extrusion?
10.1111/iej.13456Should endodontics be performed on the donor tooth during autotransplantation?
10.1111/iej.13456Do donor teeth with open apex require root canal treatment in autotransplantation?
10.1111/iej.12741What are the indications for the use of antibiotics in endodontics?
10.1111/iej.12741When should we administer antibiotic prophylaxis in endodontics?
10.1111/iej.12741What are the recommended antibiotics in endodontics?
10.1111/iej.13916When is performing a CBCT in endodontics indicated?
10.1111/iej.13916How much FOV should we use in endodontics when performing a CBCT?
10.1111/iej.13916Should we evaluate the complete CBCT volume?
10.1111/iej.13916Can CBCTs be performed on children in endodontics?
Table 2. Frequency of expert grading for Gemini answers.
Table 2. Frequency of expert grading for Gemini answers.
Incorrect (0)Partially Correct or Incomplete (1)Correct (2)
Questionn%n%n%
100.0030100.0000.00
200.0030100.0000.00
300.001136.671963.33
400.0000.0030100.00
500.0030100.0000.00
600.001653.331446.67
700.0030100.0000.00
800.001240.001860.00
9000000.0030100.00
1030100.0000.0000.00
112996.6713.3300.00
1200.0030100.0000.00
1300.002273.33826.67
142790.0000.00310.00
1500.0000.0030100.00
1600.002996.6713.33
1700.0000.0030100.00
18413.332686.6700.00
1913.332996.6700.00
2030100.0000.0000.00
2100.002273.33826.67
222273.33516.67310.00
2300.0000.0030100.00
2400.001756.671343.33
2500.002790.00310.00
262583.3313.33413.33
2700.0000.0030100.00
2800.0000.0030100.00
292790.00310.0000.00
3000.0000.0030100.00
Table 3. Overall accuracy of the answers.
Table 3. Overall accuracy of the answers.
AccuracyStandard ErrorCI 95% (Wilson)
Overall 900 responses37.11%0.0234.02%40.32%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Díaz-Flores García, V.; Freire, Y.; Tortosa, M.; Tejedor, B.; Estevez, R.; Suárez, A. Google Gemini’s Performance in Endodontics: A Study on Answer Precision and Reliability. Appl. Sci. 2024, 14, 6390. https://doi.org/10.3390/app14156390

AMA Style

Díaz-Flores García V, Freire Y, Tortosa M, Tejedor B, Estevez R, Suárez A. Google Gemini’s Performance in Endodontics: A Study on Answer Precision and Reliability. Applied Sciences. 2024; 14(15):6390. https://doi.org/10.3390/app14156390

Chicago/Turabian Style

Díaz-Flores García, Victor, Yolanda Freire, Marta Tortosa, Beatriz Tejedor, Roberto Estevez, and Ana Suárez. 2024. "Google Gemini’s Performance in Endodontics: A Study on Answer Precision and Reliability" Applied Sciences 14, no. 15: 6390. https://doi.org/10.3390/app14156390

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop