Skip to Content
HealthcareHealthcare
  • Article
  • Open Access

15 September 2023

ChatGPT’s Skills in Statistical Analysis Using the Example of Allergology: Do We Have Reason for Concern?

Department of Pharmacotherapy and Pharmaceutical Care, Faculty of Pharmacy, Medical University of Warsaw, Banacha 1 Str., 02-097 Warsaw, Poland

Abstract

Background: Content generated by artificial intelligence is sometimes not truthful. To date, there have been a number of medical studies related to the validity of ChatGPT’s responses; however, there is a lack of studies addressing various aspects of statistical analysis. The aim of this study was to assess the validity of the answers provided by ChatGPT in relation to statistical analysis, as well as to identify recommendations to be implemented in the future in connection with the results obtained. Methods: The study was divided into four parts and was based on the exemplary medical field of allergology. The first part consisted of asking ChatGPT 30 different questions related to statistical analysis. The next five questions included a request for ChatGPT to perform the relevant statistical analyses, and another five requested ChatGPT to indicate which statistical test should be applied to articles accepted for publication in Allergy. The final part of the survey involved asking ChatGPT the same statistical question three times. Results: Out of the 40 general questions asked that related to broad statistical analysis, ChatGPT did not fully answer half of them. Assumptions necessary for the application of specific statistical tests were not included. ChatGPT also gave completely divergent answers to one question about which test should be used. Conclusion: The answers provided by ChatGPT to various statistical questions may give rise to the use of inappropriate statistical tests and, consequently, the subsequent misinterpretation of the research results obtained. Questions asked in this regard need to be framed more precisely.

1. Introduction

At the end of July 2023, guidance related to the use of artificial intelligence, language models and chatbots was published by JAMA Network. The authors indicated that non-human artificial intelligence tools should not be credited as co-authors. The use of such tools should be transparently reported when writing manuscripts and other content [1]. A similar position has been expressed by the World Association of Medical Editors (WAME) [2]. An article published in June 2023 in Allergy pointed out that ChatGPT can generate reliable but incorrect and meaningless responses. For this reason, it is essential to keep expert supervision in mind when publishing this type of material [3]. A number of articles have been published in recent months assessing the accuracy of ChatGPT’s answers. In an example study, a team of three oncologists with extensive professional experience assessed the correctness of answers to 30 questions. In total, 14% of the answers were incorrect, and thus there was a high risk of receiving incorrect recommendations [4]. Another study investigated the correctness of answers to questions on a highly demanding postgraduate exam, namely the European Exam in Core Cardiology (EECC). The results indicated significant success of ChatGPT in the EECC; however, it cannot replace qualities that doctors should possess such as critical thinking, innovation and creativity [5]. An interesting study was also conducted on 15 different sub-disciplines of otolaryngology in which the dataset contained 2576 questions, i.e., 479 multiple-choice and 2097 single-choice. The percentage of multiple-choice questions answered correctly by ChatGPT was lower than that for single-choice questions. The highest percentage of correct answers given was related to allergology, while the lowest was related to legal aspects of otolaryngology [5]. The provision of incorrect answers by ChatGPT with the occurrence of lower levels of statistical analysis carried out in recent years may further reduce the confidence of the public in science. In one of the most cited papers published in PLOS Medicine, Professor John P. A. Ioannidis argued that the results of the majority of scientific publications are fraudulent [6]. Other data published in PLOS ONE show that between 1998 and 2020, there was unfortunately no significant change in the frequency with which medical journal editors used statistical reviewers [7]. In addition, another factor contributing to the poorer quality of the analyses conducted is the non-practical way in which biostatistics is taught. The statistical recommendations published in 2016 by PLOS Biology point to the need for education through critical analysis of accepted manuscripts [8]. Such aspects mentioned above, combined with misinformation provided by ChatGPT related to the way statistical analysis is carried out, may unfortunately contribute to an even greater deterioration in the quality of published research results and thus the subsequent implementation of inappropriate recommendations. What is lacking in the literature is an analysis of the answers provided by ChatGPT about various aspects of statistical analysis. To the best of my current knowledge, this is the first study of its kind to be carried out. As Allergy is the official journal of the European Academy of Allergy and Clinical Immunology, the first aim of this research was to determine the validity of the statistical analyses used by the authors of the articles published in this journal. The second objective was to analyse the validity of the statistical analyses carried out by ChatGPT on the datasets created, the subjects of which were related to allergology. The third objective of this research was to investigate the correctness of ChatGPT’s answers to questions on various statistical issues, using the medical field of allergology as an example. The fourth objective of the manuscript was, using one statistical question asked of ChatGPT as an example, to indicate the variety of answers given. The final objective was to provide statistical recommendations to bear in mind when considering the use of ChatGPT for conducting statistical analyses.

2. Materials and Methods

2.1. ChatGPT Statistical Skills

Due to its wide and general availability, ChatGPT-3.5 was used in the present study. The study was divided into four parts. ChatGPT-4 was excluded from this study due to its cost and limited availability. The questions that ChatGPT answered were related to broad statistical analysis and were based on different groups of allergy patients. The first part of the survey consisted of asking ChatGPT 30 different questions about which statistical tests should be applied to particular cases. Other questions asked included the assumptions of the statistical tests necessary for their application, as well as other aspects related to the performance of the statistical analysis. The second part of the research carried out involved a request to ChatGPT for it to carry out relevant statistical analyses based on five specific datasets. The validity of the statistical tests used was assessed. The third part of the study involved five articles published in Allergy. ChatGPT provided answers to questions about what kind of statistical test it would apply to specific studies conducted by these authors, the results of which were subsequently published in Allergy. The articles were selected based on the various statistical tests applied. In order to show the variety of answers given, one question related to the conduct of a statistical analysis was selected and then asked several times of ChatGPT. Based on the answers given by ChatGPT to the 40 questions, the statistical recommendations mentioned in the introduction were created. The answers given were scored using a 0–1 system, where 0 was the wrong answer and 1 was the correct answer. Incorrect answers were related to the wrong choice of statistical tests, the tests’ assumptions, and other aspects related to conducting a statistical analysis, such as the correct description of the obtained test results, removal of outliers, etc. In an additional supplement file (Supplementary Materials) to this manuscript, screenshots of the questions asked of ChatGPT are included, along with its answers.

2.2. Statistical Analysis

For each part of the survey, the percentage of correct and incorrect answers given was calculated. A chi-square test was used to check whether ChatGPT gave more correct or incorrect answers. Statistical analysis was conducted using the IBM SPSS Statistics 25 Package (IBM SPSS Statistics Armonk, NY, USA: IBM Corp.). A p value < 0.05 was taken as the level of statistical significance.

3. Results

3.1. Correctness of ChatGPT’s Answers to Questions Related to Statistical Analysis

The following table lists the 30 questions asked of ChatGPT related to statistical analysis in the broadest sense. The number of correct answers out of 30 possible questions was 18 (60%). Of the answers given, 12 were incorrect. The difference in the number of correct and incorrect answers given was not statistically significant, χ2(1) = 1.2; p = 0.27. However, the responses associated with the need to use non-parametric equivalents of statistical tests were noteworthy. This refers to ChatGPT’s lack of consideration of the scale of the analysed variables, for example, related to ordinal nature, low sample size and other assumptions necessary for the application of specific tests. The remaining wrong answers given were to do with the wrong choice of statistical test, the wrong measure of effect size and the wrong descriptive statistics to be included in the description of the results obtained (Table 1). For this reason, statistical questions need to be formulated more precisely, bearing in mind the assumptions necessary for the application of specific statistical tests (equality of groups compared, normality of distribution, etc.).
Table 1. Correctness of ChatGPT’s answers to 30 different questions related to statistical analysis.

3.2. Correctness of ChatGPT’s Statistical Analyses for Five Specific Studies

In terms of the correctness of ChatGPT’s five different statistical analyses, as was observed for the 30 questions in Table 1, ChatGPT did not consider sample size or the scale on which the variable being analysed was measured. Failure to take this type of information into account also resulted in the incorrect indication of the appropriate measure of effect size to be used in a particular task (Table 2). In order to carry out a statistical analysis when asking ChatGPT questions involving concrete data, questions should be asked in a precise way. It is recommended to include information on the need to test the assumption of normality of distribution, assessment of sample size, homogeneity of variance, etc.
Table 2. Correctness of ChatGPT’s statistical analyses for five specific studies.

3.3. Addressing the Statistical Tests That Should Be Used in Published Articles in Allergy

ChatGPT gave correct answers to the questions related to three of the five selected articles published in Allergy. Once again, the incorrect responses were related to the fact that ChatGPT did not consider aspects such as sample size and the equality of the groups of people being compared, and therefore did not select the appropriate effect size measure (Table 3) [9,10,11,12,13]. The recommendations related to the following statements are the same as those made when describing Table 2 (test assumptions, measurement scale of variables, etc.). It is important to remember to specify the questions being asked.
Table 3. Correctness of ChatGPT’s indication of answers to statistical questions related to articles accepted for publication in Allergy.

3.4. Repeatability of the Answer Given to a Specific Statistical Problem

When asking ChatGPT questions related to different aspects of the statistical analyses being conducted, it was noted that there were discrepancies in the answers given to the same questions. One question was selected from the 30 questions presented in the first table and then presented to ChatGPT three times to show the discrepancies in the answers given. Of the 30 questions presented in Table 1, the second was selected for analysis due to the scale of the variable under analysis. The first answer given recommended the use of a repeated-measure analysis of variance, without indicating the validity of using the Friedman test in this particular case. The second answer given to the same question indicated a chi-square test, but the subsequent response acknowledged that this test is used for independent observations. For this reason, ChatGPT noted the validity of using the McNemar test in this particular case, which was also not an appropriate choice. A surprising answer was received the third time. When the question was asked a second time, the response obtained, despite mentioning the chi-square test, did not support the use of this test. When asked the same question a third time, ChatGPT indicated that “Since you have multiple measurements for each patient taken at different times, a suitable statistical test to use in this scenario is the Chi-squared test for independence.” Despite previously pointing out the unreasonableness of using a chi-square test, the subsequent answer given contradicted this earlier response. Scans of the responses provided by ChatGPT are included in a supplement file (Supplementary Materials).

4. Discussion

4.1. ChatGPT’s Statistical Skills

The present research is, to the best of my knowledge, the first on ChatGPT’s statistical skills. Using the medical field of allergology as an example, including articles published in Allergy, a study was carried out to analyse the accuracy of ChatGPT’s answers to questions related to statistical analysis in a broad sense. Unfortunately, the results obtained indicate that out of 40 possible questions, only half were correctly answered, which had to do with the lack of specification of questions in which assumptions such as sample size, measurement scale of analysed variables, etc., had to be taken into account. A review article published in Health and Technology pointed out that the use of artificial intelligence tools can help to facilitate comprehensive statistical analysis, including troubleshooting missing data [14]. ChatGPT can play a significant role in conducting statistical analyses, although without human oversight of every stage of a piece of research, it is not possible to adequately report the results of the research [15]. A study on how AI-based transformers can help with the design of epidemiological studies also indicated, as this manuscript does, that ChatGPT can give methodologically flawed answers. The lowest scores were recorded for this section [16]. An interesting article was recently published in Schizophrenia with the following title: “ChatGPT: these are not hallucinations—they’re fabrications and falsifications”. Emsley R. pointed out that questions on general methodology produced general and down-to-earth answers. Difficulties were encountered in obtaining slightly more detailed answers related to a statistical analysis plan [17]. This confirms the results of the studies conducted in this work. At present, unfortunately, the quality of statistical reporting, interpretation and presentation of data in scientific articles is poor [18]. One recently published survey indicated that only 39% of accepted articles on various aspects of COVID-19 met the requirements of statistical validity [19]. An increasing number of journals are publishing statistical recommendations; however, it is difficult to know at this point whether such articles are improving reporting standards [20,21,22]. The research conducted in this work indicated that the main errors associated with the advice given related to the failure to consider a number of assumptions when selecting specific statistical tests. One of the main mistakes most frequently made in journals is the application of parametric statistical tests to non-parametric data [19,23]. The ordinal scale of the variables analysed, low size and unequal nature of the groups of individuals compared, and disorders of normality of distribution are all indications for the use of non-parametric equivalents of statistical tests [24,25]. Unfortunately, using the example of many questions related to the medical field of allergology, the advice provided by ChatGPT did not take such assumptions into account, which, with the low level of statistical analysis being performed today, may unfortunately exacerbate the problem instead of helping. According to the recommendations indicated in PLOS Biology, it is now generally recommended to provide practical instruction in performing statistical analyses, e.g., using examples of published articles in which specific statistical tests have been applied [8]. One published comparison of student performance in essay writing with or without the use of ChatGPT-3 indicated that the use of this type of tool did not play a major role in this task. This was related to an increased risk of plagiarism, a lack of originality and a lack of acceleration in essay writing [26]. An article published at the end of 2022 in Nature pointed out researchers’ concerns about students writing homework with the help of artificial intelligence [27]. The non-practical way in which statistical analysis is taught today, coupled with ChatGPT’s misguided advice on the subject, may unfortunately result in the implementation of inappropriate solutions based on published articles in the future. The responses obtained from ChatGPT also indicated incorrect measures of effect size in specific commands. Among other things, ChatGPT did not take into account the number of levels at which a particular variable was measured, nor its scale, and these are among the factors that play a role in the selection of an appropriate parameter [28]. ChatGPT’s indication of an inappropriate measure of effect size may also be a cause of misinterpretation of the research results obtained. It is noteworthy that, in the survey conducted, ChatGPT gave divergent answers to the same question related to which statistical test should be used. This made it all the more surprising to see that ChatGPT, after initially analysing the lack of validity of using a particular statistical test, indicated the same test about which it had expressed the opposite opinion a moment earlier, when presented with the question a second time. This confirms information published previously indicating that ChatGPT displays different answers when repeatedly presented with the same prompt [29]. Asking the same question may result in different answers [30]. A command such as “regenerate answer” may lead to different conclusions [31]. The provision of different answers by ChatGPT, combined with inappropriate advice related to the conduct of statistical analysis, may unfortunately result in a decrease in the public’s confidence in science. The results of a survey of adults in the United States indicated that over-reliance on ChatGPT can contribute to the spread of misinformation and subsequent risks to our health. This risk may be related to over-reliance on chatbots [32]. This manuscript is about the statistical skills of ChatGPT. However, alternative artificial intelligence tools such as BARD by Google are recommended for the future. A comparison with other artificial intelligence models would allow the reliability of the research to be assessed. In other words, the use of multiple artificial intelligence tools to answer the same questions related to conducting a statistical analysis would illustrate the existing consistency and variability in the recommendations given.

4.2. Statistical Recommendations Related to the Use of ChatGPT

Based on the research conducted, statistical recommendations related to the use of ChatGPT are indicated below. An initial recommendation for authors wishing to use ChatGPT’s indicated answers to questions related to statistical analysis in the broadest sense concerns the initial analysis of the assumptions necessary to apply specific tests. Before the authors decide, for example, to apply a statistical test indicated by ChatGPT, they should first examine the individual assumptions. Assumptions such as normality of distribution, group size, equality of groups, the scale on which the variable being analysed is measured, etc., must be taken into account. ChatGPT should not be asked general questions related to what kind of analysis should be applied, and the highlighted assumptions should be considered. In other words, it is advisable to ask more precise questions about whether 16 samples are enough to apply an analysis, or, for example, whether the measurement scale on which a variable is measured allows the application of a particular test. Without a specific question, the percentage of statistical advice provided by ChatGPT that is appropriate may be low. The second recommendation is to ask the same ChatGPT question several times because, as indicated in this manuscript, there is the possibility that it will provide quite divergent answers and thus subsequently suggest the wrong advice. Once different answers to a single statistical question have been obtained, authors should analyse these in depth, choose the most optimal answer to the best of their knowledge and then, if there are further doubts, relate the chosen answer to the relevant references. One option is also to refer to published statistical recommendations in medical journals such as those published by this journal, among others [33]. This is another such recommendation. In these times of developing artificial intelligence, it is all the more advisable to include statistical guidance on the websites of medical journals, or for recommendations to be published in said journals. Before authors submit a manuscript to a particular journal, they should first familiarise themselves with these recommendations. It should also be noted that authors who have used ChatGPT may not specifically have included this information in the submitted manuscript. A review by the authors of guidance obtained in this way should take place before submission of a manuscript and be indicated in the cover letter. The next recommendation is to forward to an expert in biostatistics any statistical analyses carried out for which the authors have applied advice from ChatGPT. Information about the statistical review carried out by this expert should also be included in the cover letter. The last recommendation has to do with presenting at major scientific conferences/congresses on the reliability of the information provided by ChatGPT related to statistical analysis in the broadest sense. These include major events attended by members of the International Committee of Medical Journal Editors and the World Association of Medical Editors. Making the scientific community aware of the most common possible errors in statistical advice provided by ChatGPT could contribute to even greater attention being paid to the aspects discussed in this manuscript.

5. Conclusions

At a time when there is a lower level of statistical analysis being conducted, we should have limited confidence in the advice provided by ChatGPT related to conducting statistical analyses.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/healthcare11182554/s1, Answers Provided by ChatGPT.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The author has no conflict of interest to declare.

References

  1. Flanagin, A.; Kendall-Taylor, J.; Bibbins-Domingo, K. Guidance for Authors, Peer Reviewers, and Editors on Use of AI, Language Models, and Chatbots. JAMA 2023, 330, 702–703. [Google Scholar] [CrossRef]
  2. Zielinski, C.; Winker, M.A.; Aggarwal, R.; Ferris, L.E.; Heinemann, M.; Lapeña, J.F., Jr.; Pai, S.A.; Ing, E.; Citrome, L.; Alam, M.; et al. Chatbots, Generative AI, and Scholarly Manuscripts. WAME Recommendations on Chatbots and Generative Artificial Intelligence in Relation to Scholarly Publications. WAME. 2023. Available online: https://wame.org/page3.php?id=106 (accessed on 1 August 2023).
  3. Jutel, M.; Zemelka-Wiacek, M.; Ordak, M.; Pfaar, O.; Eiwegger, T.; Rechenmacher, M.; Akdis, C.A. The artificial intelligence (AI) revolution: How important for scientific work and its reliable sharing. Allergy 2023, 78, 2085–2088. [Google Scholar] [CrossRef]
  4. Kassab, J.; Nasr, L.; Gebrael, G.; El Helou, M.C.; Saba, L.; Haroun, E.; El Dahdah, J.; Nasr, F. AI-based online chat and the future of oncology care: A promising technology or a solution in search of a problem? Front. Oncol. 2023, 13, 1176617. [Google Scholar] [CrossRef] [PubMed]
  5. Hoch, C.C.; Wollenberg, B.; Lüers, J.C.; Knoedler, S.; Knoedler, L.; Frank, K.; Cotofana, S.; Alfertshofer, M. ChatGPT’s quiz skills in different otolaryngology subspecialties: An analysis of 2576 single-choice and multiple-choice board certification preparation questions. Eur. Arch. Otorhinolaryngol. 2023, 280, 4271–4278. [Google Scholar] [CrossRef] [PubMed]
  6. Ioannidis, J.P. Why most published research findings are false. PLoS Med. 2005, 2, e124. [Google Scholar] [CrossRef]
  7. Hardwicke, T.E.; Goodman, S.N. How often do leading biomedical journals use statistical experts to evaluate statistical methods? The results of a survey. PLoS ONE 2020, 15, e0239598. [Google Scholar] [CrossRef]
  8. Weissgerber, T.L.; Garovic, V.D.; Milin-Lazovic, J.S.; Winham, S.J.; Obradovic, Z.; Trzeciakowski, J.P.; Milic, N.M. Reinventing Biostatistics Education for Basic Scientists. PLoS Biol. 2016, 14, e1002430. [Google Scholar] [CrossRef]
  9. Diver, S.; Haldar, K.; McDowell, P.J.; Busby, J.; Mistry, V.; Micieli, C.; Brown, V.; Cox, C.; Yang, F.; Borg, C.; et al. Medical Research Council: Refractory Asthma Stratification Programme (RASP-UK Consortium). Relationship between inflammatory status and microbial composition in severe asthma and during exacerbation. Allergy 2022, 77, 3362–3376. [Google Scholar] [CrossRef]
  10. Lee, S.E.; Hopkins, C.; Mullol, J.; Guillemin, I.; Amin, N.; Mannent, L.P.; Li, Y.; Siddiqui, S.; Chuang, C.; Kamat, S.; et al. Dupilumab improves health related quality of life: Results from the phase 3 SINUS studies. Allergy 2022, 77, 2211–2221. [Google Scholar] [CrossRef]
  11. Kallen, E.J.J.; Revers, A.; Fernández-Rivas, M.; Asero, R.; Ballmer-Weber, B.; Barreales, L.; Belohlavkova, S.; de Blay, F.; Clausen, M.; Dubakiene, R.; et al. A European-Japanese study on peach allergy: IgE to Pru p 7 associates with severity. Allergy 2023, 78, 2497–2509. [Google Scholar] [CrossRef]
  12. Le Floc’h, A.; Allinne, J.; Nagashima, K.; Scott, G.; Birchard, D.; Asrat, S.; Bai, Y.; Lim, W.K.; Martin, J.; Huang, T.; et al. Dual blockade of IL-4 and IL-13 with dupilumab, an IL-4Rα antibody, is required to broadly inhibit type 2 inflammation. Allergy 2020, 75, 1188–1204. [Google Scholar] [CrossRef] [PubMed]
  13. Berjont, N.; Floch, V.V.B.; O’Hehir, R.E.; Canonica, W.G.; van Zelm, M.C.; Batard, T.; Mascarell, L. Early increase in serum specific IgG2 upon allergen immunotherapy with a 300 IR sublingual house dust mite tablet. Allergy 2023, 78, 2040–2043. [Google Scholar] [CrossRef] [PubMed]
  14. Askin, S.; Burkhalter, D.; Calado, G.; El Dakrouni, S. Artificial Intelligence Applied to clinical trials: Opportunities and challenges. Health Technol. 2023, 13, 203–213. [Google Scholar] [CrossRef]
  15. Macdonald, C.; Adeloye, D.; Sheikh, A.; Rudan, I. Can ChatGPT draft a research article? An example of population-level vaccine effectiveness analysis. J. Glob. Health 2023, 13, 01003. [Google Scholar] [CrossRef] [PubMed]
  16. Sanmarchi, F.; Bucci, A.; Nuzzolese, A.G.; Carullo, G.; Toscano, F.; Nante, N.; Golinelli, D. A step-by-step researcher’s guide to the use of an AI-based transformer in epidemiology: An exploratory analysis of ChatGPT using the STROBE checklist for observational studies. Z. Gesundh. Wiss. 2023, 1–36. [Google Scholar] [CrossRef] [PubMed]
  17. Emsley, R. ChatGPT: These are not hallucinations—They’re fabrications and falsifications. Schizophr 2023, 9, 52. [Google Scholar] [CrossRef] [PubMed]
  18. Diong, J.; Butler, A.A.; Gandevia, S.C.; Héroux, M.E. Poor statistical reporting, inadequate data presentation and spin persist despite editorial advice. PLoS ONE 2018, 13, e0202121. [Google Scholar] [CrossRef]
  19. Ordak, M. COVID-19 research: Quality of biostatistics. Arch. Med. Sci. 2021, 18, 257–259. [Google Scholar] [CrossRef]
  20. Curran-Everett, D.; Benos, D.J. Guidelines for reporting statistics in journals published by the American Physiological Society: The sequel. Adv. Physiol. Educ. 2007, 31, 295–298. [Google Scholar] [CrossRef]
  21. Stevanovic, A.; Schmitz, S.; Rossaint, R.; Schürholz, T.; Coburn, M. CONSORT item reporting quality in the top ten ranked journals of critical care medicine in 2011: A retrospective analysis. PLoS ONE 2015, 10, e0128061. [Google Scholar] [CrossRef]
  22. Héroux, M. Inadequate reporting of statistical results. J. Neurophysiol. 2016, 116, 1536–1537. [Google Scholar] [CrossRef] [PubMed]
  23. Yim, K.H.; Nahm, F.S.; Han, K.A.; Park, S.Y. Analysis of statistical methods and errors in the articles published in the korean journal of pain. Korean J. Pain. 2010, 23, 35–41. [Google Scholar] [CrossRef] [PubMed]
  24. Schober, P.; Vetter, T.R. Nonparametric Statistical Methods in Medical Research. Anesth. Analg. 2020, 131, 1862–1863. [Google Scholar] [CrossRef] [PubMed]
  25. Politi, M.T.; Ferreira, J.C.; Patino, C.M. Nonparametric statistical tests: Friend or foe? J. Bras. Pneumol. 2021, 47, e20210292. [Google Scholar] [CrossRef] [PubMed]
  26. Bašić, Ž.; Banovac, A.; Kružić, I.; Jerković, I. Better by You, better than Me? ChatGPT-3 as writing assistance in students’ essays. arXiv 2023, arXiv:2302.04536. [Google Scholar] [CrossRef]
  27. Graham, F. Daily briefing: Will ChatGPT kill the essay assignment? Nature 2022. ahead of print. [Google Scholar] [CrossRef]
  28. Ordak, M. Multiple comparisons and effect size: Statistical recommendations for authors planning to submit an article to Allergy. Allergy 2023, 78, 1145–1147. [Google Scholar] [CrossRef]
  29. Hopkins, A.M.; Logan, J.M.; Kichenadasse, G.; Sorich, M.J. Artificial intelligence chatbots will revolutionize how cancer patients access information: ChatGPT represents a paradigm-shift. JNCI Cancer Spectr. 2023, 7, pkad010. [Google Scholar] [CrossRef]
  30. Alberts, I.L.; Mercolli, L.; Pyka, T.; Prenosil, G.; Shi, K.; Rominger, A.; Afshar-Oromieh, A. Large language models (LLM) and ChatGPT: What will the impact on nuclear medicine be? Eur. J. Nucl. Med. Mol. Imaging. 2023, 50, 1549–1552. [Google Scholar] [CrossRef]
  31. Beutel, G.; Geerits, E.; Kielstein, J.T. Artificial hallucination: GPT on LSD? Crit. Care 2023, 27, 148. [Google Scholar] [CrossRef]
  32. Choudhury, A.; Shamszare, H. Investigating the Impact of User Trust on the Adoption and Use of ChatGPT: Survey Analysis. J. Med. Internet Res. 2023, 25, e47184. [Google Scholar] [CrossRef] [PubMed]
  33. Ordak, M. Biostatistics in allergy—Recommendations for authors. Allergy 2022, 77, 3493–3495. [Google Scholar] [CrossRef] [PubMed]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.