Next Article in Journal
Venous Malformation of the Maxilla: A Systematic Review and a Report of an Unusual Case
Previous Article in Journal
Can Better Surgical Education Lead to the Improved Acquisition of Young Trauma Surgeons? A Prospective Survey of Medical Students Concerning the Impact of Teaching Quality on the Future Choice of Medical Discipline
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Artificial Intelligence Tools in Surgical Research: A Narrative Review of Current Applications and Ethical Challenges

1
Department of Plastic Surgery, Peninsula Health, Melbourne, VIC 3199, Australia
2
Department of Plastic Surgery, Royal Melbourne Hospital, Melbourne, VIC 3004, Australia
3
Plastic Surgery Unit, Department of Medicine, Surgery and Neuroscience, University of Siena, 53100 Siena, Italy
*
Author to whom correspondence should be addressed.
Surgeries 2025, 6(3), 55; https://doi.org/10.3390/surgeries6030055
Submission received: 23 April 2025 / Revised: 25 June 2025 / Accepted: 7 July 2025 / Published: 9 July 2025

Abstract

Background/Objectives: Artificial intelligence (AI) holds great potential to reshape the academic paradigm. They can process large volumes of information, assist in academic literature reviews, and augment the overall quality of scientific discourse. This narrative review examines the application of various AI tools in surgical research, its present capabilities, future directions, and potential challenges. Methods: A search was performed by two independent authors for relevant studies on PubMed, Cochrane Library, Web of Science, and EMBASE databases from January 1901 until March 2025. Studies were included if they were written in English and discussed the use of AI tools in surgical research. They were excluded if they were not in English and discussed the use of AI tools in medical research. Results: Forty-two articles were included in this review. The findings underscore a range of AI tools such as writing enhancers, LLMs, search engine optimizers, image interpreters and generators, content organization and search systems, and audio analysis tools, along with their influence on medical research. Despite the multitude of benefits presented by AI tools, risks such as data security, inherent biases, accuracy, and ethical dilemmas are of concern and warrant attention. Conclusions: AI could offer significant contributions to medical research in the form of superior data analysis, predictive abilities, personalized treatment strategies, enhanced diagnostic accuracy, amplified research, educational, and publication processes. However, to unlock the full potential of AI in surgical research, we must institute robust frameworks and ethical guidelines.

1. Introduction

Recent breakthroughs in artificial intelligence (AI), especially large language models (LLMs), have shown impressive potential in surgical research [1,2,3,4,5,6]. This includes synthesizing large swaths of data, understanding natural language, and recognizing patterns. Consequently, AI′s role in enhancing the quality of scientific literature has become crucial due to the ever-growing body of articles and data, challenging researchers to thoroughly review the literature. AI holds substantial potential to aid surgeons and are instrumental in facilitating rapid understanding and application of the literature [7].
The integration of AI tools into academic workflows can expedite scientific communication. AI writing tools facilitate the articulation of complex ideas with increased clarity and precision, making them more comprehensible to a broader audience [8]. Additionally, English being the prevailing language of scientific communication can disadvantage those with limited proficiency, leading to underrepresentation of their ideas. AI tools can help bypass these linguistic barriers, aiding effective translation for researchers and promoting greater contribution.
This study synthesizes the current literature regarding the integration of AI in surgical research. By exploring and summarizing how these tools assist in research, we will provide insight into their future contributions and highlight potential challenges when integrating AI into surgical academia.

2. Materials and Methods

The primary aim is to conduct a narrative review of the application of artificial intelligence in surgery, focusing on identifying diverse AI tools that have advanced medical research. The secondary aim is to investigate the extent to which widely used AI tools enhance medical research by comparing findings from AI-intervention studies with those conducted solely by human researchers.
Two independent authors (BL and IS) searched for relevant studies on PubMed, Cochrane Library, Web of Science, and EMBASE from January 1901 until March 2025. The search terms included, but were not limited to, “Artificial Intelligence,” “AI,” “machine learning”, “large language models”, “LLMs”, “surgical research”, “scientific writing”, “grammar enhancement”, “content generation”, “data organization”, “image interpretation”, and “content search”. These terms were used in various combinations using Boolean operators “AND” and “OR”. Google and Bing search engines were also searched to uncover relevant articles, blogs, and reports discussing popular AI tools in surgical research. This ensured the incorporation of non-academic sources, which, while not always peer-reviewed, may offer practical insight into the use of AI tools in real-world settings. Articles were included if they were published in English, peer-reviewed, and focused on the use of artificial intelligence tools in surgical research, including preclinical, clinical, or outcome-based studies. Eligible studies applied AI methods such as machine learning (ML) or natural language processing to surgical datasets or research workflows. Exclusion criteria included non-English language, non-surgical focus, editorials, reviews without original data, and studies describing AI use solely for administrative purposes or clinical practice without a research objective. Commonly used AI tools in scientific research are depicted in Table 1, and compared to each other in Table 2.

3. Results

3.1. Writing Enhancement Tools

Writing enhancement technologies such as Trinka and Jenni, refine syntax and style, in addition to rectifying spelling mistakes [9]. They do so by proposing succinct structures for convoluted sentences, creating a more engaging and accessible article. They are especially beneficial for non-native English-speaking researchers by ameliorating various linguistic aspects of their work [10]. Osteosynthesis, a writing instrument by PubMed, is distinguished by its ability to generate in-depth literature reviews, systematic reviews, and exam questions tailored to specific user requests. If similar functionalities were incorporated into LLMs like ChatGPT, using databases such as Scopus and Cochrane, the process of conducting literature searches and generating systematic reviews could be remarkably expedited. Such a technological upgrade could catalyze faster breakthroughs and advancements. While these tools are beneficial, they should not supersede manual review due to potential limitations in discerning complex terminology or context-dependent phrasing.

3.2. Large Language Models

LLMs such as ChatGPT, Gemini, and the recently released Chinese-developed Deepseek AI, facilitate the process of article creation, assess medical records and images, predict potential complications, offer management recommendations, and promote research on various topics, thereby advancing surgical practices [1,2,3,4,5,6,11]. LLMs are increasingly integrated into surgery, reducing administrative burdens and enhancing research efficiency as surgeons can leverage LLMs for drafting emails, refining documents, and even generating letters of recommendation [12]. In surgical research, LLMs help in writing literature reviews, study design, and manuscript drafting, with studies showing AI-generated abstracts can closely resemble those from high-impact journals [13,14]. Despite concerns over academic integrity and AI-generated text policies, specialized biomedical LLMs such as BioBERT and Med7 demonstrate strong potential in medicine [15,16,17]. LLMs also facilitate hypothesis generation and experimental design in surgical research [18,19]. By analyzing large datasets and recognizing patterns across diverse sources, LLMs can propose novel research questions, optimize study methodologies, and even predict potential outcomes based on existing evidence [18,19,20]. With ongoing advancements in natural language processing, LLMs have the potential to enhance access to high-quality scientific insights, empowering both established researchers and early-career academics to contribute meaningfully to the evolving landscape of surgical innovation.

3.3. Search Engine Optimization Tools

Search Engine Optimization (SEO) tools such as Consensus and Elicit increase the visibility of articles by optimizing keywords, titles, click-through rates, dwell time, and meta descriptions, using AI to understand user behavior and suggest improvements. These suggestions refine content to better meet user expectations, increasing readership and accelerating the spread of medical knowledge. Additionally, the link analysis function of SEOs can potentially serve as a digital counterpart to citation analysis in academic literature [21]. They can track the digital footprint of research articles by identifying web pages, institutional repositories, and social media platforms that link to a particular study or academic profile. This broadens the perspective on research influence beyond traditional citation metrics, uncovering new dissemination pathways. Furthermore, AI-driven analytics can assess engagement metrics, such as reader interactions and geographic reach, providing insights into how surgical research is being consumed and discussed globally. While these tools offer valuable enhancements to research visibility and impact assessment, their findings should complement traditional academic metrics to ensure a comprehensive evaluation of research significance.

3.4. Image Interpretation and Generation Tools

AI image generators such as Dall-E3 and DermEngine analyze numerous image datasets of dermatological conditions [22]. They are also capable of monitoring changes in skin conditions, thereby facilitating more accurate, early detection and management strategies. JasperAI functions differently, capable of generating images based on textual input. Its potential could be instrumental to plastic surgeons and patients who are planning reconstructive or cosmetic surgery by generating digital approximations of the final surgical outcome based on patient data. V7, on the other hand, allows users to upload medical images and videos, subsequently employing ML algorithms to interpret the information. After which, it can identify and annotate a variety of anatomical structures in X-Ray and Computed Tomography (CT) scans. This amalgamation of image recognition technology and medicine represents a future characterized by improved patient outcomes, optimized healthcare workflows, and preventative care becoming more effective and widely accessible.

3.5. Organization and Content Search

Organization tools aim to streamline article searches by capturing relevant articles and helping researchers better organize their work. Lateral allows users to quickly search keywords, organize findings on a dashboard, and navigate directly to relevant sections, speeding up searches. ResearchRabbit can track citations, summarize papers, and illustrate connections on a tree graph, enhancing understanding of research relationships. ConnectedPapers uses Semantic Scholar to find and distill thousands of related articles into the most relevant ones. Scite and Iris process keywords against databases to extract relevant articles, with Iris also generating a research map to visually organize articles into topical clusters [23]. Perplexity, founded in 2022 by Aravind Srinivas—a former researcher at OpenAI and DeepMind—is a conversational search engine that combines AI chatbot capabilities with real-time web search, providing users with natural language responses accompanied by cited sources. These tools help users find the most relevant scholarly works, enriching their research.

3.6. Audio Analysis

AI tools present a transformative opportunity by harnessing ML algorithms to interpret audio data. Scribing tools like Fireflies and PowerScribe360 can understand human speech across a range of accents and transcribe them into text, which can streamline notetaking and facilitate team communication and rigorous data analysis. These tools ensure the preservation of critical dialog from medical interactions, maintaining accuracy in documentation which can be crucial for data collection. The resulting transcripts improve accessibility for those with hearing impairments or for whom English is not a primary language. More importantly, these tools support regulatory and ethical compliance through rigorous record-keeping in clinical research [24]. audEERING, another audio AI tool, innovatively diagnoses diseases like Parkinson′s and COVID-19 by analyzing patient audio data. However, automated services may not achieve complete precision, especially with intricate medical jargon or in instances of pronounced accents or ambient noise [25]. Confidentiality and data security are also major concerns, especially with sensitive medical data and third-party cloud services. Researchers must ensure tools comply with regulations like the Australian Privacy Principles or the European Union′s General Data Protection Regulation to safeguard participant confidentiality and data security.

3.7. Publication and Collaborative AI Tools

Publication and collaborative AI tools such as SciSpace provide a comprehensive platform to collaborate with other researchers, promote the dissemination of one’s work, track the impact of published papers, and comment on other scholarly articles. They also provide an updated list of journals actively accepting submissions and help reformat the researcher’s article according to their guidelines. This can reduce the time spent on formatting and journal selection, thereby streamlining the research process. AI-driven search platforms excel in systematically filtering through diverse scientific literature sources, ensuring comprehensive coverage and efficient data retrieval. AI tools and algorithms that can understand the contextual use of words significantly enhance the accuracy and relevance of search results [26]. Essential features, such as author name, query refinement, and automatic article indexing aid in generating concise summaries of pertinent studies. Notable examples of databases using such features include PubMed and Embase [27]. However, over-reliance on AI could result in the attrition of crucial research skills, underlining the importance of these tools as supplements to, rather than substitutes for, human expertise [28]. Data security and privacy concerns can also arise due to the extensive collection and storage of sensitive data.

3.8. Ontologies

Ontology-based AIs use structured representations of knowledge and relationships, provide a standardized vocabulary and capture relationships between concepts, allowing heterogeneous data to be unified [29,30]. This facilitates automated reasoning, enabling researchers to gain new insights from existing structured data. Ontologies also promote knowledge reuse and sharing across systems, reducing redundant effort [31]. They help organize guidelines, research findings, and terminology to support clinical decision-making [32]. They also allow linkage of diverse biomedical data to generate new hypotheses. Ontologies augment scientific research by making knowledge computable so that information systems can be interconnected and advanced analytics can be performed to accelerate discoveries, positioning them as a core enabling technology for augmenting scientific research across many domains [33].

3.9. Video Generation

The Sora model, a new video generator AI created by OpenAI, utilizes transformer architectures and unified visual data representations to interpret text instructions and generate detailed, accurate visual content [34]. It could allow researchers to create sophisticated visual representations of data and concepts, potentially revolutionizing how studies are communicated and understood. By facilitating the visualization of complex medical scenarios and research findings, Sora has the potential to enhance data presentation, communication, and explain research ideas more easily. Despite challenges such as potential inaccuracies due to the model′s text interpretation limitations, Sora′s ability to produce high-quality visual outputs from textual prompts offers promising avenues for advancing medical research capabilities [34]. Further exploration is needed to fully explore and optimize Sora′s impact in this critical domain.

4. Discussion

AI tools have augmented medical research, driving efficiency, accuracy, and innovation [35]. They can process vast amounts of data more efficiently than humans, identify patterns and predict outcomes that exceed human capabilities, accelerating data analysis and revealing hidden insights that are otherwise laborious to uncover [36,37]. AI tools can conduct literature reviews, saving researchers a significant amount of time by sieving through thousands of relevant studies [38]. Audio analysis is also progressively influenced by AI. Transcribing tools like FireFly enable radiologists to produce radiographic reports faster, and audEERING spearheads innovative approaches to diagnosing pathologies based on auditory data, potentially advancing scientific research, fostering innovation, and enhancing communication.
LLMs such as ChatGPT have demonstrated potential in improving writing efficiency, with studies showing that AI-assisted abstracts and lay summaries are often rated as comparable in quality to those written by human researchers, particularly for first drafts or non-native English speakers [39,40,41]. In comparative studies, tools like Elicit and Scite have outperformed traditional search engines in generating relevant literature suggestions, aiding hypothesis formation, and streamlining systematic review protocols [42,43]. While these cases reflect growing interest and early success, robust empirical studies using standardized outcome measures remain essential to formally evaluate long-term efficacy, reproducibility, and integration into surgical research practice.
However, AI tools face potential barriers to integration such as patient privacy, data security, accuracy, reliability, and potential biases in training data. Addressing these is crucial for their ethical and effective use as many AI tools rely on ML and Deep Learning, requiring access to extensive datasets. However, institutions hesitate to share confidential patient records due to security, privacy concerns, and insufficient patient consent [44,45,46], limiting accessibility for ML and DL training. Certain AI tools such as ChatGPT lack access to the internet limiting them to a fixed database for responses and this can result in less relevant answers, necessitating database updates [46]. Therefore, a critical need exists for innovative solutions that uphold data privacy without inhibiting the progression of AI [47].
Another concern involves inadequate data from marginalized or misrepresented populations for training AI tools using ML and DL, which can lead to inaccurate health predictions and epidemiological trends [48,49,50,51,52,53,54,55]. For example, tools trained primarily on data from one geographic or ethnic group may not accurately reflect disease patterns or health conditions in other populations. These disparities can inadvertently bias research findings and medical hypotheses. Without careful dataset selection, AI systems could reinforce historical biases, similar to those in 1900 San Francisco′s Chinatown or apartheid-era South Africa [56]. The “Black Box” problem, where AI algorithms are opaque, can worsen this issue [57,58,59]. Using diverse, generalizable training datasets and developing stereotype-resistant neural networks could alleviate this issue, but their effectiveness is still unproven [60].
Plagiarism remains a significant concern as LLMs may reproduce content from training data without proper citation [61]. This is compounded by difficulties in tracing the origins of AI-generated content. It underscores the need for robust strategies to monitor and mitigate plagiarism risks in scientific writing and research when using AI tools. Additionally, inaccuracies such as generating non-existent references, termed “AI Hallucination”, can compromise the validity of outputs [61]. For instance, ChatGPT-4, PaLM 2, and Llama 2 generated blogs with false health claims, such as sunscreen causing skin cancer and the alkaline diet curing cancer [62]. Consequently, when AI is used to summarize evidence, they are still prone to errors. To address content suspected of being inaccurate, consultation with experts or their work is often necessary [63]. If an AI-generated hallucination is included in a published article, and subsequent articles continue to cite that original source, it can create a “snowball” effect that is difficult to correct. This therefore raises the question of who or what sources should be consulted if the experts and authoritative resources were educated using the same AI algorithms.
Image generation is becoming a popular topic for medical applications. Sora, an upcoming video generation software by OpenAI, exemplifies the immense potential of such AI tools in augmenting clinical practice and academic research [34]. However, many barriers remain, particularly concerning the potential for misinformation. AI-generated videos can be exploited by malicious actors, confusing patients, undermining trust in medical professionals, and leading to harmful treatment decisions. Such misinformation can compromise medical research by introducing false data and misleading results [56]. Despite OpenAI′s efforts to mitigate misuse by employing a text classifier to check text prompts and developing tools to detect AI-generated videos, such as the inclusion of C2PA metadata, these measures have limitations. Metadata can be easily removed, and most social media platforms automatically strip metadata, rendering these protections ineffective against misinformation [64]. Therefore, combating potential health misinformation from Sora-generated videos will require a multifaceted approach, including robust verification mechanisms, government regulation, and patient education. Waisberg et al.′s letter to the editor underscores this risk by highlighting that during the COVID-19 pandemic, misleading or inaccurate information was found in nearly one in four of the most popular COVID-19 YouTube videos [64].
Despite their sophistication, AI tools may have difficulty distinguishing relevant data from irrelevant data. To ensure data validity, information generated by models should be subjected to rigorous validation by experienced surgeons and researchers. Ethical considerations are crucial when using AI tools like ChatGPT for research content creation, necessitating disclosure of using AI during submission. Furthermore, the validity of information from AI tools can be outdated, requiring users to verify ChatGPT′s recommendations against current and reliable guidelines for up-to-date and relevant advice.

4.1. Limitations

This study has several limitations: a lack of systematic literature search methodology may introduce selection bias, and the absence of quality assessment for included studies could lead to overgeneralization from a small sample size. Additionally, there is no quantitative evaluation of AI tools using metrics like “effectiveness” and “efficacy” in data analysis. These terms are primarily conceptualized here in terms of speed and precision, but broader factors like purpose, rigor, reproducibility, and impact were not rigorously assessed, potentially hindering replication.
Given that this review examined AI tools rather than primary research studies, formal quality appraisal frameworks were not applicable. Nonetheless, we sought to provide a structured assessment by outlining each tool’s function and accessibility (Table 1), thereby offering insight into their potential utility and limitations in surgical research. Future studies should aim to quantitatively evaluate these tools using standardized metrics to more rigorously determine their impact on research quality and outcomes.
Despite these limitations, this study lays a foundational understanding of AI tools in medical research, paving the way for future comprehensive reviews.

4.2. The Future of AI in Medical Research

Despite their drawbacks, AI is poised to advance surgical research. They can analyze vast datasets using ML and DL to detect trends and patterns, thereby facilitating accurate predictions and revealing new disease insights. This capability extends to biopharmaceuticals where AI models utilize convolutional neural networks to predict drug binding affinity, accelerating more effective research into medications [65,66,67,68]. Additionally, AI can develop personalized treatments by analyzing patients′ medical history and relevant data, enhancing treatment success, and minimizing complications [65,66,67,68]. Their increasing proficiency in utilizing visual and auditory data positions them as ideal supplementary tools for clinicians, researchers, and patients alike.
Ethical concerns in AI use are essential. Frameworks like the Development Plan for the New Generation of Artificial Intelligence should be created and applied for oversight [69]. New regulations are crucial for ensuring that AI systems are technically robust, private, and safe, with harmonized safeguards. Human oversight is still needed to maintain meaningful involvement and prevent bias, while transparency is vital for trust and reproducibility [70,71]. Diverse datasets are necessary to avoid discrimination, and accountability mechanisms must be established [72]. We must also consider the practical and ethical risks of using limited AI tools to avoid generating discriminatory or harmful outputs [73,74]. Rashidian et al. found that AI holds significant promise for enhancing outcomes in Hepato-Pancreato-Biliary (HPB) surgery, yet ethical considerations and the technology’s reliability remain underexplored, as highlighted by an E-AHPBA survey showing high overall agreement on regulatory needs but mixed views on addressing bias and promoting equity [75]. These findings suggest that while HPB surgeons recognize the importance of oversight, greater clarity and consensus are needed around AI’s role in ensuring fairness in surgical care [75].
The integration of AI into surgery offers tremendous potential but raises complex medico-legal and ethical challenges. Treger et al. postulates that surgeons must continue to meet evolving standards of care, carefully document AI involvement in clinical decisions, and maintain transparency regarding their rationale for accepting or rejecting AI advice [76]. Similarly, developers must ensure the safety and reliability of their tools, including the communication of uncertainty in AI-generated recommendations. Without clear standards, there is a risk of malpractice claims and erosion of trust in AI applications. Early legislative efforts in the United States, such as the proposed but unsuccessful Healthy Technology Act of 2023 and President Biden’s Executive Order #14110, reflect growing federal attention to AI governance, emphasizing privacy, equity, and human oversight [77]. Recent guidance from the Centers for Medicare and Medicaid Services mandates human review in key clinical decisions, signaling an increasingly cautious regulatory approach to AI in healthcare [78].
Capelli et al. also discussed the growing role of AI and AI-powered robotics in surgery, emphasizing the need for trust and rigorous assessment to ensure patient safety [79]. They highlight the challenge of data security and privacy, especially with the increasing digitalization of healthcare data and the complexities of AI′s access to vast datasets. They stress the importance of minimizing bias in AI systems, which can unintentionally perpetuate societal inequalities [80,81]. The authors also address the transparency issues in AI algorithms, particularly the “black box” effect, and the need for clear explanations of AI-driven decisions [82,83]. Lastly, they suggest that AI can help address disparities in surgical training and healthcare accessibility, particularly in under-resourced regions.
Advancements in AI are transforming medical writing by improving natural language processing accuracy, enabling real-time data integration, and increasing transparency through explainable AI [84]. While AI enhances efficiency and reduces costs in data processing, thorough economic evaluations are necessary to determine its broader impact. With AI managing routine tasks, medical writers will focus more on strategic decision-making, ethical considerations, and advanced editing, making it essential to maintain a balance through training and well-defined intervention strategies [85]. Key best practices involve integrating AI within ethical frameworks, utilizing it for literature synthesis, and fostering collaboration between human expertise and AI capabilities. AI tools in surgical research are increasingly supporting robotic surgery and real-time intraoperative decision-making by integrating multimodal data analysis [86,87]. During surgery, AI could analyze real-time patient data, such as vital signs and imaging, to provide immediate feedback and suggest optimal approaches by conducting immediate literature searches on the latest evidence. Pre-operatively, AI can scan the latest literature and clinical guidelines to recommend the most effective treatment strategies, ensuring that the surgical team is equipped with the best possible approach before entering the operating room. This continuous integration of real-time data and evidence enhances decision-making and overall surgical outcomes.
By streamlining literature reviews and draft creation, AI supports medical writers in producing high-quality, patient-focused content while ensuring that critical analysis and ethical judgment remain firmly in human hands, ultimately strengthening healthcare communication and research [85].

5. Conclusions

AI tools are transforming surgical academia by enhancing efficiency, accuracy, and innovation in medical research, with future advancements expected in natural language processing, predictive modeling, and multimodal data analysis. These tools hold promise for automating literature reviews, analyzing large-scale clinical datasets, generating hypothesis-driven insights, and streamlining manuscript preparation. However, addressing challenges like data privacy, bias, and ethical concerns is crucial, requiring rigorous frameworks to ensure responsible AI integration while preserving human oversight in scientific discovery and healthcare communication.

Author Contributions

Conceptualization, B.L., I.S., F.S., R.C., and W.M.R.; methodology, B.L., I.S., F.S., R.C., and W.M.R.; formal analysis, B.L., I.S., and J.C.; investigation, B.L. and I.S.; data curation, B.L. and I.S.; writing—original draft preparation, B.L., I.S., J.C., and X.M.; writing—review and editing, B.L., I.S., J.C., X.M., F.S., R.C., and W.M.R.; supervision, F.S., R.C. and W.M.R.; project administration, B.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All available data is present in this manuscript.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

AIArtificial Intelligence
COVID-19Coronavirus Disease 2019
LLMLarge language model
SEOSearch Engine Optimization

References

  1. Cox, A.; Seth, I.; Xie, Y.; Hunter-Smith, D.J.; Rozen, W.M. Utilizing ChatGPT-4 for Providing Medical Information on Blepharoplasties to Patients. Aesthet. Surg. J. 2023, 43, NP658–NP662. [Google Scholar] [CrossRef] [PubMed]
  2. Park, Y.J.; Pillai, A.; Deng, J.; Guo, E.; Gupta, M.; Paget, M.; Naugler, C. Assessing the research landscape and clinical utility of large language models: A scoping review. BMC Med. Inform. Decis. Mak. 2024, 24, 72. [Google Scholar] [CrossRef] [PubMed]
  3. Meng, X.; Yan, X.; Zhang, K.; Liu, D.; Cui, X.; Yang, Y.; Zhang, M.; Cao, C.; Wang, J.; Wang, X. The application of large language models in medicine: A scoping review. iScience 2024, 27, 109713. [Google Scholar] [CrossRef] [PubMed]
  4. Goh, E.; Gallo, R.; Hom, J.; Strong, E.; Weng, Y.; Kerman, H.; Cool, J.A.; Kanjee, Z.; Parsons, A.S.; Ahuja, N. Large Language Model Influence on Diagnostic Reasoning: A Randomized Clinical Trial. JAMA Netw. Open 2024, 7, e2440969. [Google Scholar] [CrossRef]
  5. Shah, N.H.; Entwistle, D.; Pfeffer, M.A. Creation and Adoption of Large Language Models in Medicine. JAMA 2023, 330, 866–869. [Google Scholar] [CrossRef]
  6. Li, Z.; Li, F.; Fu, Q.; Wang, X.; Liu, H.; Zhao, Y.; Ren, W. Large language models and medical education: A paradigm shift in educator roles. Smart Learn. Environ. 2024, 11, 26. [Google Scholar] [CrossRef]
  7. Patel, S.B.; Lam, K.; Liebrenz, M. ChatGPT: Friend or Foe. Lancet Digit. Health 2023, 5, e102. [Google Scholar] [CrossRef]
  8. ONeill, R.; Russell, A. Stop! Grammar time: University students’ perceptions of the automated feedback program Grammarly. Australas. J. Educ. Technol. 2019, 35, 42–56, AJET. [Google Scholar] [CrossRef]
  9. Ananda, M.; Nisa, R.; Safura, S. Students’ perceptions toward the use of Grammarly in checking grammar in assignments. Accent. J. Engl. Lang. Educ. 2021, 1, 72–77. [Google Scholar] [CrossRef]
  10. Ngo, T.T.-N.; Chen, H.H.-J.; Lai, K.K.-W. The effectiveness of automated writing evaluation in EFL/ESL writing: A three-level meta-analysis. Interact. Learn. Environ. 2024, 32, 727–744. [Google Scholar] [CrossRef]
  11. ElHawary, H.; Gorgy, A.; Janis, J. Large Language Models in Academic Plastic Surgery: The Way Forward. PRS Glob Open 2023, 11, e4949. [Google Scholar] [CrossRef] [PubMed]
  12. Rengers, T.A.; Thiels, C.A.; Salehinejad, H. Academic Surgery in the Era of Large Language Models. JAMA Surg. 2024, 159, 445–450. [Google Scholar] [CrossRef] [PubMed]
  13. Carobene, A.; Padoan, A.; Cabitza, F.; Banfi, G.; Plebani, M. Rising adoption of artificial intelligence in scientific publishing: Evaluating the role, risks, and ethical implications in paper drafting and review process. Clin. Chem. Lab. Med. 2023, 62, 835–843. [Google Scholar] [CrossRef] [PubMed]
  14. Sanmarchi, F.; Bucci, A.; Nuzzolese, A.G.; Carullo, G.; Toscano, F.; Nante, N.; Golinelli, D. A step-by-step researcher’s guide to the use of an AI-based transformer in epidemiology: An exploratory analysis of ChatGPT using the STROBE checklist for observational studies. Z. Gesundh. Wiss. 2023, 26, 1761–1796. [Google Scholar] [CrossRef]
  15. Lee, J.; Yoon, W.; Kim, S.; Kim, D.; Kim, S.; So, C.H.; Kang, J. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 2020, 36, 1234–1240. [Google Scholar] [CrossRef]
  16. Kormilitzin, A.; Vaci, N.; QiangLiu, Y.; Nevado-Holgado, A. Med7: A transferable clinical natural language processing model for electronic health recordsArtif. Intell. Med. 2021, 118, 102086. [Google Scholar] [CrossRef]
  17. Luo, R.; Sun, L.; Xia, Y.; Qin, T.; Zhang, S.; Poon, H.; Liu, T.Y. BioGPT: Generative pre-trained transformer for biomedical text generation and mining. Brief Bioinform. 2022, 23, bbac409. [Google Scholar] [CrossRef]
  18. Janssen, B.V.; Kazemier, G.; Besselink, M.G. The use of ChatGPT and other large language models in surgical science. BJS Open 2023, 7, zrad032. [Google Scholar] [CrossRef]
  19. Ray, P.P. Large language models in laparoscopic surgery: A transformative opportunity. Laparosc. Endosc. Rob. Surg. 2024, 7, 174–180. [Google Scholar] [CrossRef]
  20. Ray, P.P. ChatGPT: A comprehensive review on background, applications, key challenges, bias, ethics, limitations, and future scope. Internet Things Cyber-Phys. Syst. 2023, 3, 121–154. [Google Scholar] [CrossRef]
  21. Bornmann, L. Do altmetrics point to the broader impact of research? An overview of benefits and disadvantages of altmetrics. J. Informetr. 2014, 8, 895–903. [Google Scholar] [CrossRef]
  22. Kvak, D.; Březinová, E.; Biroš, M.; Hrubý, R. Synthetic Data as a Tool to Combat Racial Bias in Medical AI: Utilizing Generative Models for Optimizing Early Detection of Melanoma in Fitzpatrick Skin Types IV–VI. In International Conference on Medical Imaging and Computer-Aided Diagnosis, Leicester, UK, 20–21 November 2024; Lecture Notes in Electrical Engineering; Springer Nature: Singapore, 2022; pp. 317–330. [Google Scholar]
  23. Schoeb, D.; Suarez-Ibarrola, R.; Hein, S.; Dressler, F.F.; Adams, F.; Schlager, D.; Miernik, A. Use of Artificial Intelligence for Medical Literature Search: Randomized Controlled Trial Using the Hackathon Format. Interact. J. Med. Res. 2020, 9, e16606. [Google Scholar] [CrossRef] [PubMed]
  24. Davidson, C. Transcription: Imperatives for qualitative research. Int. J. Qual. Methods 2009, 8, 35–52. [Google Scholar] [CrossRef]
  25. Al-Aynati, M.M.; Chorneyko, K.A. Comparison of voice-automated transcription and human transcription in generating pathology reports. Arch. Path Lab. Med. 2003, 127, 721–725. [Google Scholar] [CrossRef]
  26. de la Torre-López, J.; Ramírez, A.; Romero, J.R. Artificial intelligence to automate the systematic review of scientific literature. Computing 2023, 105, 2171–2194. [Google Scholar] [CrossRef]
  27. Jin, Q.; Leaman, R.; Lu, Z. PubMed and beyond: Biomedical literature search in the age of artificial intelligence. eBioMed. 2024, 100, 104988. [Google Scholar] [CrossRef]
  28. Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef]
  29. Filice, R.W.; Kahn, C.E. Biomedical Ontologies to Guide AI Development in Radiology. J. Digit. Imaging 2021, 34, 1331–1341, [published correction appears in J. Digit. Imaging 2022, 35, 1419. https://doi.org/10.1007/s10278-021-00568-6]. [Google Scholar] [CrossRef]
  30. Sadeghi-Ghyassi, F.; Damanabi, S.; Kalankesh, L.R.; Van de Velde, S.; Feizi-Derakhshi, M.R.; Hajebrahimi, S. How are ontologies implemented to represent clinical practice guidelines in clinical decision support systems: Protocol for a systematic review. Syst. Rev. 2022, 11, 183. [Google Scholar] [CrossRef]
  31. Jing, X.; Min, H.; Gong, Y.; Biondich, P.; Robinson, D.; Law, T.; Nohr, C.; Faxvaag, A.; Rennert, L.; Hubig, N.; et al. Ontologies Applied in Clinical Decision Support System Rules: Systematic Review. JMIR Med. Inform. 2023, 11, e43053. [Google Scholar] [CrossRef]
  32. Moghadam, S.T.; Hooman, N.; Sheikhtaheri, A. Patient safety classifications, taxonomies and ontologies: A systematic review on development and evaluation methodologies. J. Biomed. Inform. 2022, 133, 104150. [Google Scholar] [CrossRef]
  33. Okikiola, F.M.; Ikotun, A.M.; Adelokun, A.P.; Ishola, P.E. A Systematic Review of Health Care Ontology. Asian J. Res. Comput. Sci. 2020, 5, 15–28. [Google Scholar] [CrossRef]
  34. Sora: Creating Video from Text. Available online: https://openai.com/sora (accessed on 1 February 2025).
  35. Adams, D.; Chuah, K.-M. Artificial Intelligence-Based Tools in Research Writing: Current Trends and Future Potentials. In Artificial Intelligence in Higher Education; CRC Press: Boca Raton, FL, USA, 2022; pp. 169–184. [Google Scholar] [CrossRef]
  36. Tai, M.C. The impact of artificial intelligence on human society and bioethics. Tzu Chi Med. J. 2020, 32, 339–343. [Google Scholar] [CrossRef] [PubMed]
  37. Eriksen, A.V.; Möller, S.; Ryg, J. Use of GPT-4 to Diagnose Complex Clinical Cases. NEJM AI 2023, 1, AIp2300031. [Google Scholar] [CrossRef]
  38. Golan, R.; Reddy, R.; Muthigi, A.; Ramasamy, R. Artificial intelligence in academic writing: A paradigm-shifting technological advance. Nat. Rev. Urol. 2023, 20, 327–328. [Google Scholar] [CrossRef]
  39. Rinderknecht, E.; Schmelzer, A.; Kravchuk, A.; Goßler, C.; Breyer, J.; Gilfrich, C.; Burger, M.; Engelmann, S.; Saberi, V.; Kirschner, C. Leveraging Large Language Models for High-Quality Lay Summaries: Efficacy of ChatGPT-4 with Custom Prompts in a Consecutive Series of Prostate Cancer Manuscripts. Curr. Oncol. 2025, 32, 102. [Google Scholar] [CrossRef]
  40. Meyer, J.G.; Urbanowicz, R.J.; Martin, P.C.N.; O’Connor, K.; Li, R.; Peng, P.C.; Bright, T.J.; Tatonetti, N.; Won, K.J.; Gonzalez-Hernandez, G. ChatGPT and large language models in academia: Opportunities and challenges. BioData Min. 2023, 16, 20. [Google Scholar] [CrossRef]
  41. Mogavi, R.H.; Deng, C.; Kim, J.J.; Zhou, P.; Kwon, Y.D.; Metwally, A.H.S.; Tlili, A.; Bassanelli, S.; Bucchiarone, A.; Gujar, S.; et al. ChatGPT in education: A blessing or a curse? A qualitative study exploring early adopters’ utilization and perceptions. Comput. Human Behav. Artif. Humans 2024, 2, 100027. [Google Scholar] [CrossRef]
  42. Bernard, N.; Sagawa, Y., Jr.; Bier, N.; Lihoreau, T.; Pazart, L.; Tannou, T. Using artificial intelligence for systematic review: The example of elicit. BMC Med. Res. Methodol. 2025, 25, 75. [Google Scholar] [CrossRef]
  43. Fabiano, N.; Gupta, A.; Bhambra, N.; Luu, B.; Wong, S.; Maaz, M.; Fiedorowicz, J.G.; Smith, A.L.; Solmi, M. How to optimize the systematic review process using AI tools. JCPP Adv. 2024, 4, e12234. [Google Scholar] [CrossRef]
  44. Baowaly, M.K.; Lin, C.C.; Liu, C.L.; Chen, K.T. Synthesizing electronic health records using improved generative adversarial networks. J. Am. Med. Inform. Assoc. 2019, 26, 228–241. [Google Scholar] [CrossRef] [PubMed]
  45. Simpson, E.; Brown, R.; Sillence, E.; Coventry, L.; Lloyd, K.; Gibbs, J.; Tariq, S.; Durrant, A.C. Understanding the Barriers and Facilitators to Sharing Patient-Generated Health Data Using Digital Technology for People Living with Long-Term Health Conditions: A Narrative Review. Front. Public Health 2021, 9, 641424. [Google Scholar] [CrossRef] [PubMed]
  46. Datt, M.; Sharma, H.; Aggarwal, N.; Sharma, S. Role of ChatGPT-4 for Medical Researchers. Ann. Biomed. Eng. 2024, 52, 1534–1536. [Google Scholar] [CrossRef] [PubMed]
  47. Bulloch, G.; Seth, I.; Zhu, Z.; Stapleton, F.J.; Fung, A.; Tan, Z.; Taylor, H.R. Artificial Intelligence Applications and Performance Metrics in Ophthalmology: A Systematic Review and Meta-Analysis. Ann. Ophthalmol. Vis. Sci. 2023, 6, 1037. [Google Scholar]
  48. Quinn, T.P.; Senadeera, M.; Jacobs, S.; Coghlan, S.; Le, V. Trust and medical AI: The challenges we face and the expertise needed to overcome them. J. Am. Med. Inform. Assoc. 2021, 28, 890–894. [Google Scholar] [CrossRef]
  49. Chustecki, M. Benefits and Risks of AI in Health Care: Narrative Review. Interact. J. Med. Res. 2024, 13, e53616. [Google Scholar] [CrossRef]
  50. Khan, M.M.; Shah, N.; Shaikh, N.; Thabet, A.; Alrabayah, T.; Belkhair, S. Towards secure and trusted AI in healthcare: A systematic review of emerging innovations and ethical challenges. Int. J. Med. Inform. 2025, 195, 105780. [Google Scholar] [CrossRef]
  51. Murdoch, B. Privacy and artificial intelligence: Challenges for protecting health information in a new era. BMC Med. Ethics. 2021, 22, 122. [Google Scholar] [CrossRef]
  52. Williamson, S.M.; Prybutok, V. Balancing privacy and progress: A review of privacy challenges, systemic oversight, and patient perceptions in AI-driven healthcare. Appl. Sci. 2024, 14, 675. [Google Scholar] [CrossRef]
  53. Alowais, S.A.; Alghamdi, S.S.; Alsuhebany, N.; Alqahtani, T.; Alshaya, A.I.; Almohareb, S.N.; Aldairem, A.; Alrashed, M.; Bin Saleh, K.; Badreldin, H.A. Revolutionizing healthcare: The role of artificial intelligence in clinical practice. BMC Med. Educ. 2023, 23, 689. [Google Scholar] [CrossRef]
  54. Chen, Z. Ethics and discrimination in artificial intelligence-enabled recruitment practices. Humanit. Soc. Sci. Commun. 2023, 10, 567. [Google Scholar] [CrossRef]
  55. Challen, R.; Denny, J.; Pitt, M.; Gompels, L.; Edwards, T.; Tsaneva-Atanasova, K. Artificial intelligence, bias and clinical safety. BMJ Qual. Saf. 2019, 28, 231–237. [Google Scholar] [CrossRef] [PubMed]
  56. Parfett, A.; Townley, S.; Allerfeldt, K. AI-based healthcare: A new dawn or apartheid revisited? AI Soc. 2021, 36, 983–999. [Google Scholar] [CrossRef] [PubMed]
  57. Ross, A. AI and the expert: A blueprint for the ethical use of opaque AI. AI Soc. 2022, 39, 925–936. [Google Scholar] [CrossRef]
  58. Bélisle-Pipon, J.C.; Monteferrante, E.; Roy, M.C.; Couture, V. Artificial intelligence ethics has a black box problem. AI Soc. 2023, 38, 1507–1522. [Google Scholar] [CrossRef]
  59. Mirbabaie, M.; Hofeditz, L.; Frick, N.R.J.; Stieglitz, S. Artificial intelligence in hospitals: Providing a status quo of ethical considerations in academia to guide future research. AI Soc. 2022, 37, 1361–1382. [Google Scholar] [CrossRef]
  60. Gupta, R.; Bagdady, K.; Mailey, B.A. Ethical concerns amidst employment of ChatGPT in plastic surgery. Aesthetic Surg. J. 2023, 43, NP656–NP657. [Google Scholar] [CrossRef]
  61. Alkaissi, H.; McFarlane, S.I. Artificial Hallucinations in ChatGPT: Implications in Scientific Writing. Cureus 2023, 15, e35179. [Google Scholar] [CrossRef]
  62. Menz, B.D.; Kuderer, N.M.; Bacchi, S.; Modi, N.D.; Chin-Yee, B.; Hu, T.; Rickard, C.; Haseloff, M.; Vitry, A.; McKinnon, R.A.; et al. Current safeguards, risk mitigation, and transparency measures of large language models against the generation of health disinformation: Repeated cross sectional analysis. BMJ 2024, 384, e078538. [Google Scholar] [CrossRef]
  63. Jiang, M.; Bacchi, S.; Kovoor, J.; Stretton, B.; Chan, W.O. Response to Jin and Dobry’s “ChatGPT for health care providers and patients: Practical implications within dermatology”. JAAD 2024, 90, E105. [Google Scholar] [CrossRef]
  64. Waisberg, E.; Ong, J.; Masalkhi, M.; Lee, A.G. Concerns with OpenAI’s Sora in Medicine. Ann. Biomed. Eng. 2024, 52, 1932–1934. [Google Scholar] [CrossRef] [PubMed]
  65. Paul, D.; Sanap, G.; Shenoy, S.; Kalyane, D.; Kalia, K.; Tekade, R.K. Artificial intelligence in drug discovery and development. Drug Discov. Today 2021, 26, 80–93. [Google Scholar] [CrossRef] [PubMed]
  66. Liao, Q.; Zhang, Y.; Chu, Y.; Ding, Y.; Liu, Z.; Zhao, X.; Wang, Y.; Wan, J.; Ding, Y.; Tiwari, P. Application of artificial intelligence in drug-target interactions prediction: A review. Npj Biomed. Innov. 2025, 2, 1. [Google Scholar] [CrossRef]
  67. Wang, B.; Zhang, T.; Liu, Q.; Sutcharitchan, C.; Zhou, Z.; Zhang, D.; Li, S. Elucidating the role of artificial intelligence in drug development from the perspective of drug-target interactions. J. Pharm. Anal. 2025, 15, 101144. [Google Scholar] [CrossRef]
  68. Vora, L.K.; Gholap, A.D.; Jetha, K.; Thakur, R.R.S.; Solanki, H.K.; Chavda, V.P. Artificial Intelligence in Pharmaceutical Technology and Drug Delivery Design. Pharmaceutics 2023, 15, 1916. [Google Scholar] [CrossRef]
  69. King, M.R. The Future of AI in Medicine: A Perspective from a Chatbot. Ann. Biomed. Eng. 2023, 51, 291–295. [Google Scholar] [CrossRef]
  70. Fehr, J.; Citro, B.; Malpani, R.; Lippert, C.; Madai, V.I. A trustworthy AI reality-check: The lack of transparency of artificial intelligence products in healthcare. Front Digit Health 2024, 6, 1267290. [Google Scholar] [CrossRef]
  71. Díaz-Rodríguez, N.; Del Ser, J.; Coeckelbergh, M.; López de Prado, M.; Herrera-Viedma, E.; Herrera, F. Connecting the dots in trustworthy artificial intelligence: From AI principles, ethics, and key requirements to responsible AI systems and regulation. Inf. Fusion 2023, 99, 101896. [Google Scholar] [CrossRef]
  72. Meszaros, J.; Minari, J.; Huys, I. The future regulation of artificial intelligence systems in healthcare services and medical research in the European Union. Front. Genet. 2022, 13, 927721. [Google Scholar] [CrossRef]
  73. Liu, G.; Luo, Y.; Sheng, J. Discussion of ethical decision mode for artificial intelligence. AI Soc. 2022, 39, 767–773. [Google Scholar] [CrossRef]
  74. Morley, J.; Kinsey, L.; Elhalal, A.; Garcia, F.; Ziosi, M.; Floridi, L. Operationalising AI ethics: Barriers, enablers and next steps. AI Soc. 2021, 38, 411–423. [Google Scholar] [CrossRef]
  75. Rashidian, N.; Abu Hilal, M.; Frigerio, I.; Guerra, M.; Sterckx, S.; Tozzi, F.; Capelli, G.; Verdi, D.; Spolverato, G.; Gulla, A. Ethics and trustworthiness of artificial intelligence in Hepato-Pancreato-Biliary surgery: A snapshot of insights from the European-African Hepato-Pancreato-Biliary Association (E-AHPBA) survey. HPB 2025, 27, 502–510. [Google Scholar] [CrossRef] [PubMed]
  76. Treger, D.; Harris, G.; Thaller, S.R. Advantages and ethics of artificial intelligence in plastic and reconstructive surgery. Artif. Intell. Surg. 2025, 5, 126–132. [Google Scholar] [CrossRef]
  77. Federal Register. Safe, Secure, And Trustworthy Development and Use of Artificial Intelligence. Available online: https://www.federalregister.gov/documents/2023/11/01/2023-24283/safe-secure-and-trustworthy-development-and-use-of-artificial-intelligence (accessed on 30 December 2024).
  78. McDermott Will and Emery. CMS Releases Guidance on Coverage Criteria, Utilization Management, and Use of AI. Available online: https://www.mwe.com/insights/cms-releases-guidance-on-coverage-criteria-utilization-management-and-use-of-ai (accessed on 30 December 2024).
  79. Capelli, G.; Verdi, D.; Frigerio, I.; Rashidian, N.; Ficorilli, A.; Grasso, V.; Majidi, D.; Gumbs, A.A.; Spolverato, G.; Artificial Intelligence Surgery Editorial Board Study Group on Ethics. White paper: Ethics and trustworthiness of artificial intelligence in clinical surgery. Art. Int. Surg. 2023, 3, 111–122. [Google Scholar] [CrossRef]
  80. Kerasidou, A. Ethics of artificial intelligence in global health: Explainability, algorithmic bias and trust. J. Oral. Biol. Craniofac Res. 2021, 11, 612–614. [Google Scholar] [CrossRef]
  81. Voss, M.; Swart, O.; Abel, L.; Mahtani, K. Capacity-building partnerships for surgical post-graduate training in low- and middle-income countries: A scoping review of the literature with exploratory thematic synthesis. Health Policy Plan 2021, 35, 1385–1412. [Google Scholar] [CrossRef]
  82. Balasubramaniam, N.; Kauppinen, M.; Rannisto, A.; Hiekkanen, K.; Kujala, S. Transparency and explainability of AI systems: From ethical guidelines to requirements. Inf. Softw. Technol. 2023, 159, 107197. [Google Scholar] [CrossRef]
  83. Marey, A.; Arjmand, P.; Alerab, A.D.S.; Eslami, M.J.; Saad, A.M.; Sanchez, N.; Umair, M. Explainability, transparency and black box challenges of AI in radiology: Impact on patient care in cardiovascular radiology. Egypt J. Radiol. Nucl. Med. 2024, 55, 183. [Google Scholar] [CrossRef]
  84. Krishnan, G.; Singh, S.; Pathania, M.; Gosavi, S.; Abhishek, S.; Parchani, A.; Dhar, M. Artificial intelligence in clinical medicine: Catalyzing a sustainable global healthcare paradigm. Front AI 2023, 6, 1227091. [Google Scholar] [CrossRef]
  85. Thaichana, P.; Oo, M.Z.; Thorup, G.L.; Chansakaow, C.; Arworn, S.; Rerkasem, K. Integrating Artificial Intelligence in Medical Writing: Balancing Technological Innovation and Human Expertise, with Practical Applications in Lower Extremity Wounds Care. Int. J. Low. Extrem. Wounds 2025. [Google Scholar] [CrossRef]
  86. Guo, C.; He, Y.; Shi, Z.; Wang, L. Artificial intelligence in surgical medicine: A brief review. Ann. Med. Surg. 2025, 87, 2180–2186. [Google Scholar] [CrossRef]
  87. Knudsen, J.E.; Ghaffar, U.; Ma, R.; Hung, A.J. Clinical applications of artificial intelligence in robotic surgery. J. Robot Surg. 2024, 18, 102. [Google Scholar] [CrossRef]
Table 1. Commonly used artificial intelligence tools in scientific research.
Table 1. Commonly used artificial intelligence tools in scientific research.
Primary Function(s)AI ToolFunctionWeblink
Improving writing and grammarGrammarlyImproves grammar, spelling, and punctuation. Paid functions improve writing style and checks for plagiarism.https://www.grammarly.com/ (accessed 2 March 2025)
QuillbotImproves writing, grammar, assists with citations, and paraphrases text.https://quillbot.com/ (accessed 2 March 2025)
TrinkaOnline grammar checker that also improves syntax and spelling.https://www.trinka.ai/ (accessed 2 March 2025)
JenniAssists and improves written language.https://jenni.ai/ (accessed 2 March 2025)
HemingwayImproves the quality of writing.https://hemingwayapp.com/ (accessed 2 March 2025)
ProWritingAidhttps://prowritingaid.com/ (accessed 2 March 2025)
Rytr.meAssists with writing.https://rytr.me/ (accessed 2 March 2025)
Assisting writing and generating ideasChatGPT (ChatGPT-3.5)OpenAI′s LLM utilizes supervised and reinforcement learning techniques to generate relatively accurate and fruitful outputs from user prompts.https://chat.openai.com/
BARDGoogle′s LLM which functions similarly to ChatGPT.https://bard.google.com/ (accessed 2 March 2025)
BERTGoogle′s family of language models introduced in 2018 can answer user′s questions.https://github.com/google-research/bert (accessed 2 March 2025)
Bing AIBing’s LLM that utilized live internet data to provide outputs from user prompts.https://www.bing.com/ (accessed 2 March 2025)
ClaudeA transformer-based LLM which enables it to generate human-like responses to user input.https://console.anthropic.com/login (accessed 2 March 2025)
Med-PaLM 2Google’s LLM dedicated to answering medical queries and assisting in medical research. As of this paper′s publication date, it has not been made public yet.https://sites.research.google/med-palm/ (accessed 2 March 2025)
OsteosynthesisGenerates research articles based on keywords that user’s input.https://www.osteosynthesis.org/pubmed-search (accessed 2 March 2025)
AI-WriterGenerates articles from a simple headline.https://ai-writer.com/ (accessed 2 March 2025)
WriteSonicAssists with writing research articles with SEO-optimized keywords and meta-descriptions.https://writesonic.com/ (accessed 2 March 2025)
Search engine optimizationConsensusSearches the literature and provides a list of relevant articles and their summaries based on user input.https://consensus.app/ (accessed 2 March 2025)
ElicitShows the user relevant systematic reviews and their summaries based on user input, enabling one to filter through based on intervention and side effects.https://elicit.org/ (accessed 2 March 2025)
SEO.aiSearch engine optimization.https://seo.ai/ (accessed 2 March 2025)
Generate/
interpret images
Dall-E3Generates images based on user input.https://openai.com/index/dall-e-3/ (accessed 2 March 2025)
Jasper AIhttps://www.jasper.ai/ (accessed 2 March 2025)
DermEngineGenerates differentials based on the image of skin lesion uploaded.https://www.dermengine.com/ (accessed 2 March 2025)
V7Process images, videos, imaging files, and documents based on data uploaded by the user.https://www.v7labs.com (accessed 2 March 2025)
MidjourneyGenerates images from natural language descriptions.https://www.midjourney.com (accessed 2 March 2025)
SkinVisionUtilizes AI algorithms to diagnose skin moles and classify them as benign or malignant.https://www.skinvision.com/ (accessed 2 March 2025)
Organization and Content SearchLateralAutomatically organize papers for use in future research.https://www.lateral.io/ (accessed 2 March 2025)
Research RabbitTracks citations, creates bibliographies, and generates summaries of papers, allowing researchers to organize their research better.https://www.researchrabbit.ai/ (accessed 2 March 2025)
ReadCube PapersHelps manage bibliographies and references when writing papers.https://www.papersapp.com/ (accessed 2 March 2025)
Connected PapersCreates graphs by analyzing thousands of papers from Semantic Scholar, then selects a few dozen that possess the strongest connection to the original paper.https://www.connectedpapers.com (accessed 2 March 2025)
SciteAble to automatically extract key information from relevant scientific articles such as research questions, methods, citations, and results.https://scite.ai/ (accessed 2 March 2025)
IrisCreates a research map by extracting keywords, contextual synonyms and hypernyms of the research question, then compares them to over 100 million OpenAccess papers to find relevant matches.https://iris.ai/ (accessed 2 March 2025)
ChatPDFAnalyzes PDF files and provides brief summaries of the contents.https://www.chatpdf.com/ (accessed 2 March 2025)
ScholarcyGenerates summary flashcards from documents, links to references, and extract visuals like figures, tables and images.https://www.scholarcy.com/ (accessed 2 March 2025)
Audio analysisFireflies.aiAnalyze, record, and transcribe voice inputs.https://fireflies.ai/ (accessed 2 March 2025)
PowerScribe360Understands user’s speech and transcribes them onto textual documentation.https://www.nuance.com/en-au/healthcare/medical-imaging/powerscribe-360-reporting.html (accessed 2 March 2025)
audEERINGIntended to diagnose pathologies like Parkinson’s and COVID-19 by collecting audio data from patients.https://www.audeering.com/technology/health-ai/ (accessed 2 March 2025)
Publication and CollaborationSciSpaceAids researchers in publishing their works by facilitating the manuscript publication and peer reviewing.https://scispace.com/ (accessed 2 March 2025)
PubMedA versatile search engine for biomedical literature.https://pubmed.ncbi.nlm.nih.gov/ (accessed 2 March 2025)
EmbaseA database for biomedical and pharmacological literature aiding regulatory compliance for licensed drugs.https://www.embase.com/ (accessed 2 March 2025)
Video generationSoraOpenAI′s upcoming AI model specializes in generating videos from text inputs.https://openai.com/index/sora/ (accessed 2 March 2025)
Table 2. Comparative table assessing each tool’s strengths and weaknesses.
Table 2. Comparative table assessing each tool’s strengths and weaknesses.
Tool NamePrimary FunctionKey StrengthsKnown Limitations/Evidence Gaps
GrammarlyWriting improvementRobust grammar correction; tone suggestions; plagiarism checker.Limited domain-specific support; mostly non-medical; no formal academic validation.
QuillbotParaphrasing, citationsUseful for rephrasing and summarization; citation tools.Paraphrasing may distort scientific meaning; no peer-reviewed assessment.
TrinkaGrammar, spelling, and syntaxAcademic tone correction; includes technical vocabulary suggestions.Limited clinical/biomedical domain validation.
ChatGPTContent generation, Q&ABroad utility; conversational; custom prompts can generate lay summaries.Risk of hallucinations; no transparency on sources.
Claude/Bard/Bing AISimilar to ChatGPTIntegration with real-time search (Bing), enhanced safety (Claude).Lack of real-world performance evaluation in health-specific tasks.
Med-PaLM 2Medical Q&ASpecialized in healthcare; outperformed general LLMs in benchmarks.Not publicly available; independent peer-reviewed data lacking.
ElicitResearch Q&A with source filtersExtracts claims from papers; filters by population/intervention/outcome.Still experimental; no formal validation of claim extraction accuracy.
Scite.aiCitation-based insightsDifferentiates supporting/contradicting citations; aids evidence tracking.Performance varies across fields; citation sentiment can be ambiguous.
Connected PapersCitation mappingHelps visualize related works in a citation graph.Limited to Semantic Scholar database; not all connections are semantically meaningful.
DALL-E3/MidjourneyImage generation from textHigh-quality image output; artistic or illustrative applications.Not clinically validated; inappropriate for medical images or diagnostic tasks.
DermEngine/SkinVisionDiagnostic imagingLesion classification using image input; CE-marked in EU (SkinVision).Studies show variable accuracy 4; not meant to replace clinician judgment.
Scholarcy/ChatPDFPDF summarizationSummarizes large texts; identifies key results and visuals.Can miss nuance; accuracy of automatic extraction unverified in the literature.
Fireflies/PowerScribe360Audio transcriptionAccurate speech-to-text for meetings or reporting.Limited biomedical validation; performance affected by accents and background noise.
SciSpacePublishing supportHelps with formatting, peer review integration.Little formal evaluation of its impact on publication quality or speed.
Sora (OpenAI)Text-to-video generationInnovative multimedia generation from text prompts.No clinical applications; use in science communication remains theoretical.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lim, B.; Seth, I.; Cevik, J.; Mu, X.; Sofiadellis, F.; Cuomo, R.; Rozen, W.M. Artificial Intelligence Tools in Surgical Research: A Narrative Review of Current Applications and Ethical Challenges. Surgeries 2025, 6, 55. https://doi.org/10.3390/surgeries6030055

AMA Style

Lim B, Seth I, Cevik J, Mu X, Sofiadellis F, Cuomo R, Rozen WM. Artificial Intelligence Tools in Surgical Research: A Narrative Review of Current Applications and Ethical Challenges. Surgeries. 2025; 6(3):55. https://doi.org/10.3390/surgeries6030055

Chicago/Turabian Style

Lim, Bryan, Ishith Seth, Jevan Cevik, Xin Mu, Foti Sofiadellis, Roberto Cuomo, and Warren M. Rozen. 2025. "Artificial Intelligence Tools in Surgical Research: A Narrative Review of Current Applications and Ethical Challenges" Surgeries 6, no. 3: 55. https://doi.org/10.3390/surgeries6030055

APA Style

Lim, B., Seth, I., Cevik, J., Mu, X., Sofiadellis, F., Cuomo, R., & Rozen, W. M. (2025). Artificial Intelligence Tools in Surgical Research: A Narrative Review of Current Applications and Ethical Challenges. Surgeries, 6(3), 55. https://doi.org/10.3390/surgeries6030055

Article Metrics

Back to TopTop