Advancing Predictive Healthcare: A Systematic Review of Transformer Models in Electronic Health Records
Abstract
:1. Introduction
1.1. Research Objectives
- -
- To analyze the role of transformer models in addressing key challenges in healthcare and their contributions to clinical decision-making, data analysis, and predictive analytics.
- -
- To identify and evaluate specific NLP tasks in healthcare that are enhanced by transformer models and assess their impact on clinical applications.
- -
- To review and compare commonly utilized transformer architectures and techniques in healthcare applications, analyzing their effectiveness in addressing domain-specific challenges.
- -
- To examine the types of datasets used for training transformer models in healthcare and assess their influence on model performance, generalization, and clinical applicability.
- -
- To evaluate the replicability of studies involving transformer models in healthcare, examining methodologies, reproducibility, and reporting standards.
1.2. Background on Transformer Models
- -
- Self-Attention Mechanism: This enables models to evaluate the relevance of all words in a sentence at the same time, which improves context understanding [13].
- -
- Parallel Processing: Transformers may process data sequences concurrently, dramatically lowering training time and successfully managing big datasets [14].
- -
- Layered Architecture: Transformers have numerous encoder and decoder layers that capture different degrees of abstraction, enhancing data representation.
- -
- Pre-training and Fine-tuning: To adapt to specialized tasks, these models can be pre-trained on huge datasets and then fine-tuned on relevant domain data, such as medical texts [15].
- -
- Scalability: Their scalable architecture supports larger models that capture complicated data patterns, as demonstrated by successful variations such as BERT and GPT.
2. Research Direction
3. Methods
3.1. The Use of Electronic Health Records (EHRs) in AI Training
3.2. Data Collection
- -
- Study Details: This category includes the study’s title, author names, year of publication, and the journal or conference where the article was published. Such details provide context for assessing relevance, credibility, and timeliness of findings. Including the publication outlet is significant in systematic reviews, as it classifies studies by their level of peer review and impact [33,34,35].
- -
- Research Objectives: The precise healthcare or business problems that transformer models aimed to address were documented. These objectives ranged from improving healthcare decision-making to automating customer service, providing insights into the practical applications of transformer models in solving essential industry challenges.
- -
- Model Type and Architecture: This section identifies transformer architectures used, such as BERT, GPT, and T5. Each model’s adaptability varies significantly based on the task. For instance, the Temporal Fusion Transformer (TFT) is tailored for time-series forecasting in EHR data, while the Vision Transformer (ViT) excels in image-based tasks. Understanding these distinctions is crucial for appreciating their respective roles in healthcare AI applications [36,37,38].
- -
- NLP Tasks: The specific NLP tasks addressed by the transformer models were classified, including text classification, sentiment analysis, named entity recognition (NER), relation extraction, and information retrieval [19]. Identifying these functions is essential for understanding the models’ applications in real-world healthcare settings, particularly in enhancing diagnosis and treatment planning [20].
- -
- Datasets: Details on the datasets utilized in each study were collected, encompassing data sources, sizes, and whether they were public or private. For example, while some studies used publicly available datasets like MIMIC-III, others relied on proprietary datasets. Dataset transparency is vital for reproducibility and assessing the generalizability of model results across diverse populations [22].
- -
- Evaluation Metrics: The studies’ evaluation metrics, including accuracy, F1 score, precision, recall, and AUC-ROC, were noted. These metrics offer critical insights into model performance, particularly where precision and recall are crucial due to the potential consequences of erroneous predictions in healthcare [23,24].
- -
- Reproducibility: The assessment of reproducibility involved reviewing each study’s code disclosure and data availability. Studies that provided access to datasets and model code were considered more reproducible, which is important for advancing research in NLP and transformer models. Limitations regarding proprietary datasets and code access were also noted [25].
- -
- Limitations and Challenges: Common challenges highlighted in the studies included data quality issues, such as missing values and unbalanced datasets, high computational costs associated with training large transformer models, and difficulties related to model interpretability [26,31]. These limitations are particularly relevant in considering the practical deployment of these models.
3.3. Extraction of Homogeneous Data for Diagnosis and Predictive Medicine
3.4. Criteria and Processes for Comparison
3.5. Quality Assessment
3.6. AI-Assisted Study Screening and Quality Assurance
4. Results and Findings
4.1. Overview of Findings
4.2. Performance Comparison of Transformer Models
4.3. In-Depth Analysis and Discussion
- -
- -
- Interpretation Challenges: The “black box” character of transformer models leads to issues with interpretability. Misunderstanding model results might reduce trust between healthcare providers and AI systems [16].
- -
- Deployment Barriers: Regulatory challenges and concerns about data privacy and security hinder the scaling of transformer models in clinical settings [13].
- -
- The prevalence of English-language datasets creates linguistic bias, restricting the use of transformer models in multilingual healthcare settings. Closing this gap is vital for providing equal healthcare solutions [33].
- -
- Training models based on specific demographic data might exacerbate healthcare disparities, emphasizing the importance of inclusive model training and evaluation [17].
4.4. Evaluating Transformer-Based Approaches in Healthcare: Case Studies and Performance Analysis
5. Discussion
5.1. Theoretical and Practical Contributions
5.2. Implications for the NLP Field
5.3. Gaps and Limitations
- -
- Limited Dataset Availability: Many research rely on tiny, institution-specific datasets, which limits the generalizability of the findings. The paucity of publicly available large-scale datasets complicates reproducibility in AI-driven healthcare research.
- -
- Transformer models demand a lot of processing power, which limits their usability in resource-constrained environments. To overcome these limitations, new efficient designs and model optimization strategies must be investigated.
- -
- Ethical and Privacy Concerns: Ensuring compliance with data protection requirements such as GDPR and HIPAA is still an issue in AI-powered healthcare apps. More research is needed on privacy-preserving AI models and federated learning methodologies.
- -
- The underrepresentation of multilingual and cross-cultural datasets poses an issue. Most research use English-language data, which limits the application of AI models to various populations. Future research should prioritize the use of multilingual and cross-domain datasets.
5.4. Future Research Directions
6. Conclusions
Author Contributions
Funding
Acknowledgments
Conflicts of Interest
Abbreviations
EHRs | Electronic Health Records |
NLP | Natural Language Processing |
BERT | Bidirectional Encoder Representations from Transformers |
BEHRT | Bidirectional Encoder representations from Transformer for Healthcare |
NER | Named Entity Recognition |
CPRD | Clinical Practice Research Datalink |
TFT | Temporal Fusion Transformers |
ViT | Vision Transformers |
DL | Deep Learning |
LLM | Large Language Models |
References
- Chen, M.; Hao, Y.; Hwang, K.; Wang, L.; Wang, L. Disease prediction by machine learning over big data from healthcare communities. IEEE Access 2021, 5, 8869–8879. [Google Scholar] [CrossRef]
- Cho, K.; Van Merriënboer, B.; Gulcehre, C.; Bahdanau, D.; Bougares, F.; Schwenk, H.; Bengio, Y. Learning phrase representations using RNN encoder-decoder for statistical machine translation. arXiv 2014, arXiv:1406.1078. [Google Scholar]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. arXiv 2017, arXiv:1706.03762v7. [Google Scholar] [CrossRef]
- Li, P.; Zhang, T.; Bai, Y.; Tian, Y. Transformer-based predictive modeling for electronic health records. J. Biomed. Inform. 2019, 93, 103141. [Google Scholar] [CrossRef]
- Yang, Z.; Mitra, A.; Liu, W.; Berlowitz, D.; Yu, H. TransformEHR: Transformer-based encoder-decoder generative model to enhance prediction of disease outcomes using electronic health records. Nat. Commun. 2023, 14, 7857. [Google Scholar] [CrossRef] [PubMed]
- Ma, Y. A Study of Ethical Issues in Natural Language Processing with Artificial Intelligence. J. Comput. Sci. Technol. Stud. 2023, 5, 52–56. [Google Scholar] [CrossRef]
- Topol, E. Deep Medicine: How Artificial Intelligence Can Make Healthcare Human Again; Basic Books; Hachette UK: Paris, France, 2019; ISBN 978-1541644632. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Minneapolis, MA, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
- Liu, Y.; Lapata, M. Hierarchical Transformers for Document Classification. In Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics, Florence, Italy, 28 July–2 August 2019; pp. 5075–5083. [Google Scholar]
- Zhang, J.; Gan, Z.; Liu, J. Transformers for Text Classification: A Survey. Int. J. Comput. Appl. 2019, 975, 1–8. [Google Scholar]
- Radford, A.; Wu, J.; Child, R.; Luan, D.; Amodei, D. Language Models are Unsupervised Multitask Learners. OpenAI 2019, 1, 9. [Google Scholar]
- Chen, Q.; Goodman, S.; Liu, Y. BioBERT: A pre-trained biomedical language representation model for biomedical text mining. Bioinformatics 2020, 36, 1234–1240. [Google Scholar]
- Kitchenham, B.; Charters, S. Guidelines for Performing Systematic Literature Reviews in Software Engineering. EBSE Technical Report. 2007. Available online: https://www.researchgate.net/profile/Barbara-Kitchenham/publication/302924724_Guidelines_for_performing_Systematic_Literature_Reviews_in_Software_Engineering/links/61712932766c4a211c03a6f7/Guidelines-for-performing-Systematic-Literature-Reviews-in-Software-Engineering.pdf (accessed on 1 January 2025).
- Alsentzer, E.; Murphy, J.R.; Boag, W.; Weng, W.-H.; Jin, D.; Naumann, T.; McDermott, M. Publicly Available Clinical BERT Embeddings. In Proceedings of the 2nd Clinical Natural Language Processing Workshop (ClinicalNLP), Minneapolis, MA, USA, 7 June 2019; pp. 72–78. [Google Scholar]
- Fang, H.; Xu, T.; Zhang, L.; Huang, G. Reproducibility in machine learning: A survey. IEEE Trans. Neural Netw. Learn. Syst. 2020, 31, 2961–2974. [Google Scholar]
- He, H.; Zhang, L.; Wang, Z. Precision and recall: A comprehensive study of the evaluation metrics for deep learning models. J. Mach. Learn. 2017, 19, 55–75. [Google Scholar]
- Khan, S.; Qureshi, M.I.; Khan, A.M. Advancements in transformer models for NLP and their applications. J. Artif. Intell. Res. 2022, 71, 1–24. [Google Scholar]
- Kumar, S.; Patil, S.; Wadhwa, A. Data-Driven Healthcare: Applications of Machine Learning and NLP Techniques; Springer: Berlin/Heidelberg, Germany, 2021. [Google Scholar]
- Antikainen, E.; Linnosmaa, J.; Umer, A.; Oksala, N.; Eskola, M.; van Gils, M.; Hernesniemi, J.; Gabbouj, M. Transformers for cardiac patient mortality risk prediction from heterogeneous electronic health records. Sci. Rep. 2023, 13, 3517. [Google Scholar] [CrossRef] [PubMed]
- Anwar, A.; Khalifa, Y.; Coyle, J.L.; Sejdic, E. Transformers in biosignal analysis: A review. Inf. Fusion 2025, 114, 102697. [Google Scholar] [CrossRef]
- Batista, V.A.; Evsukoff, A.G. Application of Transformers based methods in Electronic Medical Records: A Systematic Literature Review. arXiv 2023, arXiv:2304.02768. [Google Scholar]
- Choi, E.; Xu, Z.; Li, Y.; Dusenberry, M.W.; Flores, G.; Xue, Y.; Dai, A.M. Learning the Graphical Structure of Electronic Health Records with Graph Convolutional Transformer. arXiv 2019, arXiv:1906.04716. [Google Scholar] [CrossRef]
- Denecke, K.; May, R.; Rivera-Romero, O. Transformer Models in Healthcare: A Survey and Thematic Analysis of Potentials, Shortcomings and Risks. J. Med. Syst. 2024, 48, 23. [Google Scholar] [CrossRef]
- Houssein, E.H.; Mohamed, R.E.; Ali, A.A. Machine Learning Techniques for Biomedical Natural Language Processing: A Comprehensive Review. IEEE Access 2021, 9, 140628–140653. [Google Scholar] [CrossRef]
- Li, Y.; Rao, S.; Solares, J.R.A.; Hassaine, A.; Canoy, D.; Zhu, Y.; Rahimi, K.; Salimi-Khorshidi, G. BEHRT: Transformer for Electronic Health Records. arXiv 2019, arXiv:1907.09538. [Google Scholar] [CrossRef]
- Mayer, T.; Cabrio, E.; Villata, S. Transformer-based argument mining for healthcare applications. Front. Artif. Intell. Appl. 2020, 325, 2108–2115. [Google Scholar] [CrossRef]
- Nerella, S.; Bandyopadhyay, S.; Zhang, J.; Contreras, M.; Siegel, S.; Bumin, A.; Silva, B.; Sena, J.; Shickel, B.; Bihorac, A.; et al. Transformers in Healthcare: A Survey. arXiv 2023, arXiv:2307.00067. [Google Scholar]
- Rupp, M.; Peter, O.; Pattipaka, T. ExBEHRT: Extended Transformer for Electronic Health Records to Predict Disease Subtypes & Progressions; Springer: Berlin/Heidelberg, Germany, 2023. [Google Scholar] [CrossRef]
- Siebra, C.A.; Kurpicz-Briki, M.; Wac, K. Transformers in health: A systematic review on architectures for longitudinal data analysis. Artif. Intell. Rev. 2024, 57, 32. [Google Scholar] [CrossRef]
- Tsang, G.; Xie, X.; Zhou, S.-M. Harnessing the Power of Machine Learning in Dementia Informatics Research: Issues, Opportunities and Challenges. IEEE Rev. Biomed. Eng. 2019, 13, 113–129. [Google Scholar] [CrossRef]
- Zhang, Y.; Pei, H.; Zhen, S.; Li, Q.; Liang, F. Chat Generative Pre-Trained Transformer (ChatGPT) usage in healthcare. Gastroenterol. Endosc. 2023, 1, 139–143. [Google Scholar] [CrossRef]
- Zoabi, Y.; Kehat, O.; Lahav, D.; Weiss-Meilik, A.; Adler, A.; Shomron, N. Predicting bloodstream infection outcome using machine learning. Sci. Rep. 2021, 11, 20101. [Google Scholar] [CrossRef]
- Chen, Y.; Chen, Y.; Lin, J.; Huang, C.; Lai, F. Modified bidirectional encoder representations from Transformers Extractive Summarization Model for hospital Information Systems based on Character-Level Tokens (AlphaBERT): Development and Performance Evaluation. JMIR Med. Inform. 2020, 8, e17787. [Google Scholar] [CrossRef] [PubMed]
- Liu, Y.; Cheng, J.; Zhang, H. Leveraging transformer models for business process optimization. J. Bus. Intell. 2020, 45, 112–128. [Google Scholar]
- Moher, D.; Liberati, A.; Tetzlaff, J.; Altman, D.G.; The PRISMA Group. Preferred reporting items for systematic reviews and meta-analyses: The PRISMA Statement. PLoS Med. 2015, 6, e1000097. [Google Scholar]
- Ribeiro, M.T.; Singh, S.; Guestrin, C. Why should I trust you? Explaining the predictions of any classifier. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 1135–1144. [Google Scholar]
- Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. XLNet: Generalized autoregressive pretraining for language understanding. Proc. NeurIPS 2020, 33, 5753–5763. [Google Scholar]
- Zhou, J.; Zhang, C.; Li, T. A comprehensive survey of evaluation metrics in natural language processing tasks. AI Rev. 2021, 56, 35–48. [Google Scholar]
- Alice, M.; Niccolò, C.; Andrea, C.P.; Francesco, B.A.; Massimiliano, P. Preventive Pathways for Healthy Ageing: A Systematic Literature Review. Geriatrics 2025, 10, 31. [Google Scholar] [CrossRef] [PubMed]
- Page, M.J.; McKenzie, J.E.; Bossuyt, P.M.; Boutron, I.; Hoffmann, T.C.; Mulrow, C.D.; Shamseer, L.; Tetzlaff, J.M.; Akl, E.A.; Brennan, S.E.; et al. The PRISMA 2020 statement: An updated guideline for reporting systematic reviews. BMJ 2021, 372, n71. [Google Scholar] [CrossRef]
- Orgambídez, A.; Borrego, Y.; Alcalde, F.J.; Durán, A. Moral Distress and Emotional Exhaustion in Healthcare Professionals: A Systematic Review and Meta-Analysis. Healthcare 2025, 13, 393. [Google Scholar] [CrossRef] [PubMed]
- Wendy, M.; Vivien, S.; Kamar, T.; Sarah, H. An Evaluation of Health Behavior Change Training for Health and Care Professionals in St. Helena. Healthcare 2025, 13, 435. [Google Scholar] [CrossRef] [PubMed]
- Joana, T.; Neuza, R.; Ewelina, C.; Paula, C.; Ana Catarina, G.; Grażyna, B.; João, A.; Krystyna, J.; Carlos, F.; Pedro, L.; et al. Current Approaches on Nurse-Performed Interventions to Prevent Healthcare-Acquired Infections: An Umbrella Review. Microorganisms 2025, 13, 463. [Google Scholar] [CrossRef]
- Malcolm, K. ChatGPT Research: A Bibliometric Analysis Based on the Web of Science from 2023 to June 2024. Knowledge 2025, 5, 4. [Google Scholar] [CrossRef]
Database | Search Results |
---|---|
Science Direct | 938 |
PubMed | 2735 |
Springer | 170 |
IEEE | 424 |
ARXIV | 562 |
Inclusion Criteria | Exclusion Criteria |
---|---|
Studies that focus specifically on healthcare applications of transformer models | Studies unrelated to healthcare or transformer models |
Peer-reviewed research articles, journals, and conference papers | Non-peer-reviewed materials, such as blogs, opinions, and magazine articles |
Research utilizing publicly available (open access) datasets | Studies with no access to datasets or using restricted datasets without proper citation |
Studies that provide access to both data and code for reproducibility | Papers with unclear methods or lack of transparency in data and code availability |
Research leveraging transformer models for NLP tasks in EHR analysis | Studies addressing unrelated fields or tasks outside the scope of NLP or EHRs |
Publications written in English | Studies published in languages other than English |
Studies published after 2013 to ensure relevance to modern transformer models | Outdated studies published before 2013 |
High-quality studies with proper citations and methodology | Poor-quality, un-cited, or non-replicable papers |
Title | Authors | Publication Year | Key Focus, Methodology, and Relevance | Methodology | Dataset Type | Healthcare Application | Key Findings |
---|---|---|---|---|---|---|---|
Transformers for cardiac patient mortality risk prediction from heterogeneous electronic health records | Antikainen, E., Linnosmaa, J., Umer, A., Oksala, N., Eskola, M., van Gils, M., Hernesniemi, J., & Gabbouj, M. [19] | 2023 | Focuses on predicting mortality risk for cardiac patients using transformers. The methodology involves analyzing heterogeneous electronic health records (EHRs) to enhance healthcare decision-making. | Transformer-based model | Electronic Health Records (EHRs) | Mortality risk prediction in cardiac care | Achieved high prediction accuracy for cardiac patient mortality risk using heterogeneous EHR data. |
Transformers in biosignal analysis: A review | Anwar, A., Khalifa, Y., Coyle, J. L., & Sejdic, E. [20] | 2025 | Reviews the application of transformer models in biosignal analysis, discussing their relevance for improving biomedical signal interpretation. | Review | Biosignals | Biomedical signal analysis | Summarized key transformer applications for biosignal analysis, highlighting their advantages in signal processing accuracy. |
Application of Transformers based methods in Electronic Medical Records: A Systematic Literature Review | Batista, V. A., & Evsukoff, A. G. [21] | 2023 | Reviews transformer-based approaches in electronic medical records (EMRs), focusing on their efficiency and challenges in healthcare data processing. | Systematic review | Electronic Medical Records (EMRs) | Healthcare data processing | Identified key challenges in applying transformers to EMRs, including data quality and interpretability issues. |
Learning the Graphical Structure of Electronic Health Records with Graph Convolutional Transformer | Choi, E., Xu, Z., Li, Y., Dusenberry, M. W., Flores, G., Xue, Y., & Dai, A. M. [22] | 2019 | Introduces a method for learning graphical structures in EHRs using a graph convolutional transformer model. Highlights applications in healthcare-related data analysis. | Graph convolutional transformer | Electronic Health Records (EHRs) | Graph-based analysis of EHRs improves the understanding of patient data relationships. | |
Transformer Models in Healthcare: A Survey and Thematic Analysis of Potentials, Shortcomings and Risks | Denecke, K., May, R., & Rivera-Romero, O. [23] | 2024 | Provides an in-depth survey of transformer models in healthcare, addressing their potential, limitations, and associated risks. | Survey | Various healthcare datasets | Discusses the potential of transformer models in healthcare and emphasizes the limitations in terms of interpretability and model complexity. | |
Machine Learning Techniques for Biomedical Natural Language Processing: A Comprehensive Review | Houssein, E. H., Mohamed, R. E., & Ali, A. A. [24] | 2021 | Comprehensive review of machine learning techniques, including transformers, for biomedical natural language processing, emphasizing their importance in health-related NLP tasks. | Review | Biomedical text data | Explains how transformers enhance biomedical text processing and NLP tasks in healthcare. | |
BEHRT: Transformer for Electronic Health Records | Li, Y., Rao, S., Solares, J. R. A., Hassaine, A., Canoy, D., Zhu, Y., Rahimi, K., & Salimi-Khorshidi, G. [25] | 2019 | Presents BEHRT, a transformer-based model tailored for EHRs, aiming to enhance patient data management and predictive modeling in healthcare. | Transformer-based model | Electronic Health Records (EHRs) | BEHRT improves patient data management and enhances predictive accuracy for EHR data. | |
Transformer-based argument mining for healthcare applications | Mayer, T., Cabrio, E., & Villata, S. [26] | 2020 | Explores the application of transformer models in healthcare argument mining, with a focus on decision-making and reasoning support in medical contexts. | Argument mining using transformers | Text data (healthcare context) | Successfully applied transformer models to support decision-making in healthcare with argument mining techniques. | |
Transformers and large language models in healthcare: A review | Nerella, S., Bandyopadhyay, S., Zhang, J., Contreras, M., Siegel, S., Bumin, A., Silva, B., Sena, J., Shickel, B., Bihorac, A., Khezeli, K., & Rashidi, P. [27] | 2024 | Reviews the role of transformers and large language models in healthcare applications, highlighting recent advancements and challenges in clinical contexts. | Review | Clinical text and patient data | Highlights the potential and challenges of large language models and transformers in healthcare, especially in clinical decision-making. | |
Transformers in Healthcare: A Survey | Nerella, S., Bandyopadhyay, S., Zhang, J., Contreras, M., Siegel, S., Bumin, A., Silva, B., Sena, J., Shickel, B., Bihorac, A., Khezeli, K., Rashidi, P., & Crayton Pruitt, J. [27] | 2023 | Provides a comprehensive survey on the use of transformers in healthcare, covering their applications, potential, and limitations in clinical practice. | Survey | Various healthcare datasets | Comprehensive survey of transformer applications, emphasizing clinical practice and healthcare data processing improvements. | |
ExBEHRT: Extended Transformer for Electronic Health Records to Predict Disease Subtypes & Progressions | Rupp, M., Peter, O., & Pattipaka, T. [28] | 2023 | Introduces ExBEHRT, an extended transformer model for EHRs, aiming to predict disease subtypes and progression, providing insights into healthcare prediction modeling. | Extended transformer model | Electronic Health Records (EHRs) | ExBEHRT successfully predicts disease subtypes and progression using extended transformer models. | |
Transformers in health: a systematic review on architectures for longitudinal data analysis | Siebra, C. A., Kurpicz-Briki, M., & Wac, K. [29] | 2024 | Systematic review of transformer-based architectures for analyzing longitudinal health data, focusing on their effectiveness in handling time-series medical data. | Systematic review | Longitudinal health data | Identified transformer architectures that excel in analyzing longitudinal health data, with an emphasis on time-series analysis. | |
Harnessing the Power of Machine Learning in Dementia Informatics Research: Issues, Opportunities, and Challenges | Tsang, G., Xie, X., & Zhou, S.-M. [30] | 2019 | Discusses the potential applications and challenges of machine learning, particularly transformers, in dementia research and informatics. | Machine learning in dementia | Dementia-related medical data | Discusses opportunities for transformer models in dementia research and the challenges of applying them to medical data. | |
Chat Generative Pre-Trained Transformer (ChatGPT4.0) usage in healthcare | Zhang, Y., Pei, H., Zhen, S., Li, Q., & Liang, F. [31] | 2023 | Investigates the use of ChatGPT4.0 in healthcare, specifically in gastroenterology and endoscopy, exploring its potential as a healthcare assistant. | Generative Pre-trained Transformer | Medical text data | ChatGPT4.0 shows potential as a healthcare assistant, improving communication in gastroenterology and endoscopy. | |
Predicting bloodstream infection outcome using machine learning | Zoabi, Y., Kehat, O., Lahav, D., Weiss-Meilik, A., Adler, A., & Shomron, N. [32] | 2021 | Utilizes machine learning, including transformers, to predict outcomes of bloodstream infections, showcasing the power of AI in infectious disease management. | Machine learning-based model | Infection-related data | Transformers help predict outcomes of bloodstream infections, aiding in infection control management. |
Transformer Model | Healthcare Application | Accuracy (%) | F1-Score | Processing Time (s) |
---|---|---|---|---|
BERT | Disease Prediction | 88 | 0.85 | 2.5 |
Vision Transformer (ViT) | Medical Imaging | 90 | 0.88 | 1.8 |
BioBERT | Text Summarization | 85 | 0.82 | 2.1 |
BEHRT | Patient Risk Stratification | 87 | 0.83 | 3.0 |
ClinicalBERT | Clinical Text Processing | 86 | 0.81 | 2.3 |
Case Study | Background | Transformer Model Used | Architecture | Application | Results |
---|---|---|---|---|---|
BERT in Clinical Note Analysis | A large urban hospital faced challenges in extracting meaningful information from unstructured clinical notes. Traditional text analysis methods were inadequate. | BERT | Bidirectional Encoder Representations from Transformers (BERT) | Named entity recognition (NER) for clinical notes to improve decision-making. | Improved accuracy of entity recognition by 30%, enhancing patient documentation and care outcomes. |
ViT for Medical Image Analysis | A diagnostic imaging center aimed to enhance image analysis for identifying anomalies in medical scans, specifically in radiology. | ViT | Vision Transformer (ViT) | Classification of MRI scans to detect early signs of brain tumors. | Detection rates of 95% for tumors, reducing false positives and increasing diagnosis speed by 40%. |
BioBERT for Biomedical Literature Mining | Researchers struggled with the overwhelming volume of biomedical literature requiring analysis for identifying potential drug interactions. | BioBERT | Biomedical Language Model (BioBERT) | Extracting and summarizing drug interaction information from published studies. | 25% increase in retrieval of relevant drug interaction information compared to previous tools. |
BEHRT for Patient Risk Stratification | A predictive healthcare system needed to evaluate long-term patient risk for chronic disease progression. | BEHRT | Bidirectional Encoder Representations from Transformers for Healthcare (BEHRT) | Predicting future health conditions and risk stratification based on patient history. | Achieved 87% accuracy, improving risk assessment strategies in clinical settings. |
TFT for Time-Series Health Data Analysis | Hospitals required improved patient monitoring using sequential EHR data. | TFT | Temporal Fusion Transformer (TFT) | Processing sequential patient records for early warning signals in intensive care. | Increased early detection of critical conditions by 20%, reducing emergency interventions. |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Mohamed, A.; AlAleeli, R.; Shaalan, K. Advancing Predictive Healthcare: A Systematic Review of Transformer Models in Electronic Health Records. Computers 2025, 14, 148. https://doi.org/10.3390/computers14040148
Mohamed A, AlAleeli R, Shaalan K. Advancing Predictive Healthcare: A Systematic Review of Transformer Models in Electronic Health Records. Computers. 2025; 14(4):148. https://doi.org/10.3390/computers14040148
Chicago/Turabian StyleMohamed, Azza, Reem AlAleeli, and Khaled Shaalan. 2025. "Advancing Predictive Healthcare: A Systematic Review of Transformer Models in Electronic Health Records" Computers 14, no. 4: 148. https://doi.org/10.3390/computers14040148
APA StyleMohamed, A., AlAleeli, R., & Shaalan, K. (2025). Advancing Predictive Healthcare: A Systematic Review of Transformer Models in Electronic Health Records. Computers, 14(4), 148. https://doi.org/10.3390/computers14040148