Next Article in Journal
Unveiling the Immunogenicity of Ovarian Tumors as the Crucial Catalyst for Therapeutic Success
Next Article in Special Issue
The Rise of Hypothesis-Driven Artificial Intelligence in Oncology
Previous Article in Journal
TMEM97/Sigma 2 Receptor Increases Estrogen Receptor α Activity in Promoting Breast Cancer Cell Growth
Previous Article in Special Issue
Radiomics and Clinicopathological Characteristics for Predicting Lymph Node Metastasis in Testicular Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Breaking Barriers: AI’s Influence on Pathology and Oncology in Resource-Scarce Medical Systems

by
Alon Vigdorovits
1,2,
Maria Magdalena Köteles
3,
Gheorghe-Emilian Olteanu
2,4,5,* and
Ovidiu Pop
1
1
Department of Pathology, County Clinical Emergency Hospital, Faculty of Medicine and Pharmacy, University of Oradea, 1 December Sq. No. 10, 410087 Oradea, Romania
2
Center for Research and Innovation in Personalized Medicine of Respiratory Diseases, “Victor Babes” University of Medicine and Pharmacy, 300041 Timisoara, Romania
3
Bihor County Clinical Emergency Hospital, Gheorghe Doja, Street No. 65, 410169 Oradea, Romania
4
Faculty of Pharmacy, “Victor Babes” University of Medicine and Pharmacy, Timisoara Eftimie Murgu Sq. No. 2, 300041 Timisoara, Romania
5
Research Center for Pharmaco-Toxicological Evaluations, Faculty of Pharmacy, “Victor Babes” University of Medicine and Pharmacy, Timisoara Eftimie Murgu Sq. No. 2, 300041 Timisoara, Romania
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(23), 5692; https://doi.org/10.3390/cancers15235692
Submission received: 30 October 2023 / Revised: 30 November 2023 / Accepted: 30 November 2023 / Published: 2 December 2023
(This article belongs to the Collection Artificial Intelligence in Oncology)

Abstract

:

Simple Summary

The quality of care available to cancer patients is highly constrained by the presence of comprehensive pathology services. However, pathology infrastructure is severely lacking in most resource-scarce medical systems. In recent years, progress made in the digitalization of pathology workflows and the advent of clinical-grade AI tools for pathology have offered pathologists new ways to increase their efficiency and offer better care to more patients. In this review, we aim to provide an overview of the current state of digitalization and AI in pathology and how it impacts the practice of oncology. We also identify the main challenges of implementing digital pathology and AI in resource-poor settings and point out the different approaches needed to tackle these challenges.

Abstract

The application of artificial intelligence to improve the access of cancer patients to high-quality medical care is one of the goals of modern medicine. Pathology constitutes the foundation of modern oncologic treatment, and its role has expanded far beyond diagnosis into predicting treatment response and overall survival. However, the funding of pathology is often an afterthought in resource-scarce medical systems. The increased digitalization of pathology has paved the way towards the potential use of artificial intelligence tools for improving pathologist efficiency and extracting more information from tissues. In this review, we provide an overview of the main research directions intersecting with artificial intelligence and pathology in relation to oncology, such as tumor classification, the prediction of molecular alterations, and biomarker quantification. We then discuss examples of tools that have matured into clinical products and gained regulatory approval for clinical use. Finally, we highlight the main hurdles that stand in the way of the digitalization of pathology and the application of artificial intelligence in pathology while also discussing possible solutions.

Graphical Abstract

1. Introduction

Using artificial intelligence (AI) to better understand the complexities of cancer, as well as to facilitate accurate diagnoses and effective treatments, is one of the most coveted goals in modern medicine. After several decades of “AI winter”, the last several years have seen increasing enthusiasm for the potential solutions that AI could offer patients [1,2]. Several innovations, such as the advent of deep learning (DL), the increased digitalization of patient data, and cloud computing, have swayed expert opinion and generated a greater level of openness towards to the adoption of AI [3].
Pathology is the cornerstone of modern oncology, and its role has expanded far beyond offering correct diagnoses. Pathology is playing an ever-increasing role in determining the treatment and prognoses of cancer patients [4]. Advances in whole-slide imaging (WSI) technology and computer hardware have increased the adoption of digital pathology and paved the way for the implementation of AI solutions [5]. These solutions can aid pathologists in tumor classification and tumor grading and can offer patients additional prognostic information derived from traditional histopathology, assisting the prediction of survival or molecular alterations [6]. Despite extensive research in this field, very few solutions have made their way into routine clinical practice due to financial constraints and various regulatory hurdles [7]. Furthermore, there is a lack of studies assessing the impact of AI in clinical settings [8]. When taking into consideration the lack of digitalization and infrastructure characteristic of resource-poor medical settings, the challenge of bridging the gap between research and patients seems even more daunting. Examples of important clinical applications that have been developed in the field can be seen in the grading of prostate cancer, the evaluation of biomarkers such as Ki-67, and the prediction of microsatellite instability in colorectal cancer (CRC) from routine H&E WSI [9,10,11].
The primary objective of this review is to provide a comprehensive overview of the current research landscape in the field of computational pathology, with a focus on applications with that could have an immediate impact on oncology patients. These applications are summarized in Table S1 of the Supplementary Materials. This paper also discusses the issues associated with the clinical implementation of AI and summarizes the most important applications that have gained regulatory approval and made their way to the clinic while also highlighting the problem of cost-efficiency. Finally, we also explore the implementation of AI in pathology in resource-poor settings, discussing challenges and potential avenues for increasing digitalization and the adoption of AI.

2. Artificial Intelligence in Pathology: Research

2.1. AI Tools for Tumor Classification

The classification of tumors is one of the main challenges pathologists face in routine practice. DL provides powerful algorithms for image classification, such as convolutional neural networks (CNNs) that have been widely used in classifying tumors [12]. A DL model trained on 579 whole-slide images from one academic hospital has been used to classify whole-slide images of transbronchial lung biopsies into adenocarcinomas, squamous cell carcinomas, small-cell lung carcinomas, and non-neoplastic diseases, achieving an area under the curve (AUC) of 0.94 to 0.99 in four independent test sets [13]. Another study reported similar results for a model trained on 741 whole-slide images and tested on independent datasets collected from multiple centers (AUCs of 0.918–0.978) in a six-category classification task that included the main types of lung cancer, as well as tuberculosis, organizing pneumonia and normal lung disease [14]. In gastrointestinal pathology, researchers have used CNNs and recurrent neural networks (RNNs) trained on 4128 gastric cancer whole-slide images and 4036 colon cancer whole-slide images to classify biopsies into adenocarcinomas, adenomas, and non-neoplastic diseases in the stomach and colon, with AUCs ranging from 0.96 to 0.99 [15]. A CNN model trained on 1865 images from 703 patients performed a six-class classification task for both CRC adenocarcinomas and the main types of benign colonic polyps, with a mean diagnostic accuracy of 97.3% [16]. Applications in prostate cancer grading have been highly successful, with a study in The Lancet Oncology reporting a model trained on 5759 biopsies with an AUC of 0.99 for classifying benign versus malignant lesions, 0.978 in identifying a grade group of 2 or more, and 0.974 in identifying a grade group of 3 or more while also scoring higher than a panel of 15 pathologists, outperforming 10 of them [9]. A paper from the same group showed that a panel of 14 pathologists had greater agreement with an expert reference standard when performing Gleason grading with the help of AI on a dataset of 160 biopsies (quadratically weighted Cohen’s kappa 0.799 vs. 0.872; p = 0.019) [17]. A DL model trained on 32,537 whole-slide images has been used to predict the origin of cancers of unknown primaries, achieving concordance in 61% of cases and a top-three agreement of 82% [18].

2.2. AI Tools for Biomarker Quantification

The quantification of biomarkers has an increasingly important role in the diagnosis, prognosis, and treatment of cancer patients. However, the process of quantification is plagued by high interobserver variability, which has been documented for numerous biomarkers [19,20,21]. A HER2 scoring algorithm that performed cell identification via the heatmap regression of a CNN combined with augmented reality microscopy was used to improve the accuracy of HER2 interpretation (0.93 vs. 0.80) and improve consistency for HER2 1+ cases [22]. A contest for HER2 scoring using AI was started to provide a benchmark for comparing the performance of various algorithms [23]. One study developed an DL-assisted method for PD-L1 scoring in breast cancer using 100 tumor resection samples, and the method improved concordance across the evaluating pathologists from moderate (0.674) to excellent (0.950) [24]. In the context of non-small-cell lung cancer (NSCLC), one study used QuPath, a free open-source tool, in order to automate PD-L1 scoring by training a random tree classifier using 67 different morphological features [25]. The reported concordance correlation coefficient between the automated and manual scores was 0.925. Automated scoring showed the same level of concordance with the average of three pathologists’ scores as an individual pathologist at both the 1% and 50% thresholds [26]. Another study developed a deep learning architecture with three modules—tumor segmentation (Res50-UNet), nuclei detection (MicroNet), and positive membrane detection (PMD)—that predicted the tumor proportion score (TPS) with a high correlation with a subspecialist pathologist (mean absolute error of 8.65 and Pearson correlation coefficient of 0.9436). The method outperformed generalist and trainee pathologists [27]. In another study, the AI-based evaluation of Ki-67 in breast cancer was performed using a workflow based on the Inception v3 and ResNet architectures, and the findings were compared with the Ki-67 standard reference card assessment, with the authors reporting an intra-group correlation coefficient larger than 0.905 among the pathologists that used these two methods [28]. Another study used a deep learning model based on DenseNet in concert with fuzzy-set interpretations for hot-spot identification to assess the Ki-67 proliferation index in in situ ductal breast carcinomas and reported a mean absolute error of 0.024 [29].

2.3. AI Tools for Survival Prediction

The estimation of survival time is of vital importance in patient management. The semi-parametric Cox proportional hazards model is the preferred method for modeling patient survival. This approach is based on a linear model, which makes it unable to capture non-linearities between independent variables and the risk of death. DL models provide a way of capturing non-linear relationships by using non-linear activation functions; therefore, they can refine the Cox model [30,31]. One study approached this problem by extracting tumor patches (images of 512 × 512 pixels) from whole-slide images of bladder urothelial carcinomas, invasive breast carcinomas, and glioblastomas using a pre-trained tumor classifier and applying graph-based CNNs to aggregate the information across patches and predict survival. The mean absolute errors (MAEs) of the model were 123.2, 167.5, and 303.3 for bladder carcinomas, breast carcinomas, and glioblastomas, respectively [32]. Another paper explored the use of adaptive sampling to extract patches from whole-slide images. The patches were later clustered, and a DL model was trained on these clusters to make patient-level predictions, with C-index values ranging from 0.510 to 0.703 [33]. MultiSurv is a model architecture that was developed to take advantage of multimodal data, including WSI patches, clinical data, copy number variation, gene and miRNA expression, and DNA methylation. This model was applied to 33 different cancer types, producing non-proportional estimates of survival, with time-dependent C-indexes ranging from 0.554 to 0.988 [34].

2.4. AI Tools for Predicting Molecular Alterations

With the advent of the genomic era in medicine, molecular pathology has occupied an ever-increasing role in the management of cancer patients. However, given the expertise and equipment needed to run molecular assays, accessibility remains an issue. Multiple research efforts have explored whether AI could help predict the presence or absence of clinically relevant mutations from H&E-stained whole-slide images (Figure 1).
A paper by Coudray et al. in 2018 showed that a deep CNN based on the Inception v3 architecture and trained on 1.634 WSI from TCGA can predict sx common mutations in NSCLC from whole-slide images, including EGFR, FAT1, SETBP1, KRAS, and TP53, with AUCs ranging from 0.733 to 0.856 [35]. A transformer-based pipeline was trained and evaluated on a multicenter cohort of over 13,000 patients for the end-to-end prediction of microsatellite instability on CRC surgical resection specimens, achieving a sensitivity of 0.99 and a negative predictive value of over 0.99 [36]. An interpretable attention-based DL model trained on images from 1240 patients was used to predict the major molecular subgroups in endometrial cancer, with AUROCs of 0.849 for POLE ultra-mutation, 0.844 for DNA mismatch repair (MMR) deficiency, 0.928 for p53-abnormal, and 0.883 for no specific molecular profile (NSMP) [37]. One study attempted to predict the ten most common mutations in hepatocellular carcinoma (HCC) using a deep learning model based on the Inception v3 architecture and trained on 491 whole-slide images. The model predicted four of them (CTNNB1, FMN2, TP53, and ZFX4), with AUCs on external cohorts ranging from 0.71 to 0.89 [38]. Another study developed a ResNet-18 based on a model trained on images from 307 patients in order to predict IDH mutation status, an important prognostic biomarker in low-grade gliomas, with an AUC of 0.667 [39]. Hierarchical deep multi-instance learning was used to predict mutations of five clinically relevant genes in the TCGA bladder cancer dataset (ATM, PIK3CA, ERBB2, FGFR3, ERCC2), with AUCs above 0.83 [40]. Regarding breast cancer, a study showed that a DL model based on the ResNet-101 architecture and trained on 461 whole-slide images can be used to predict mutations in TP53, RB1, CDH1, NF1, and NOTCH2, with AUCs of 0.729 to 0.852 [41], while another study managed to predict germline BRCA mutations from whole-slide images with an AUC of 0.766 using a ResNet architecture [42]. One group reported somatic BRCA mutation prediction in high-grade ovarian cancer with an AUC of 0.681 using a ResNet-based architecture trained on images from 867 patients [43]. In an effort to help differentiate noninvasive follicular thyroid neoplasms with papillary-like nuclear features (NIFTP) from other neoplasms, an Xception-based DL model was trained on 115 whole-slide images to predict the BRAF-RAS score. This predicted score had an AUC of 0.99 and 0.97 when restricted to follicular pattern neoplasms [44]. Another study attempted to predict BRAF mutation in melanoma by using deep learning alone (AUC = 0.71) and by combining deep learning (Inception v3 architecture) with clinical information and pathomics, achieving an AUC of 0.89 [45]. Lung adenocarcinoma whole-slide images were used to whether the predict tumor mutational burden (TMB) was high or low via a model based on the Inception v3 architecture and trained on images from 499 patients, achieving an AUC of 0.92 on a held-out dataset [46].

2.5. Generative AI and Synthetic Data

Generating ground truth for the development of AI algorithms is both expensive and time-consuming. Generative adversarial networks (GANs) could provide models with labeled synthetic data, bypassing these issues [47]. PathologyGAN is an approach that uses a relativistic average discriminator that generates high-quality images of breast cancer and CRC, achieving a Frechet Inception Distance (FID) of 16.65 (breast cancer) and 32.05 (CRC). Ratings from two expert pathologists found no significant differences between the generated images and real ones, [48]. Wei et al. used CycleGANs to generate synthetic images of colorectal polyps which at least two out of four gastrointestinal pathologists could not discriminate from real images during a Turing test while also using the synthetic images to improve a ResNet classifier AUC by over 10% [49]. Another potential application of GANs is the virtual staining of histological slides, which could reduce costs as well as variability in staining [50].

3. Artificial Intelligence in Pathology: Clinical-Grade Tools

3.1. Implementation of AI in Routine Practice

Despite the large amount of published research on the applications of AI in pathology, a very small number of research projects go on to become mature commercial products that can be used in routine clinical practice. Research generates prototypes that must be further developed, validated on multiple independent datasets, and integrated into existing clinical laboratory workflows [51]. Clinical AI algorithms are considered medical devices, therefore necessitating regulatory approval before being made commercially available.
One major initial technical challenge is developing high-quality software that is easy to maintain and update. The computer code used in research projects is usually poorly structured and error-prone. Software design principles like object orientation and rigorous testing must be implemented to ensure code that is stable and works properly [52]. The level of maturity of AI solutions can be evaluated using technology readiness levels (TRLs) [53]. Homeyer et al. adapted TRLs for AI solutions in pathology. There are a total of nine levels, with higher ones indicating more mature products (Figure 2). A research prototype which is only evaluated on a limited dataset in a peer-reviewed publication is level 3, a clinical prototype that has been evaluated in routine pathology practice is level 6, while a solution that has cleared regulation and has achieved the status of commercial product is level 9 [54]. The validation of AI algorithms requires multiple datasets that are representative of the target population to avoid biases [55]. Moreover, obtaining regulatory approval is challenging due to the lack of transparency and model explainability, as well as adaptive training, which can cause changes in an AI solution after it gains regulatory clearance. The final step is reimbursement, which can only be obtained after the solution’s clinical utility has been demonstrated.

3.2. Examples of AI Tools Approved for Clinical Use in the USA and/or EU

The Galen platform, developed by Ibex Medical Analytics, represents a suite of AI solutions used to facilitate diagnosis across multiple organs and pathologies. The first of these solutions is Galen Breast, currently CE-marked under the In Vitro Diagnostic Directive (IVDD) in the EU and registered with the UK Medicines and Healthcare products Regulatory Agency (MHRA), as well as with the Brazilian Health Regulatory Agency (ANVISA). The validation and real-world clinical application study of the solution was published in December 2022. The study showed that Galen Breast can detect invasive carcinoma with an AUC of 0.99 (specificity of 93.57% and sensitivity of 95.51%) and DCIS with an AUC of 0.98 (specificity of 93.79% and sensitivity of 93.20%). The algorithm can also differentiate between invasive ductal carcinoma and invasive lobular carcinoma with an AUC of 0.97 and make the distinction between high-grade DCIS and low-grade DCIS/atypical ductal hyperplasia with an AUC of 0.92. Tumor-infiltrating lymphocytes (TILs) were also identified with an AUC of 0.965 [56]. Galen Prostate is CE-marked under IVDD, as well as under IVDR, and registered with the UK MHRA and ANVISA. The validation study for this tool was published in August 2020. The model detected cancer with an AUC of 0.991 in the external validation set. It differentiated between low-grade (Gleason score 6 or atypical small acinar proliferation) and high-grade (Gleason score 7–10) with an AUC of 0.941. Gleason pattern 5 was detected with an AUC of 0.971, and the cancer percentage calculated by the pathologists in the study, and the algorithm showed good correlation (r = 0.882, p < 0.0001), with a mean bias of −4.14%. The algorithm also achieved an AUC of 0.957 for the detection of perineural invasion [57]. Galen Gastric is CE-marked under IVDD, with a validation study presented as a poster at USCAP 2022. On real-world data, the solution demonstrated an AUC of 0.945 for detecting gastric adenocarcinoma and high-grade dysplasia and an AUC of 0.966 for the detection of Helicobacter pylori [58].
HALO Prostate AI, developed by Indica Labs is a DL-based screening tool currently CE-marked under IVDD, and its validation study was published in August 2023. The model identified biopsies that contained tumors with sensitivity, specificity, and negative predictive values ranging from 0.971 to 1.000, 0.875 to 0.976, and 0.988 to 1.000, respectively, across multiple test cohorts [59]. Histotype Px, by DoMore Diagnostics (CE-marked under IVDD), uses deep learning to provide risk stratification for patients with CRC using WSI. The validation study of the product was published in September 2022. This method provided hazard ratios of 10.71 for high-risk versus low-risk patients and 3.06 for intermediate- versus low-risk patients. Estimates of 3-year cancer-specific survival were 97.2% for the low-risk group, 94.8% for the intermediate-risk group, and 77.6% for the high-risk group. Risk stratification based on current guidelines was less prognostic, with the low-risk group being represented by only 13% of the validation cohort versus 41% in the DL based risk stratification [60]. MSIntuit, produced by Owkin and CE-marked under IVDD, is a tool that uses DL to identify MSI status from H&E WSI, and it achieved a sensitivity and specificity of 97% in a blinded validation study, with excellent agreement on two different scanners (Cohen’s kappa: 0.82) [61].
Paige’s Prostate Suite applications are CE-marked under IVDD, and Paige Prostate Detect is also the only FDA-approved AI solution for pathology. In one validation study, Paige Prostate Detect was reported to have a sensitivity of 97.4% and a specificity of 94.8% at the WSI level, with an AUC of 0.99. Pathologist sensitivity and specificity in diagnosis with and without Paige Prostate were measured, with an increase in accuracy from 88.7% unassisted to 96.6% when assisted, reducing detection errors by 70% [62]. Another study reported that pathologists diagnosed atypical small acinar proliferation about 30% less often while also requesting fewer IHC studies (around 20% less) and fewer second opinions (around 40% less) when assisted by Paige Prostate. The median time required to read and report slides was about 20% lower when assisted [63].
Visiopharm offers EU IVDR-certified tools for breast cancer, including the IHC quantification of Ki-67, estrogen receptor (ER), progesterone receptor (PR), and HER2, as well as the detection of tumors in pan cytokeratin-stained slides, hot spot detection, invasive tumor detection, and metastasis detection [64,65]. Aiforia has developed multiple AI models that are CE-marked under IVDD for breast cancer (ER, PR, Ki-67 scoring), prostate cancer (Gleason grading), and lung cancer (PD-L1 scoring) [66]. Mindpeak offers a suite of tools that have received CE marking under IVDD for lung cancer (PD-L1 scoring) and breast cancer (ER, PR, Ki-67, HER2 scoring) [10]. Deep Bio has developed a CE-marked under IVDD deep learning tool for prostate cancer grading, showing high concordance between the model and the reference standard (quadratic-weighted Cohen’s kappa of 0.907) [67].

3.3. The Problem of Cost-Efficiency

One of the main advantages invoked for digital pathology and AI implementation is increased efficiency, particularly in fully digital workflows. One study calculated projected cost savings in the first 5 years after the implementation of a digital pathology system. Based on potential improvements in productivity and laboratory consolidation, the potential savings were estimated to be USD 12.4 million in an institution with 219,000 annual accessions [68]. Another cost–benefit analysis was performed for a department of 45 full-time consultant pathologists, 80,000 specimens per year, and a yearly budget of GDP 9 million. The results showed that an increase in productivity of 10% would allow the institution to break even in 2 years, while an increase of 15% would allow the institution to break even after 1 year. A pathology laboratory with half the volume would make a profit in 4 years after a productivity improvement of 10% [69]. An analysis from the Memorial Sloan Kettering Cancer Center, published in 2019, reported that digital pathology implementation could lead to a 93% reduction in the requests for physical slides. In addition, the authors estimated yearly IHC savings of USD 114,000 due to fewer ancillary tests being ordered. Overall cost savings were estimated at USD 267,000 per year. Factoring in digital workflow setup and maintenance costs, it was estimated that the breakeven point would be 7 years after the initial transition [70]. A Delphi study on the role of computational pathology in 2030 revealed strong agreement among experts regarding the fact that the regulatory requirements will generate important costs in the future because of the need for large-scale prospective studies. No consensus was reached concerning a potential decrease in cost-per-case and second-opinion consultations. Most experts predicted that cost-per-case would not decrease in the following 8–10 years [71]. Another study evaluated the financial impact of implementing an AI-based evaluation of MSI status in a hypothetical population of 32,549 patients receiving first-line treatment for newly diagnosed metastatic CRC. Eight different testing strategies were compared: (1) next-generation sequencing (NGS), (2) high-sensitivity polymerase chain (PCR) reaction or IHC panel, (3) high-specificity IHC panel, (4) high-specificity AI, (5) high-sensitivity AI followed by NGS, (6) high-specificity AI followed by NGS, (7) high-sensitivity AI and high-sensitivity IHC panel, and (8) high-sensitivity AI and high-specificity panel. High-sensitivity AI followed by confirmatory high-specificity PCR or IHC panel for the patients that test negative using AI proved to be the strategy that resulted in the largest savings (USD 400 million, 12.9%) when compared to NGS alone [72].

4. Digitalization and AI in Countries with Scarce Resources

4.1. The Challenge of Digitalization

The process of digitalization represents the first step needed to implement AI in pathology. Digital pathology is being progressively deployed in more and more departments across the world, with several papers reporting the experience of both public and private pathology laboratories [73,74,75,76]. However, securing funding for digitalization is difficult. The large initial investments associated with WSI scanners, associated IT infrastructure, and the storage requirements needed for the large sizes of whole-slide images (up to several Gb per image) are important hurdles, particularly in low-resource settings, where the funding of pathology departments is lacking.
Digitalization generally follows one of two implementation models: a “hybrid” approach where the digital workflow is gradually introduced [74] and a second model that consists of adopting a fully digital workflow very shortly after implementation [77]. The main advantage of the “hybrid” model is that it allows pathologists to gradually tackle the learning curve of working with whole-slide images while still having access to glass slides. The fully digital workflow has the major advantage of saving the time of histotechnicians that would otherwise be spent on slide sorting and distribution while also allowing pathologists to benefit immediately from various digital pathology tools [77].
Before implementing the digital workflow, it is necessary to have a dedicated information technology infrastructure in place. This includes a broadband internet connection, a sample tracking system integrated with the laboratory information system (LIS), a WSI scanner, and pathologists’ workstations [78]. Implementing the required infrastructure, together with subsequent quality-control tests and validation, can take up to several months [74,76]. The adoption of digital pathology allows for easier sharing of cases, enabling telepathology and remote consultations (Figure 3). Furthermore, it provides improved efficiency and productivity, as well as a smaller rate of interpretive errors [68]. Integration with molecular pathology and genomic research databases could further facilitate the use of digital pathology in clinical practice and research [75].

4.2. The Challenges of Pathology and Digitalization in Developing Countries

The increasing burden of cancer in developing countries is creating new challenges given the lack of economic and healthcare infrastructure. The lack of proper diagnoses and treatments decreases the effectiveness of cancer control, which is based on surveillance and screening programs. These approaches only benefit patients if there is access to appropriate patient management once an early diagnosis of cancer has been made [79]. It has been found that the human development index has a significant negative correlation with the cancer mortality/incidence ratio [80].
As a critical first step in the management of cancer patients, pathology services are also severely underdeveloped in many parts of the world. Many of the scarce resources available are invested in offering treatment rather than diagnostic services [81]. According to a report from 2023, there are over 108,000 active pathologists in 162 countries and territories, corresponding to an average of 14 pathologists per million inhabitants. The disparities are glaring, with the United States having access to 65 pathologists per million, while African countries, on average, have less than 3 pathologists per million. Over two-thirds of the global pathologist workforce resides in 10 countries [82]. Other challenges faced by pathology as a speciality in developing countries include the variable standards of training, long turnaround times, lack of quality synoptic reporting, and the lack of visibility of pathology as a key component in the multidisciplinary clinical team [81]. Even though digitalization could increase accessibility to quality pathology services via telepathology or increase productivity in areas with insufficient pathologists, it requires a significant initial investment before benefits can be reaped. This might deter pathology laboratories in countries with scarce resources from making the transition to a digital workflow [74].

4.3. Telepathology as a Stepping Stone for Digitalization in Developing Countries

Telepathology is considered one of the main tools for overcoming pathologist shortages, with several successful examples of implementation being present in the literature [83,84,85]. The are four main approaches to telepathology: static images, whole-slide image scanning, dynamic nonrobotic telemicroscopy and dynamic robotic telemicroscopy. The static image approach requires minimal technical infrastructure but necessitates an appropriate selection of relevant diagnostic fields. Following the training period, static image telepathology diagnoses showed a concordance of 97% with a subsequent review of slides. WSI does not require the selection of diagnostic fields and allows pathologists to examine entire specimens at different magnifications. However, the costs of acquiring slide scanners and setting up the appropriate IT infrastructure and support prove quite prohibitive in low-resource environments. Dynamic nonrobotic telemicroscopy represents the video transmission of a microscope image in real-time across the internet, with a pathologist “driving” the microscope. This method is dependent on the quality of the internet connection and on image resolution but provides whole-slide imaging at a much lower cost. Dynamic robotic telemicroscopy allows a pathologist to control the microscope remotely. This approach is used less frequently due to costs and technical issues [86].

4.4. The Potential Uses of Social Media

Social media could help bridge knowledge gaps among pathologists and foster an inclusive environment for pathologists from developing countries. One retrospective survey-based study aimed to assess whether medical students that used Twitter during their pathology rotation found that it facilitated learning. Most respondents thought that Twitter (currently X) helped them retain more information while also being useful for professional development [87]. Twitter has also been described as a potential tool for case consultation and research collaborations, especially for pathologists in the developing world, who have less well-developed professional networks [88]. An article addressing the ethical issues of pathology and social media highlighted the fact that there are no significant differences regarding patient privacy when images are shared on social media platforms compared to image publication in a medical journal and that, with responsible use, the benefits far outweigh the risks [89].
Recently, given the growing number of pathologists creating content, social media has also proved to be a useful resource for the training of AI models. A group of researchers used pathology images from tweets, as well as images from published literature, to develop multimodal DL models that identify histopathology stains, classify tissues, and differentiate disease states with AUROCs of 0.805 to 0.996. Based on these models, the researchers created a social media bot that can help pathologists obtain real-time feedback on the cases that they post [90]. A recent study developed a method called pathology language-image pretraining (PLIP), which is a multimodal AI model trained on OpenPath, a dataset of pathology images curated from Twitter hashtags paired with natural language descriptions and pathology data from other websites. PLIP was then tested on four external datasets and achieved state-of-the-art results for zero-shot classification tasks, with F1 scores of 0.565–0.832. PLIP also allows users to find similar cases using image or natural language searches [91].

5. Conclusions and Future Directions

In this review, we provided an overview of the main research directions in computational pathology applied to oncology. We then discussed the challenges associated with bridging the gap from research to the clinic and highlighted the most important tools that have made their way through regulatory hurdles and are available for clinical use. We also discussed the avenues by which digital pathology and AI can be implemented in low-resource settings while also considering the existing limitations.
A major issue identified was the striking discrepancy between the burgeoning research landscape of computational pathology and the scarcity of tools available for clinical use by pathologists. Most research projects produce intriguing prototypes that must be further developed to become clinical-grade tools. This development process requires making significant investments and passing ever-increasing regulatory hurdles, which discourages public sector involvement. In the future, more public and non-profit interest in computational pathology could greatly increase access to pathology services in developing countries and lower resource settings.
In conclusion, the potential of AI in the field of pathology has yet to be fully realized. The lack of digitalization of pathology services in developing countries generates difficulties when it comes to being involved in the creation of AI tools. It is essential for researchers, clinicians, companies, and various other stakeholders to collaborate to overcome future challenges and offer cancer patients equitable access to AI-enhanced pathology services.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers15235692/s1, Table S1: The applications of computational pathology with an immediate potential impact on oncology patients are summarized.

Author Contributions

Conceptualization, G.-E.O., A.V., M.M.K. and O.P.; methodology, G.-E.O.; software, A.V.; investigation, A.V.; resources, G.-E.O.; data curation, A.V.; writing—original draft preparation, A.V.; writing—review and editing, A.V. and G.-E.O.; visualization, A.V. and M.M.K.; supervision, G.-E.O. and O.P.; project administration, G.-E.O.; funding acquisition, O.P. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The graphical abstract and figures were created using BioRender.com.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Shreve, J.T.; Khanani, S.A.; Haddad, T.C. Artificial Intelligence in Oncology: Current Capabilities, Future Opportunities, and Ethical Considerations. Am. Soc. Clin. Oncol. Educ. Book 2022, 42, 842–851. [Google Scholar] [CrossRef] [PubMed]
  2. Topol, E.J. High-performance medicine: The convergence of human and artificial intelligence. Nat. Med. 2019, 25, 44–56. [Google Scholar] [CrossRef] [PubMed]
  3. Pedro, A.R.; Dias, M.B.; Laranjo, L.; Cunha, A.S.; Cordeiro, J.V. Artificial intelligence in medicine: A comprehensive survey of medical doctor’s perspectives in Portugal. PLoS ONE 2023, 18, e0290613. [Google Scholar] [CrossRef]
  4. on behalf of the European Society of Pathology (ESP); Matias-Guiu, X.; Stanta, G.; Carneiro, F.; Ryska, A.; Hoefler, G.; Moch, H. The leading role of pathology in assessing the somatic molecular alterations of cancer: Position Paper of the European Society of Pathology. Virchows Arch. 2020, 476, 491–497. [Google Scholar] [CrossRef]
  5. Jahn, S.W.; Plass, M.; Moinfar, F. Digital Pathology: Advantages, Limitations and Emerging Perspectives. J. Clin. Med. 2020, 9, 3697. [Google Scholar] [CrossRef] [PubMed]
  6. Wu, Y.; Cheng, M.; Huang, S.; Pei, Z.; Zuo, Y.; Liu, J.; Yang, K.; Zhu, Q.; Zhang, J.; Hong, H.; et al. Recent Advances of Deep Learning for Computational Histopathology: Principles and Applications. Cancers 2022, 14, 1199. [Google Scholar] [CrossRef]
  7. Steiner, D.F.; Chen, P.-H.C.; Mermel, C.H. Closing the translation gap: AI applications in digital pathology. Biochim. Biophys. Acta (BBA) Rev. Cancer 2021, 1875, 188452. [Google Scholar] [CrossRef]
  8. Kearney, S.J.; Lowe, A.; Lennerz, J.K.; Parwani, A.; Bui, M.M.; Wack, K.; Giannini, G.; Abels, E. Bridging the Gap: The Critical Role of Regulatory Affairs and Clinical Affairs in the Total Product Life Cycle of Pathology Imaging Devices and Software. Front. Med. 2021, 8, 765385. [Google Scholar] [CrossRef]
  9. Bulten, W.; Pinckaers, H.; van Boven, H.; Vink, R.; de Bel, T.; van Ginneken, B.; van der Laak, J.; Hulsbergen-van de Kaa, C.; Litjens, G. Automated deep-learning system for Gleason grading of prostate cancer using biopsies: A diagnostic study. Lancet Oncol. 2020, 21, 233–241. [Google Scholar] [CrossRef]
  10. Abele, N.; Tiemann, K.; Krech, T.; Wellmann, A.; Schaaf, C.; Länger, F.; Peters, A.; Donner, A.; Keil, F.; Daifalla, K.; et al. Noninferiority of Artificial Intelligence–Assisted Analysis of Ki-67 and Estrogen/Progesterone Receptor in Breast Cancer Routine Diagnostics. Mod. Pathol. 2023, 36, 100033. [Google Scholar] [CrossRef]
  11. Echle, A.; Laleh, N.G.; Quirke, P.; Grabsch, H.; Muti, H.; Saldanha, O.; Brockmoeller, S.; Brandt, P.v.D.; Hutchins, G.; Richman, S.; et al. Artificial intelligence for detection of microsatellite instability in colorectal cancer—A multicentric analysis of a pre-screening tool for clinical application. ESMO Open 2022, 7, 100400. [Google Scholar] [CrossRef]
  12. Echle, A.; Rindtorff, N.T.; Brinker, T.J.; Luedde, T.; Pearson, A.T.; Kather, J.N. Deep learning in cancer pathology: A new generation of clinical biomarkers. Br. J. Cancer 2021, 124, 686–696. [Google Scholar] [CrossRef]
  13. Kanavati, F.; Toyokawa, G.; Momosaki, S.; Takeoka, H.; Okamoto, M.; Yamazaki, K.; Takeo, S.; Iizuka, O.; Tsuneki, M. A deep learning model for the classification of indeterminate lung carcinoma in biopsy whole slide images. Sci. Rep. 2021, 11, 8110. [Google Scholar] [CrossRef]
  14. Yang, H.; Chen, L.; Cheng, Z.; Yang, M.; Wang, J.; Lin, C.; Wang, Y.; Huang, L.; Chen, Y.; Peng, S.; et al. Deep learning-based six-type classifier for lung cancer and mimics from histopathological whole slide images: A retrospective study. BMC Med. 2021, 19, 80. [Google Scholar] [CrossRef]
  15. Iizuka, O.; Kanavati, F.; Kato, K.; Rambeau, M.; Arihiro, K.; Tsuneki, M. Deep Learning Models for Histopathological Classification of Gastric and Colonic Epithelial Tumours. Sci. Rep. 2020, 10, 1504. [Google Scholar] [CrossRef]
  16. Byeon, S.-J.; Park, J.; Cho, Y.A.; Cho, B.-J. Automated histological classification for digital pathology images of colonoscopy specimen via deep learning. Sci. Rep. 2022, 12, 12804. [Google Scholar] [CrossRef] [PubMed]
  17. Bulten, W.; Balkenhol, M.; Belinga, J.-J.A.; Brilhante, A.; Çakır, A.; Egevad, L.; Eklund, M.; Farré, X.; Geronatsiou, K.; Molinié, V.; et al. Artificial intelligence assistance significantly improves Gleason grading of prostate biopsies by pathologists. Mod. Pathol. 2021, 34, 660–671. [Google Scholar] [CrossRef]
  18. Lu, M.Y.; Chen, T.Y.; Williamson, D.F.K.; Zhao, M.; Shady, M.; Lipkova, J.; Mahmood, F. AI-based pathology predicts origins for cancers of unknown primary. Nature 2021, 594, 106–110. [Google Scholar] [CrossRef]
  19. Baez-Navarro, X.; van Bockstal, M.R.; Nawawi, D.; Broeckx, G.; Colpaert, C.; Doebar, S.C.; Hogenes, M.C.; Koop, E.; Lambein, K.; Peeters, D.J.; et al. Interobserver Variation in the Assessment of Immunohistochemistry Expression Levels in HER2-Negative Breast Cancer: Can We Improve the Identification of Low Levels of HER2 Expression by Adjusting the Criteria? An International Interobserver Study. Mod. Pathol. 2023, 36, 100009. [Google Scholar] [CrossRef] [PubMed]
  20. Robert, M.E.; Rüschoff, J.; Jasani, B.; Graham, R.P.; Badve, S.S.; Rodriguez-Justo, M.; Kodach, L.L.; Srivastava, A.; Wang, H.L.; Tang, L.H.; et al. High Interobserver Variability Among Pathologists Using Combined Positive Score to Evaluate PD-L1 Expression in Gastric, Gastroesophageal Junction, and Esophageal Adenocarcinoma. Mod. Pathol. 2023, 36, 100154. [Google Scholar] [CrossRef] [PubMed]
  21. Butter, R.; Hondelink, L.M.; van Elswijk, L.; Blaauwgeers, J.L.; Bloemena, E.; Britstra, R.; Bulkmans, N.; van Gullik, A.L.; Monkhorst, K.; de Rooij, M.J.; et al. The impact of a pathologist’s personality on the interobserver variability and diagnostic accuracy of predictive PD-L1 immunohistochemistry in lung cancer. Lung Cancer 2022, 166, 143–149. [Google Scholar] [CrossRef] [PubMed]
  22. Wu, S.; Yue, M.; Zhang, J.; Li, X.; Li, Z.; Zhang, H.; Wang, X.; Han, X.; Cai, L.; Shang, J.; et al. The Role of Artificial Intelligence in Accurate Interpretation of HER2 Immunohistochemical Scores 0 and 1+ in Breast Cancer. Mod. Pathol. 2023, 36, 100054. [Google Scholar] [CrossRef]
  23. Qaiser, T.; Mukherjee, A.; Pb, C.R.; Munugoti, S.D.; Tallam, V.; Pitkäaho, T.; Lehtimäki, T.; Naughton, T.; Berseth, M.; Pedraza, A.; et al. HER2 challenge contest: A detailed assessment of automated HER2 scoring algorithms in whole slide images of breast cancer tissues. Histopathology 2018, 72, 227–238. [Google Scholar] [CrossRef]
  24. Wang, X.; Wang, L.; Bu, H.; Zhang, N.; Yue, M.; Jia, Z.; Cai, L.; He, J.; Wang, Y.; Xu, X.; et al. How can artificial intelligence models assist PD-L1 expression scoring in breast cancer: Results of multi-institutional ring studies. NPJ Breast Cancer 2021, 7, 61. [Google Scholar] [CrossRef]
  25. Bankhead, P.; Loughrey, M.B.; Fernández, J.A.; Dombrowski, Y.; McArt, D.G.; Dunne, P.D.; McQuaid, S.; Gray, R.T.; Murray, L.J.; Coleman, H.G.; et al. QuPath: Open source software for digital pathology image analysis. Sci. Rep. 2017, 7, 16878. [Google Scholar] [CrossRef] [PubMed]
  26. Naso, J.R.; Povshedna, T.; Wang, G.; Banyi, N.; MacAulay, C.; Ionescu, D.N.; Zhou, C. Automated PD-L1 Scoring for Non-Small Cell Lung Carcinoma Using Open-Source Software. Pathol. Oncol. Res. 2021, 27, 609717. [Google Scholar] [CrossRef]
  27. Liu, J.; Zheng, Q.; Mu, X.; Zuo, Y.; Xu, B.; Jin, Y.; Wang, Y.; Tian, H.; Yang, Y.; Xue, Q.; et al. Automated tumor proportion score analysis for PD-L1 (22C3) expression in lung squamous cell carcinoma. Sci. Rep. 2021, 11, 15907. [Google Scholar] [CrossRef]
  28. Li, L.; Han, D.; Yu, Y.; Li, J.; Liu, Y. Artificial intelligence-assisted interpretation of Ki-67 expression and repeatability in breast cancer. Diagn. Pathol. 2022, 17, 20. [Google Scholar] [CrossRef]
  29. Fulawka, L.; Blaszczyk, J.; Tabakov, M.; Halon, A. Assessment of Ki-67 proliferation index with deep learning in DCIS (ductal carcinoma in situ). Sci. Rep. 2022, 12, 3166. [Google Scholar] [CrossRef]
  30. Zhu, W.; Xie, L.; Han, J.; Guo, X. The Application of Deep Learning in Cancer Prognosis Prediction. Cancers 2020, 12, 603. [Google Scholar] [CrossRef] [PubMed]
  31. Katzman, J.L.; Shaham, U.; Cloninger, A.; Bates, J.; Jiang, T.; Kluger, Y. DeepSurv: Personalized treatment recommender system using a Cox proportional hazards deep neural network. BMC Med. Res. Methodol. 2018, 18, 24. [Google Scholar] [CrossRef]
  32. Xu, Z.; Lim, S.; Shin, H.-K.; Uhm, K.-H.; Lu, Y.; Jung, S.-W.; Ko, S.-J. Risk-aware survival time prediction from whole slide pathological images. Sci. Rep. 2022, 12, 21948. [Google Scholar] [CrossRef]
  33. Zhu, X.; Yao, J.; Zhu, F.; Huang, J. WSISA: Making Survival Prediction from Whole Slide Histopathological Images. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6855–6863. [Google Scholar]
  34. Vale-Silva, L.A.; Rohr, K. Long-term cancer survival prediction using multimodal deep learning. Sci. Rep. 2021, 11, 13505. [Google Scholar] [CrossRef]
  35. Coudray, N.; Ocampo, P.S.; Sakellaropoulos, T.; Narula, N.; Snuderl, M.; Fenyö, D.; Moreira, A.L.; Razavian, N.; Tsirigos, A. Classification and mutation prediction from non–small cell lung cancer histopathology images using deep learning. Nat. Med. 2018, 24, 1559–1567. [Google Scholar] [CrossRef]
  36. Wagner, S.J.; Reisenbüchler, D.; West, N.P.; Niehues, J.M.; Zhu, J.; Foersch, S.; Veldhuizen, G.P.; Quirke, P.; Grabsch, H.I.; Brandt, P.A.v.D.; et al. Transformer-based biomarker prediction from colorectal cancer histology: A large-scale multicentric study. Cancer Cell 2023, 41, 1650–1661.e4. [Google Scholar] [CrossRef] [PubMed]
  37. Fremond, S.; Andani, S.; Wolf, J.B.; Dijkstra, J.; Melsbach, S.; Jobsen, J.J.; Brinkhuis, M.; Roothaan, S.; Jurgenliemk-Schulz, I.; Lutgens, L.C.H.W.; et al. Interpretable deep learning model to predict the molecular classification of endometrial cancer from haematoxylin and eosin-stained whole-slide images: A combined analysis of the PORTEC randomised trials and clinical cohorts. Lancet Digit. Health 2023, 5, e71–e82. [Google Scholar] [CrossRef] [PubMed]
  38. Chen, M.; Zhang, B.; Topatana, W.; Cao, J.; Zhu, H.; Juengpanich, S.; Mao, Q.; Yu, H.; Cai, X. Classification and mutation prediction based on histopathology H&E images in liver cancer using deep learning. npj Precis. Oncol. 2020, 4, 14. [Google Scholar] [CrossRef] [PubMed]
  39. Jiang, S.; Zanazzi, G.J.; Hassanpour, S. Predicting prognosis and IDH mutation status for patients with lower-grade gliomas using whole slide images. Sci. Rep. 2021, 11, 16849. [Google Scholar] [CrossRef] [PubMed]
  40. Yan, R.; Shen, Y.; Zhang, X.; Xu, P.; Wang, J.; Li, J.; Ren, F.; Ye, D.; Zhou, S.K. Histopathological bladder cancer gene mutation prediction with hierarchical deep multiple-instance learning. Med. Image Anal. 2023, 87, 102824. [Google Scholar] [CrossRef] [PubMed]
  41. Qu, H.; Zhou, M.; Yan, Z.; Wang, H.; Rustgi, V.K.; Zhang, S.; Gevaert, O.; Metaxas, D.N. Genetic mutation and biological pathway prediction based on whole slide images in breast carcinoma using deep learning. npj Precis. Oncol. 2021, 5, 87. [Google Scholar] [CrossRef] [PubMed]
  42. Wang, X.; Zou, C.; Zhang, Y.; Li, X.; Wang, C.; Ke, F.; Chen, J.; Wang, W.; Wang, D.; Xu, X.; et al. Prediction of BRCA Gene Mutation in Breast Cancer Based on Deep Learning and Histopathology Images. Front. Genet. 2021, 12, 661109. [Google Scholar] [CrossRef] [PubMed]
  43. Bourgade, R.; Rabilloud, N.; Perennec, T.; Pécot, T.; Garrec, C.; Guédon, A.F.; Delnatte, C.; Bézieau, S.; Lespagnol, A.; de Tayrac, M.; et al. Deep Learning for Detecting BRCA Mutations in High-Grade Ovarian Cancer Based on an Innovative Tumor Segmentation Method From Whole Slide Images. Mod. Pathol. 2023, 36, 100304. [Google Scholar] [CrossRef] [PubMed]
  44. Dolezal, J.M.; Trzcinska, A.; Liao, C.-Y.; Kochanny, S.; Blair, E.; Agrawal, N.; Keutgen, X.M.; Angelos, P.; Cipriani, N.A.; Pearson, A.T. Deep learning prediction of BRAF-RAS gene expression signature identifies noninvasive follicular thyroid neoplasms with papillary-like nuclear features. Mod. Pathol. 2021, 34, 862–874. [Google Scholar] [CrossRef] [PubMed]
  45. Kim, R.H.; Nomikou, S.; Coudray, N.; Jour, G.; Dawood, Z.; Hong, R.; Esteva, E.; Sakellaropoulos, T.; Donnelly, D.; Moran, U.; et al. Deep Learning and Pathomics Analyses Reveal Cell Nuclei as Important Features for Mutation Prediction of BRAF-Mutated Melanomas. J. Investig. Dermatol. 2022, 142, 1650–1658.e6. [Google Scholar] [CrossRef] [PubMed]
  46. Jain, M.S.; Massoud, T.F. Predicting tumour mutational burden from histopathological images using multiscale deep learning. Nat. Mach. Intell. 2020, 2, 356–362. [Google Scholar] [CrossRef]
  47. Morrison, D.; Harris-Birtill, D.; Caie, P.D. Generative Deep Learning in Digital Pathology Workflows. Am. J. Pathol. 2021, 191, 1717–1723. [Google Scholar] [CrossRef]
  48. Quiros, A.C.; Murray-Smith, R.; Yuan, K. PathologyGAN: Learning deep representations of cancer tissue. arXiv 2021, arXiv:1907.02644. [Google Scholar] [CrossRef]
  49. Wei, J.; Suriawinata, A.; Vaickus, L.; Ren, B.; Liu, X.; Wei, J.; Hassanpour, S. Generative Image Translation for Data Augmentation in Colorectal Histopathology Images. Proc. Mach. Learn. Res. 2019, 116, 10–24. [Google Scholar]
  50. Jose, L.; Liu, S.; Russo, C.; Nadort, A.; Di Ieva, A. Generative Adversarial Networks in Digital Pathology and Histopathological Image Processing: A Review. J. Pathol. Inform. 2021, 12, 43. [Google Scholar] [CrossRef]
  51. Acs, B.; Rantalainen, M.; Hartman, J. Artificial intelligence as the next step towards precision pathology. J. Intern. Med. 2020, 288, 62–81. [Google Scholar] [CrossRef]
  52. Kaur, S.; Singh, P. How does object-oriented code refactoring influence software quality? Research landscape and challenges. J. Syst. Softw. 2019, 157, 110394. [Google Scholar] [CrossRef]
  53. Eljasik-Swoboda, T.; Rathgeber, C.; Hasenauer, R. Assessing Technology Readiness for Artificial Intelligence and Machine Learning Based Innovations. In Proceedings of the 8th International Conference on Data Science, Technology and Applications, Prague, Czech Republic, 26–28 July 2019; pp. 281–288. [Google Scholar]
  54. Homeyer, A.; Lotz, J.; Schwen, L.O.; Weiss, N.; Romberg, D.; Höfener, H.; Zerbe, N.; Hufnagl, P. Artificial Intelligence in Pathology: From Prototype to Product. J. Pathol. Inform. 2021, 12, 13. [Google Scholar] [CrossRef]
  55. Abels, E.; Pantanowitz, L.; Aeffner, F.; Zarella, M.D.; van der Laak, J.; Bui, M.M.; Vemuri, V.N.; Parwani, A.V.; Gibbs, J.; Agosto-Arroyo, E.; et al. Computational pathology definitions, best practices, and recommendations for regulatory guidance: A white paper from the Digital Pathology Association. J. Pathol. 2019, 249, 286–294. [Google Scholar] [CrossRef]
  56. Sandbank, J.; Bataillon, G.; Nudelman, A.; Krasnitsky, I.; Mikulinsky, R.; Bien, L.; Thibault, L.; Shach, A.A.; Sebag, G.; Clark, D.P.; et al. Validation and real-world clinical application of an artificial intelligence algorithm for breast cancer detection in biopsies. NPJ Breast Cancer 2022, 8, 129. [Google Scholar] [CrossRef] [PubMed]
  57. Pantanowitz, L.; Quiroga-Garza, G.M.; Bien, L.; Heled, R.; Laifenfeld, D.; Linhart, C.; Sandbank, J.; Shach, A.A.; Shalev, V.; Vecsler, M.; et al. An artificial intelligence algorithm for prostate cancer diagnosis in whole slide images of core needle biopsies: A blinded clinical validation and deployment study. Lancet Digit. Health 2020, 2, e407–e416. [Google Scholar] [CrossRef] [PubMed]
  58. Sandbank, J.; Sebag, G.; Linhart, C. Validation and Clinical Deployment of an AI-Based Solution for Detection of Gastric Adenocarcinoma and Helicobacter Pylori in Gastric Biopsies. Lab. Investig. 2022, 102 (Suppl. S1), 419–565. [Google Scholar] [CrossRef]
  59. Tolkach, Y.; Ovtcharov, V.; Pryalukhin, A.; Eich, M.-L.; Gaisa, N.T.; Braun, M.; Radzhabov, A.; Quaas, A.; Hammerer, P.; Dellmann, A.; et al. An international multi-institutional validation study of the algorithm for prostate cancer detection and Gleason grading. NPJ Precis. Oncol. 2023, 7, 77. [Google Scholar] [CrossRef]
  60. Kleppe, A.; Skrede, O.-J.; De Raedt, S.; Hveem, T.S.; Askautrud, H.A.; Jacobsen, J.E.; Church, D.N.; Nesbakken, A.; Shepherd, N.A.; Novelli, M.; et al. A clinical decision support system optimising adjuvant chemotherapy for colorectal cancers by integrating deep learning and pathological staging markers: A development and validation study. Lancet Oncol. 2022, 23, 1221–1232. [Google Scholar] [CrossRef]
  61. Saillard, C.; Dubois, R.; Tchita, O.; Loiseau, N.; Garcia, T.; Adriansen, A.; Carpentier, S.; Reyre, J.; Enea, D.; von Loga, K.; et al. Validation of MSIntuit as an AI-based pre-screening tool for MSI detection from colorectal cancer histology slides. Nat. Commun. 2023, 14, 6695. [Google Scholar] [CrossRef]
  62. Raciti, P.; Sue, J.; Retamero, J.A.; Ceballos, R.; Godrich, R.; Kunz, J.D.; Casson, A.; Thiagarajan, D.; Ebrahimzadeh, Z.; Viret, J.; et al. Clinical Validation of Artificial Intelligence–Augmented Pathology Diagnosis Demonstrates Significant Gains in Diagnostic Accuracy in Prostate Cancer Detection. Arch. Pathol. Lab. Med. 2023, 147, 1178–1185. [Google Scholar] [CrossRef]
  63. Eloy, C.; Marques, A.; Pinto, J.; Pinheiro, J.; Campelos, S.; Curado, M.; Vale, J.; Polónia, A. Artificial intelligence–assisted cancer diagnosis improves the efficiency of pathologists in prostatic biopsies. Virchows Arch. 2023, 482, 595–604. [Google Scholar] [CrossRef] [PubMed]
  64. Shafi, S.; Kellough, D.A.; Lujan, G.; Satturwar, S.; Parwani, A.V.; Li, Z. Integrating and validating automated digital imaging analysis of estrogen receptor immunohistochemistry in a fully digital workflow for clinical use. J. Pathol. Inform. 2022, 13, 100122. [Google Scholar] [CrossRef] [PubMed]
  65. Anand, D.; Kurian, N.C.; Dhage, S.; Kumar, N.; Rane, S.; Gann, P.H.; Sethi, A. Deep Learning to Estimate Human Epidermal Growth Factor Receptor 2 Status from Hematoxylin and Eosin-Stained Breast Tissue Images. J. Pathol. Inform. 2020, 11, 19. [Google Scholar] [CrossRef]
  66. Hondelink, L.M.; Hüyük, M.; Postmus, P.E.; Smit, V.T.H.B.M.; Blom, S.; von der Thüsen, J.H.; Cohen, D. Development and validation of a supervised deep learning algorithm for automated whole-slide programmed death-ligand 1 tumour proportion score assessment in non-small cell lung cancer. Histopathology 2022, 80, 635–647. [Google Scholar] [CrossRef] [PubMed]
  67. Ryu, H.S.; Jin, M.-S.; Park, J.H.; Lee, S.; Cho, J.; Oh, S.; Kwak, T.-Y.; Woo, J.I.; Mun, Y.; Kim, S.W.; et al. Automated Gleason Scoring and Tumor Quantification in Prostate Core Needle Biopsy Images Using Deep Neural Networks and Its Comparison with Pathologist-Based Assessment. Cancers 2019, 11, 1860. [Google Scholar] [CrossRef] [PubMed]
  68. Ho, J.; Ahlers, S.M.; Stratman, C.; Aridor, O.; Pantanowitz, L.; Fine, J.L.; Kuzmishin, J.A.; Montalto, M.C.; Parwani, A.V. Can digital pathology result in cost savings? A financial projection for digital pathology implementation at a large integrated health care organization. J. Pathol. Inform. 2014, 5, 33. [Google Scholar] [CrossRef]
  69. Griffin, J.; Treanor, D. Digital pathology in clinical use: Where are we now and what is holding us back? Histopathology 2017, 70, 134–145. [Google Scholar] [CrossRef]
  70. Hanna, M.G.; Reuter, V.E.; Samboy, J.; England, C.; Corsale, L.; Fine, S.W.; Agaram, N.P.; Stamelos, E.; Yagi, Y.; Hameed, M.; et al. Implementation of Digital Pathology Offers Clinical and Operational Increase in Efficiency and Cost Savings. Arch. Pathol. Lab. Med. 2019, 143, 1545–1555. [Google Scholar] [CrossRef]
  71. Berbís, M.A.; McClintock, D.S.; Bychkov, A.; Van der Laak, J.; Pantanowitz, L.; Lennerz, J.K.; Cheng, J.Y.; Delahunt, B.; Egevad, L.; Eloy, C.; et al. Computational pathology in 2030: A Delphi study forecasting the role of AI in pathology within the next decade. EBioMedicine 2023, 88, 104427. [Google Scholar] [CrossRef]
  72. Kacew, A.J.; Strohbehn, G.W.; Saulsberry, L.; Laiteerapong, N.; Cipriani, N.A.; Kather, J.N.; Pearson, A.T. Artificial Intelligence Can Cut Costs While Maintaining Accuracy in Colorectal Cancer Genotyping. Front. Oncol. 2021, 11, 630953. [Google Scholar] [CrossRef]
  73. Fraggetta, F.; Caputo, A.; Guglielmino, R.; Pellegrino, M.G.; Runza, G.; L’Imperio, V. A Survival Guide for the Rapid Transition to a Fully Digital Workflow: The “Caltagirone Example”. Diagnostics 2021, 11, 1916. [Google Scholar] [CrossRef]
  74. Fraggetta, F.; Garozzo, S.; Zannoni, G.F.; Pantanowitz, L.; Rossi, E.D. Routine Digital Pathology Workflow: The Catania Experience. J. Pathol. Inform. 2017, 8, 51. [Google Scholar] [CrossRef] [PubMed]
  75. Schüffler, P.J.; Geneslaw, L.; Yarlagadda, D.V.K.; Hanna, M.G.; Samboy, J.; Stamelos, E.; Vanderbilt, C.; Philip, J.; Jean, M.-H.; Corsale, L.; et al. Integrated digital pathology at scale: A solution for clinical diagnostics and cancer research at a large academic medical center. J. Am. Med. Inform. Assoc. 2021, 28, 1874–1884. [Google Scholar] [CrossRef] [PubMed]
  76. Montezuma, D.; Monteiro, A.; Fraga, J.; Ribeiro, L.; Gonçalves, S.; Tavares, A.; Monteiro, J.; Macedo-Pinto, I. Digital Pathology Implementation in Private Practice: Specific Challenges and Opportunities. Diagnostics 2022, 12, 529. [Google Scholar] [CrossRef] [PubMed]
  77. Retamero, J.A.; Aneiros-Fernandez, J.; del Moral, R.G. Complete Digital Pathology for Routine Histopathology Diagnosis in a Multicenter Hospital Network. Arch. Pathol. Lab. Med. 2020, 144, 221–228. [Google Scholar] [CrossRef] [PubMed]
  78. Hartman, D.; Pantanowitz, L.; McHugh, J.; Piccoli, A.; Oleary, M.; Lauro, G. Enterprise Implementation of Digital Pathology: Feasibility, Challenges, and Opportunities. J. Digit. Imaging 2017, 30, 555–560. [Google Scholar] [CrossRef] [PubMed]
  79. Shah, S.C.; Kayamba, V.; Peek, M.R., Jr.; Heimburger, D. Cancer Control in Low- and Middle-Income Countries: Is It Time to Consider Screening? J. Glob. Oncol. 2019, 5, 1–8. [Google Scholar] [CrossRef] [PubMed]
  80. Sharma, R.; Nanda, M.; Fronterre, C.; Sewagudde, P.; Ssentongo, A.E.; Yenney, K.; Arhin, N.D.; Oh, J.; Amponsah-Manu, F.; Ssentongo, P. Mapping Cancer in Africa: A Comprehensive and Comparable Characterization of 34 Cancer Types Using Estimates From GLOBOCAN 2020. Front. Public Health 2022, 10, 839835. [Google Scholar] [CrossRef]
  81. Sayed, S.; Lukande, R.; Fleming, K.A. Providing Pathology Support in Low-Income Countries. J. Glob. Oncol. 2015, 1, 3–6. [Google Scholar] [CrossRef]
  82. Bychkov, A.; Schubert, M. Constant Demand, Patchy Supply. Pathologist 2023, 88, 18–27. [Google Scholar]
  83. Wamala, D.; Katamba, A.; Dworak, O. Feasibility and diagnostic accuracy of Internet-based dynamic telepathology between Uganda and Germany. J. Telemed. Telecare 2011, 17, 222–225. [Google Scholar] [CrossRef]
  84. Pagni, F.; Bono, F.; Di Bella, C.; Faravelli, A.; Cappellini, A. Virtual Surgical Pathology in Underdeveloped Countries: The Zambia Project. Arch. Pathol. Lab. Med. 2011, 135, 215–219. [Google Scholar] [CrossRef]
  85. Fischer, M.K.; Kayembe, M.K.; Scheer, A.J.; Introcaso, C.E.; Binder, S.W.; Kovarik, C.L. Establishing telepathology in Africa: Lessons from Botswana. J. Am. Acad. Dermatol. 2011, 64, 986–987. [Google Scholar] [CrossRef]
  86. Montgomery, N.D.; Tomoka, T.; Krysiak, R.; Powers, E.; Mulenga, M.; Kampani, C.; Chimzimu, F.; Owino, M.K.; Dhungel, B.M.; Gopal, S.; et al. Practical Successes in Telepathology Experiences in Africa. Clin. Lab. Med. 2018, 38, 141–150. [Google Scholar] [CrossRef]
  87. Hamnvåg, H.M.; McHenry, A.; Ahmed, A.; Trabzonlu, L.; Arnold, C.A.; Mirza, K.M. #TwitterHomework During Pathology Electives: Transforming Pathology Pedagogy. Arch. Pathol. Lab. Med. 2021, 145, 1438–1447. [Google Scholar] [CrossRef] [PubMed]
  88. Folaranmi, O.O.; Ibiyeye, K.M.; Odetunde, O.A.; Kerr, D.A. The Influence of Social Media in Promoting Knowledge Acquisition and Pathology Excellence in Nigeria. Front. Med. 2022, 9, 906950. [Google Scholar] [CrossRef]
  89. Crane, G.M.; Gardner, J.M. Pathology Image-Sharing on Social Media: Recommendations for Protecting Privacy While Motivating Education. AMA J. Ethics 2016, 18, 817–825. [Google Scholar] [CrossRef] [PubMed]
  90. Schaumberg, A.J.; Juarez-Nicanor, W.C.; Choudhury, S.J.; Pastrián, L.G.; Pritt, B.S.; Pozuelo, M.P.; Sánchez, R.S.; Ho, K.; Zahra, N.; Sener, B.D.; et al. Interpretable multimodal deep learning for real-time pan-tissue pan-disease pathology search on social media. Mod. Pathol. 2020, 33, 2169–2185. [Google Scholar] [CrossRef] [PubMed]
  91. Huang, Z.; Bianchi, F.; Yuksekgonul, M.; Montine, T.J.; Zou, J. A visual–language foundation model for pathology image analysis using medical Twitter. Nat. Med. 2023, 29, 2307–2316. [Google Scholar] [CrossRef]
Figure 1. Predicting molecular alterations from H&E-stained whole-slide images.
Figure 1. Predicting molecular alterations from H&E-stained whole-slide images.
Cancers 15 05692 g001
Figure 2. Technology readiness levels in pathology.
Figure 2. Technology readiness levels in pathology.
Cancers 15 05692 g002
Figure 3. Digital pathology workflow.
Figure 3. Digital pathology workflow.
Cancers 15 05692 g003
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Vigdorovits, A.; Köteles, M.M.; Olteanu, G.-E.; Pop, O. Breaking Barriers: AI’s Influence on Pathology and Oncology in Resource-Scarce Medical Systems. Cancers 2023, 15, 5692. https://doi.org/10.3390/cancers15235692

AMA Style

Vigdorovits A, Köteles MM, Olteanu G-E, Pop O. Breaking Barriers: AI’s Influence on Pathology and Oncology in Resource-Scarce Medical Systems. Cancers. 2023; 15(23):5692. https://doi.org/10.3390/cancers15235692

Chicago/Turabian Style

Vigdorovits, Alon, Maria Magdalena Köteles, Gheorghe-Emilian Olteanu, and Ovidiu Pop. 2023. "Breaking Barriers: AI’s Influence on Pathology and Oncology in Resource-Scarce Medical Systems" Cancers 15, no. 23: 5692. https://doi.org/10.3390/cancers15235692

APA Style

Vigdorovits, A., Köteles, M. M., Olteanu, G. -E., & Pop, O. (2023). Breaking Barriers: AI’s Influence on Pathology and Oncology in Resource-Scarce Medical Systems. Cancers, 15(23), 5692. https://doi.org/10.3390/cancers15235692

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop