Next Article in Journal
Internal Guidelines for Reducing Lymph Node Contour Variability in Total Marrow and Lymph Node Irradiation
Next Article in Special Issue
Machine Learning-Based Radiomic Features on Pre-Ablation MRI as Predictors of Pathologic Response in Patients with Hepatocellular Carcinoma Who Underwent Hepatic Transplant
Previous Article in Journal
Merkel Cell Carcinoma: An Update and Review
Previous Article in Special Issue
Longitudinal and Multimodal Radiomics Models for Head and Neck Cancer Outcome Prediction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Label Distribution Learning for Automatic Cancer Grading of Histopathological Images of Prostate Cancer

1
Department of Radiology, Kobe University Graduate School of Medicine, 7-5-2 Kusunoki-cho, Chuo-ku, Kobe 650-0017, Japan
2
Department of Diagnostic Imaging and Nuclear Medicine, Kyoto University Graduate School of Medicine, 54 Shogoin Kawahara-cho, Sakyo-ku, Kyoto 606-8507, Japan
3
Department of Informatics, Kindai University, 3-4-1 Kowakae, Higashiosaka City 577-8502, Japan
4
Department of Real World Data Research and Development, Kyoto University Graduate School of Medicine, 54 Shogoin Kawahara-cho, Sakyo-ku, Kyoto 606-8507, Japan
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(5), 1535; https://doi.org/10.3390/cancers15051535
Submission received: 13 February 2023 / Revised: 25 February 2023 / Accepted: 26 February 2023 / Published: 28 February 2023

Abstract

:

Simple Summary

We aimed to develop and evaluate an automatic prediction system for grading histopathological images of prostate cancer using a deep learning model and label distribution learning. Our results show that the label distribution learning improved the diagnostic performance of the automatic prediction system for the cancer grading.

Abstract

We aimed to develop and evaluate an automatic prediction system for grading histopathological images of prostate cancer. A total of 10,616 whole slide images (WSIs) of prostate tissue were used in this study. The WSIs from one institution (5160 WSIs) were used as the development set, while those from the other institution (5456 WSIs) were used as the unseen test set. Label distribution learning (LDL) was used to address a difference in label characteristics between the development and test sets. A combination of EfficientNet (a deep learning model) and LDL was utilized to develop an automatic prediction system. Quadratic weighted kappa (QWK) and accuracy in the test set were used as the evaluation metrics. The QWK and accuracy were compared between systems with and without LDL to evaluate the usefulness of LDL in system development. The QWK and accuracy were 0.364 and 0.407 in the systems with LDL and 0.240 and 0.247 in those without LDL, respectively. Thus, LDL improved the diagnostic performance of the automatic prediction system for the grading of histopathological images for cancer. By handling the difference in label characteristics using LDL, the diagnostic performance of the automatic prediction system could be improved for prostate cancer grading.

1. Introduction

In 2022, 34,500 Americans were predicted to die of prostate cancer [1]. Prostate cancer is the second leading cause of cancer-related death in men. It is also the leading cause of cancer-related morbidity in men, with 268,490 cases [1]. Although the 5-year survival rate of prostate cancer has improved in recent years, the number of deaths is considerably high due to the large number of patients with prostate cancer. Bone metastases from prostate cancer frequently occur and reduce the patients’ quality of life.
Prostate cancer is diagnosed by palpation, prostate-specific antigen testing, ultrasound examination, magnetic resonance imaging, and biopsy. The definitive diagnosis of prostate cancer is made by performing a pathological evaluation of the prostate tissue obtained through biopsy or surgery. A visual evaluation score called the Gleason score is used for pathological evaluation of prostate tissues [2,3]. It is determined based on the classification system of histological morphology with five-class patterns. For each specimen, the pattern with the largest area is regarded as the first pattern, while the pattern with the second largest area is regarded as the second pattern. The sum of the two patterns is then calculated to obtain the Gleason score. For example, if the values of the first and second patterns are 4 and 5, respectively, the Gleason score is 9 (4 + 5).
Various problems associated with the determination of the Gleason score have been identified. Hence, the International Society of Urological Pathology (ISUP) proposed a new grading system, called the ISUP grade group classification system (ISUP score), based on the Gleason score [3]. The appropriate ISUP score was assigned by grouping the Gleason scores, with higher ISUP scores indicating higher malignancy.
A previous study evaluated the reproducibility of the Gleason score and its interobserver variability [4,5,6]. One previous study showed moderate agreement on the Gleason score between the observers [4]. As the ISUP score is based on the Gleason score, the agreement on the ISUP score is expected to be moderate.
The recent emergence of deep learning provides powerful tools for medical image analysis [7]. Deep learning makes it possible to extract valuable features in an end-to-end manner. Many studies have used deep learning for medical image analysis in radiology [7,8,9,10] and pathology [11,12,13,14,15]. To improve the inter-observer variability of the Gleason and ISUP scores, deep learning has been used for automated grading systems [15,16,17,18,19,20,21,22,23]. Table 1 presents a summary of the results of previous studies on automatic prediction systems for Gleason and ISUP scores. Table 1 shows that the size of the dataset used in the deep learning-based systems of previous studies was less than 10,000. As shown in the previous study [10], it is difficult to construct reliable deep-learning-based systems with small-sized datasets.
This study aimed to (i) use a large dataset for the development of a deep-learning-based ISUP grading system, (ii) evaluate the generalizability of the automated ISUP grading, and (iii) incorporate label distribution learning (LDL) [24,25,26] in system development with deep learning to address a difference in label characteristics between datasets. Of these, the primary goal of the current study is to combine deep learning with LDL and evaluate the utility of LDL in the ISUP grading.
Table 1. Summary of previous studies on the automatic prediction systems of Gleason or ISUP scores.
Table 1. Summary of previous studies on the automatic prediction systems of Gleason or ISUP scores.
AuthorsOrigin of Dataset or Dataset NameSize of DatasetDiagnostic Performance of SystemsComment
Nagpal et al. [22]
  • TCGA Dataset
  • Naval Medical Center
  • San Diego Marin Medical Laboratories
  • 1226 slides for training
  • 331 slides for validation
  • 70% accuracy on the Gleason scoring task
  • DL
Arvaniti et al. [27]
  • University Hospital Zurich
  • 641 patients for training
  • 245 patients for testing
  • Cohen’s quadratic kappa was evaluated
  • the inter-pathologist agreement (kappa = 0.71).
  • kappa = 0.75 between DL and pathologist1 and kappa = 0.71 between DL and pathologist2.
  • DL
  • 6-class classification based on the Gleason score.
Lucas et al. [20]
  • Amsterdam University Medical Centers
  • 96 tissue sections from 38 patients
  • Concordance of adjusted grade groups between the automated determination method and a genitourinary pathologist was obtained in 65% (A quadratic weighted kappa = 0.70)
  • DL
  • 4-class classification based on the Gleason score.
Bulten et al. [19]
  • Radboud University Medical Center
  • 5759 biopsies from 1243 patients
  • In an observer experiment, the deep learning system scored higher (kappa 0.854) than the panel (median kappa 0.819), outperforming 10 of 15 pathologist observers.
  • DL
  • 6-class classification based on the Gleason score.
Egevad et al. [18]
  • Pathology Imagebase dataset hosted on the ISUP Web site
  • 87 needle biopsies
  • The mean weighted kappas of panel members for all cases, the consensus cases, and the non-consensus cases were 0.67, 0.77, and 0.50, respectively.
  • The weighted kappas of the AI system against the observers for all cases, the consensus cases, and the non-consensus cases were 0.63, 0.66, and 0.53, respectively.
  • DL developed in the previous study
  • All cases were graded by 23 panel members.
  • 5-class classification
  • The experts failed to reach a 2/3 consensus in 41.4% (36/87).
Kwak et al. [16]National Institutes of Health.
  • 73 benign and 89 cancer samples for training
  • 217 benign and 274 cancer samples for testing
  • AUC = 0.974
  • DL
  • Binary classification (benign/cancer)
  • Four tissue microarrays (TMAs) were used.
Singhal et al. [15]
  • PANDA challenge dataset (Radboud University Medical Center and Karolinska Institute)
  • Muljibhai Patel Urological Hospital (MPUH)
  • 580 biopsies from MPUH. The dataset was split into training (155) and testing sets (425).
  • 3586 biopsies from Radboud University Medical Center for training and 1201 for testing.
  • 1303 biopsies from the Karolinska Institute for unseen test data.
  • accuracy of 83.1% and a quadratic weighted kappa of 0.93 for the 1303 biopsies of unseen test data.
  • DL
  • Part of PANDA challenge dataset was used.
  • 6-class classification (5 classes of IUSP + benign)
Abbreviations: DL, deep learning.

2. Materials and Methods

2.1. Dataset

A total of 10,616 whole slide images (WSIs) were used in this study, which are available from the Prostate cANcer graDe Assessment (PANDA) challenge [14,28]. The 5160 and 5456 WSIs of the PANDA dataset were collected from Radboud University Medical Center and Karolinska Institute, respectively. A summary of the PANDA dataset is presented in Table 2. The details of the PANDA dataset are described in the published paper [14] and the Kaggle website [29].
The two institutions used different scanners with slightly different maximum microscopic resolutions. The annotation process for the 5160 and 5456 WSIs differed between the two institutions. At the Radboud University Medical Center, labels of the 5160 WSIs were retrieved from pathology reports, containing the original diagnosis. Trained students read all the reports and assigned a label to each WSI. At the Karolinska Institute, the 5456 WSIs were annotated by an experienced pathologist. Therefore, the labels of 5160 WSIs from Radboud University Medical Center might have a higher degree of inconsistency than those from the Karolinska Institute. In addition, Table 2 shows that the frequencies of ISUP scores were different between Radboud University Medical Center and Karolinska Institute; the frequencies of ISUP scores were more uniform at Radbound University Medical Center than at Karolinska Institute.
To develop and evaluate our deep learning-based systems, 5160 WSIs from Radboud University Medical Center and 5456 WSIs from Karolinska Institute were used as the development (training/validation sets) and unseen test sets, respectively. Table 2 shows that there was a difference in the frequency of ISUP scores between Radboud University Medical Center and Karolinska Institute. In addition, the 5160 WSIs of Radboud University Medical Center, which might have higher label inconsistency, were used for the development of a deep learning-based system. Therefore, it was expected that the dataset splitting of the current study would cause a deterioration in diagnostic performance compared with that of the original PANDA challenge.

2.2. Baseline Convolutional Neural Network

In the PANDA challenge, many deep learning-based systems have been developed using convolutional neural networks (CNNs). After the PANDA challenge, the source codes of several CNNs were made available as open sources. For example, the source code of the first-place solution used in the PANDA challenge can be obtained from the GitHub repository [30]. Their CNN ranked 22nd (metric = 0.910) and 1st (metric = 0.940) on the public and private leaderboards of the PANDA challenge, respectively. The details of the first-place solutions are described on their slide [31]. The first solution consists of several types of CNNs. One of the CNNs of the first-place solution was used as the baseline CNN in this study. To implement our baseline CNN (network structure, image preprocessing, loss function, etc.), the source code of the first-place solution was used. Based on the first-place solution, our baseline CNN used EfficientNet [32] (EfficientNet B1) for the network structure, which was pretrained with the ImageNet dataset. Because the original WSIs contained non-tissue lesions, the input images of the baseline CNN were preprocessed by creating tiled images as the first-place solution. In image tiling, the original WSIs were split into several tiles (image patches), tiles with non-tissue lesions were removed, and the tiles were concatenated as the input image. Figure 1 shows representative images of the original WSIs and their tiles. After image tiling, the original WSI was converted into a tiled image of 1536 × 1536 × 3 (1536 = 8 tile × 192) in size. Based on the first-place solution, the labels of the WSIs in our baseline CNN were represented as 10-dimensional vectors, where the first and second five dimensions were derived based on the ISUP score and the first pattern of the Gleason score, respectively. For example, the ISUP score and the first pattern of the Gleason score were 3 and 4, respectively, and the scores 3 and 4 were converted to [1, 1, 1, 0, 0] and [1, 1, 1, 1, 0], respectively. Then, the [1, 1, 1, 0, 0] and [1, 1, 1, 1, 0] were concatenated as [1, 1, 1, 0, 0, 1, 1, 1, 1, 0]. The [1, 1, 1, 0, 0, 1, 1, 1, 1, 0] was used as the label of the WSI for our baseline CNN. The loss function of our baseline CNN was regarded as the binary cross-entropy loss between the 10-dimensional vectors of the outputs and labels.

2.3. Proposed CNN with LDL

The primary goal of the current study is to combine deep learning with LDL [24,25,26]. For this purpose, LDL was incorporated into our proposed CNN. Previous studies suggested that LDL could be used to address label inconsistency issues in traditional single labels. Instead of assigning a single label, LDL covers a certain number of adjacent labels, where each label represents a different degree of description. Given the N input training WSIs with the corresponding single labels of ISUP scores and the first pattern of Gleason scores, the dataset is represented as { (x1, y1, z1), ……, (xN, yN, zN) }, where xi represents WSI, yi represents the ISUP score of WSI (yi ∈ [0, ……, 5]), and zi represents the first pattern of the Gleason score (zi ∈ [0, ……, 5]). For WSI xi, the label distribution d i c (c = 1, 2, 3, …, D, where D is the maximum number of label distributions) was generated based on the following Gaussian function:
d i c = 1 2 π σ e x p c l i 2 2 σ 2 ,
where li is the rescaled value of yi or zi and σ   is the standard deviation of the Gaussian function. In this study, the values of li were obtained from yi or zi using linear scaling based on the value of D (range of li matched with [1, …, D]). d i c represents the description degree (probability) of a label. As the maximum of the Gaussian function was obtained at c = li, the probability at li was the highest. However, the probabilities at c = l i 1   or c = l i + 1 were not 0 in general. Based on the label distribution, label inconsistency caused by label noise was handled in the LDL. The raw values of d i c were normalized to satisfy the following conditions: (i) d i c ∈ [0, 1] and (ii) c = 1 D d i c . Representative examples of label distributions are shown in Supplementary Material S1 (Figure S1A–D).
The network structure of our proposed CNN with LDL was based on EfficientNet. In our proposed CNN, the base part with convolutional layers of EfficientNet was retained, while the head part of EfficientNet was replaced with three fully connected layers with the following numbers of outputs: 512, 100, and 2D. The three layers accompanied the batch normalization and activation layers of the rectified linear unit. Our proposed CNN received tiled WSIs and generated a 2D-dimensional vector. The first and second D-dimensional vector ( o i 1 and o i 2 ) outputs from our proposed CNN corresponded to two vectors of d i c obtained from yi and zi. Training of our proposed CNN was accomplished by minimizing the KL divergence between the predicted and ground-truth label distributions, which is represented by the following equation:
KL div ( o i 1 ,   d i c _ f r o m _ y i )   + weight   ×   KL div ( o i 2 ,   d i c _ f r o m _ z i )
where d i c _ f r o m _ y i and d i c _ f r o m _ z i were obtained from yi and zi, respectively, with Equation (1) showing the label distributions. Here, KLdiv is the KL divergence between two probability distributions.

2.4. Implementation Details

PyTorch (version 1.9.0) and PyTorch Lightning (version 1.6.0) were used as the deep learning frameworks in this study. The baseline CNN was implemented based on the open-source code of the first-place solution in the PANDA challenge. As the original source code of the first-place solution used the old version of PyTorch Lightning (version 0.8.5), the source code for our baseline CNN was revised using the new version of PyTorch Lightning. The hyperparameters of our baseline CNN are available from https://github.com/kentaroy47/Kaggle-PANDA-1st-place-solution/blob/master/src/configs/final_1.yaml (accessed on 6 January 2023). Briefly, the hyperparameters of our baseline CNN were as follows: number of epochs, 30; optimizer, Adam; learning rate, 3.0 × 10−5; and size of input (tiled image), 1536 × 1536 × 3. The proposed CNN with LDL was implemented by modifying the source code of the baseline CNN. The base part of the proposed CNN was a pretrained EfficientNet (B0–B5) model. The hyperparameters specific to the proposed CNN were as follows: weight of LDL = 0.20 (Equation (2)), σ = 2.0 , and D = 18.

2.5. Evaluation of CNNs

In the baseline CNN, the predicted label was obtained by summing the first 5-dimensional vector of its output. This method was implemented in the original source code of the first solution. In the proposed CNN with LDL, the raw predicted label of the proposed CNN was determined using the following equation:
p l i = a r g m a x o i 1 ,
where p l i   is the predicted raw label of the proposed CNN. Finally, the value of p l i   was inverted via linear rescaling based on the value of D, and the final predicted label for the proposed CNN was obtained. In the baseline CNN and proposed CNN, the predicted Gleason score was not used for evaluation. Five-fold cross-validation was performed for the baseline CNN and proposed CNN using the development set (5160 WSIs from Radboud University Medical Center). After the five-fold cross-validation, the unseen test set (5456 WSIs from the Karolinska Institute) was used to evaluate the CNNs. The prediction results for the test set were obtained using an ensemble of five trained models for the baseline and proposed CNNs. As the whole process of prediction is not available in the source code of the first-place solution, it was implemented for the baseline and proposed CNNs in this study.
The evaluation metrics used in this study were quadratic weighted kappa (QWK) and accuracy. QWK, the main metric utilized in the PANDA challenge, was used to evaluate the agreement between the ground truth label and predicted label. The QWK typically varies from 0 (random agreement) to 1 (complete agreement). QWK was calculated as follows: First, a confusion matrix M was constructed from the ground truth and predicted labels. The size of M was N-by-N (N = 6, range of ground truth label; 0–5, range of predicted labels). In matrix M, Mi,j corresponded to the number of ISUP scores i (ground truth) that received a predicted value j. The N-by-N matrix of weights w was calculated based on the difference between the actual and predicted values using the following equation:
w i , j = i j 2 N 1 2 .
The N-by-N confusion matrix of the expected outcomes, E, was calculated assuming that there was no correlation between the ground truth and predicted labels. In all three matrices (M, w, and E), QWK was calculated as follows:
QWK = i , j w i , j M i , j i , j w i , j E i , j .
Accuracy was defined as the ratio of diagonal elements in matrix M calculated between the ground truth and predicted labels. In this study, the IUSP scores were used as the ground truth in the calculation of QWK and accuracy; the Gleason scores were ignored when calculating the evaluation metrics of this study.
To statistically test the difference in accuracy between the baseline and proposed CNNs, McNemar’s test was used. A p-value < 0.05 was considered to indicate statistical significance.

3. Results

Table 3 presents the results of five-fold cross-validation of the baseline and proposed CNNs (EfficientNet B0–5) using the development set. The cross-validated QWK and the accuracy of the baseline CNN were 0.820 and 0.545, respectively. The cross-validated QWK and accuracy of the proposed CNNs were 0.817–850 and 0.646–0.680, respectively. In most cases, the cross-validated QWK and accuracy of the proposed CNNs were better than those of the baseline CNN. The proposed CNN of EfficientNet B3 was the best, as shown in Table 3.
Table 4 shows the results of five-fold cross-validation of the proposed CNNs with different D values (EfficientNet B3 only). Here, D = 12, 30, and 60 were used in addition to D = 18. The cross-validated QWK and accuracy of the proposed CNNs were 0.835–850 and 0.609–0.700, respectively (Table 4). The proposed CNN with D = 60 achieved the highest accuracy (Table 4).
The cross-validated QWK and accuracy of the proposed CNN of EfficientNet B3 with LDL (D = 18) and another proposed CNN of EfficientNet B3 with LDL (D = 60) were relatively high (Table 3 and Table 4, respectively). Hence, these two types of proposed CNNs were evaluated using the test set.
Table 5 shows the diagnostic performance of the baseline and proposed CNNs in the unseen test set. The QWK and accuracy of the baseline CNN were 0.240 and 0.247, respectively, in the test set. The QWK and accuracy of the two types of proposed CNNs were 0.301 and 0.364, and 0.249 and 0.407, respectively, in the test set. The difference in accuracy between the baseline CNN and the proposed CNN (EfficientNet B3 with LDL (D = 60)) was statistically significant (p-value < 0.000001). As a result, the proposed CNNs can improve the diagnostic performance of the first-place solution for the automatic prediction of prostate cancer grade. Figure 2 shows the confusion matrix between the ground truth and the predicted label for the proposed CNN of EfficientNet B3 with LDL (D = 60).
Table 3, Table 4 and Table 5 show that when the development set (5160 WSIs from Radboud University Medical Center) was used, the diagnostic performance of the baseline and proposed CNNs severely deteriorated in the test set (5456 WSIs from the Karolinska Institute).

4. Discussion

Table 5 shows that the proposed CNNs could achieve better diagnostic performance for the automatic prediction of ISUP scores compared with the baseline CNN. Because the major difference between the proposed CNNs and the baseline CNNs was the use of LDL, LDL was useful for improving the CNNs and predicting the ISUP scores automatically. However, for the proposed CNN and baseline CNN, their diagnostic performance in the test set was worse than that in the development set.
As previously mentioned in Materials and Methods section and Table 2, the labels of 5160 WSIs from the Radboud University Medical Center were annotated by trained students. Therefore, these labels might have a higher degree of inconsistency compared with those from the Karolinska Institute. The label inconsistency of WSIs in the Radboud University Medical Center might decrease the generalizability of the model in the development of the proposed and baseline CNNs. In addition, the difference in label frequency between Radboud University Medical Center and Karolinska Institute might deteriorate the model’s generalizability. Our results and speculation indicate that when a large number of images with low-quality labels are available, the development of a deep learning model may lead to the deterioration of the generalizability of the model. This point should be carefully considered when developing deep-learning-based systems.
Issues related to label inconsistency have been reported in the original PANDA challenge. However, the discrepancy in diagnostic performance between the development and test sets was not extremely large during the original PANDA challenge because the development set consisted of WSIs obtained from the two institutions. In this study, the discrepancy in diagnostic performance between the development and test sets was significantly large because the development set was derived from one institution.
LDL has been proposed to address the issues related to label inconsistency on traditional single labels. In this study, the diagnostic performance of our system could be improved by incorporating LDL into the deep learning-based system. Owing to the characteristics of LDL, it might be useful for other medical systems to resolve label inconsistency issues in the grading scores (cancer staging, grading for cancer diagnosis, and so on).
Table 1 presents a summary of the results of previous studies on automatic prediction systems for Gleason and ISUP scores. The size of the dataset in this study was larger than that in the previous studies. However, the diagnostic performance of our CNNs was worse than that of the CNNs in the previous studies. As shown, the differences in label inconsistency and label frequency between Radbound University Medical Center and Karolinska Institute may cause the performance discrepancy.
This study has several limitations. First, it was conducted using a public dataset (the PANDA dataset). Our results should be confirmed using other datasets. Second, our results were obtained using EfficientNet. Although EfficientNet is a standard deep learning model, other deep learning models should be used to evaluate the usefulness of LDL. As the first-place solution to the PANDA challenge used EfficientNet in the source code, EfficientNet was also used in this study.

5. Conclusions

Our proposed CNN with LDL could improve the cancer grading (ISUP scores) of histopathological images in prostate cancer. This improvement could be achieved with the aid of LDL (a strategy used to address label inconsistency issues). However, when the difference in label characteristics (label inconsistency and label frequency) between the development and test sets was observed, the generalizability of the baseline and proposed CNNs deteriorated in the unseen test set. Our results should be carefully considered when developing deep-learning-based systems.

Supplementary Materials

The following supporting information can be downloaded at: https://www.mdpi.com/article/10.3390/cancers15051535/s1, Material S1 (including Figure S1A–D), Example of label distribution; Material S2 (including Table S2): Results of the five-fold cross validation (CV) for the baseline CNN and the proposed CNN.

Author Contributions

Conceptualization, M.N., Y.K., O.S. and K.F.; methodology, M.N.; software, M.N. and H.M.; validation, M.N.; formal analysis, M.N.; investigation, M.N.; resources, M.N.; data curation, M.N.; writing—original draft preparation, M.N.; writing—review and editing, all authors; visualization, M.N.; supervision, M.N.; project administration, M.N.; funding acquisition, M.N.; All authors have read and agreed to the published version of the manuscript.

Funding

The present study was partly supported by JSPS KAKENHI (grant numbers: JP19K17232 and 22K07665).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The public dataset is available on the Kaggle website. The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Siegel, R.L.; Miller, K.D.; Fuchs, H.E.; Jemal, A. Cancer statistics, 2022. CA Cancer J. Clin. 2022, 72, 7–33. [Google Scholar] [CrossRef] [PubMed]
  2. Gleason, D.F. Histologic grading of prostate cancer: A perspective. Hum. Pathol. 1992, 23, 273–279. [Google Scholar] [CrossRef] [PubMed]
  3. Epstein, J.I.; Egevad, L.; Amin, M.B.; Delahunt, B.; Srigley, J.R.; Humphrey, P.A. The 2014 international society of urological pathology (ISUP) consensus conference on gleason grading of prostatic carcinoma definition of grading patterns and proposal for a new grading system. Am. J. Surg. Pathol. 2016, 40, 244–252. [Google Scholar] [CrossRef]
  4. Ozkan, T.A.; Eruyar, A.T.; Cebeci, O.O.; Memik, O.; Ozcan, L.; Kuskonmaz, I. Interobserver variability in Gleason histological grading of prostate cancer. Scand. J. Urol. 2016, 50, 420–424. [Google Scholar] [CrossRef]
  5. Allsbrook, W.C.; Mangold, K.A.; Johnson, M.H.; Lane, R.B.; Lane, C.G.; Epstein, J.I. Interobserver reproducibility of Gleason grading of prostatic carcinoma: General pathologist. Hum. Pathol. 2001, 32, 81–88. [Google Scholar] [CrossRef]
  6. Di Loreto, C.; Fitzpatrick, B.; Underhill, S.; Kim, D.H.; Dytch, H.E.; Galera-Davidson, H.; Bibbo, M. Correlation Between Visual Clues, Objective Architectural Features, and Interobserver Agreement in Prostate Cancer. Am. J. Clin. Pathol. 1991, 96, 70–75. [Google Scholar] [CrossRef]
  7. Yamashita, R.; Nishio, M.; Do, R.K.G.; Togashi, K. Convolutional neural networks: An overview and application in radiology. Insights Imaging 2018, 9, 611–629. [Google Scholar] [CrossRef]
  8. Moribata, Y.; Kurata, Y.; Nishio, M.; Kido, A.; Otani, S.; Himoto, Y.; Nishio, N.; Furuta, A.; Onishi, H.; Masui, K.; et al. Automatic segmentation of bladder cancer on MRI using a convolutional neural network and reproducibility of radiomics features: A two-center study. Sci. Rep. 2023, 13, 628. [Google Scholar] [CrossRef]
  9. Noguchi, S.; Nishio, M.; Sakamoto, R.; Yakami, M.; Fujimoto, K.; Emoto, Y.; Kubo, T.; Iizuka, Y.; Nakagomi, K.; Miyasa, K.; et al. Deep learning-based algorithm improved radiologists’ performance in bone metastases detection on CT. Eur. Radiol. 2022, 32, 7976–7987. [Google Scholar] [CrossRef]
  10. Matsuo, H.; Nishio, M.; Kanda, T.; Kojita, Y.; Kono, A.K.; Hori, M.; Teshima, M.; Otsuki, N.; Nibu, K.-i; Murakami, T. Diagnostic accuracy of deep-learning with anomaly detection for a small amount of imbalanced data: Discriminating malignant parotid tumors in MRI. Sci. Rep. 2020, 10, 19388. [Google Scholar] [CrossRef]
  11. Steiner, D.F.; Macdonald, R.; Liu, Y.; Truszkowski, P.; Hipp, J.D.; Gammage, C.; Thng, F.; Peng, L.; Stumpe, M.C. Impact of Deep Learning Assistance on the Histopathologic Review of Lymph Nodes for Metastatic Breast Cancer. Am. J. Surg. Pathol. 2018, 42, 1636–1646. [Google Scholar] [CrossRef] [PubMed]
  12. Woerl, A.C.; Eckstein, M.; Geiger, J.; Wagner, D.C.; Daher, T.; Stenzel, P.; Fernandez, A.; Hartmann, A.; Wand, M.; Roth, W.; et al. Deep Learning Predicts Molecular Subtype of Muscle-invasive Bladder Cancer from Conventional Histopathological Slides. Eur. Urol. 2020, 78, 256–264. [Google Scholar] [CrossRef] [PubMed]
  13. Wei, J.W.; Tafe, L.J.; Linnik, Y.A.; Vaickus, L.J.; Tomita, N.; Hassanpour, S. Pathologist-level classification of histologic patterns on resected lung adenocarcinoma slides with deep neural networks. Sci. Rep. 2019, 9, 3358. [Google Scholar] [CrossRef]
  14. Bulten, W.; Kartasalo, K.; Chen, P.H.C.; Ström, P.; Pinckaers, H.; Nagpal, K.; Cai, Y.; Steiner, D.F.; van Boven, H.; Vink, R.; et al. Artificial intelligence for diagnosis and Gleason grading of prostate cancer: The PANDA challenge. Nat. Med. 2022, 28, 154–163. [Google Scholar] [CrossRef]
  15. Singhal, N.; Soni, S.; Bonthu, S.; Chattopadhyay, N.; Samanta, P.; Joshi, U.; Jojera, A.; Chharchhodawala, T.; Agarwal, A.; Desai, M.; et al. A deep learning system for prostate cancer diagnosis and grading in whole slide images of core needle biopsies. Sci. Rep. 2022, 12, 3383. [Google Scholar] [CrossRef]
  16. Kwak, J.T.; Hewitt, S.M. Nuclear Architecture Analysis of Prostate Cancer via Convolutional Neural Networks. IEEE Access 2017, 5, 18526–18533. [Google Scholar] [CrossRef]
  17. Ren, J.; Sadimin, E.; Foran, D.J.; Qi, X. Computer aided analysis of prostate histopathology images to support a refined Gleason grading system. In Proceedings of the Medical Imaging 2017, Image Processing, SPIE, Orlando, FL, USA, 24 February 2017; p. 101331V. [Google Scholar] [CrossRef]
  18. Egevad, L.; Swanberg, D.; Delahunt, B.; Ström, P.; Kartasalo, K.; Olsson, H.; Berney, D.M.; Bostwick, D.G.; Evans, A.J.; Humphrey, P.A.; et al. Identification of areas of grading difficulties in prostate cancer and comparison with artificial intelligence assisted grading. Virchows Arch. 2020, 477, 777–786. [Google Scholar] [CrossRef]
  19. Bulten, W.; Pinckaers, H.; van Boven, H.; Vink, R.; de Bel, T.; van Ginneken, B.; van der Laak, J.; Hulsbergen-van de Kaa, C.; Litjens, G. Automated deep-learning system for Gleason grading of prostate cancer using biopsies: A diagnostic study. Lancet Oncol. 2020, 21, 233–241. [Google Scholar] [CrossRef]
  20. Lucas, M.; Jansen, I.; Savci-Heijink, C.D.; Meijer, S.L.; de Boer, O.J.; van Leeuwen, T.G.; de Bruin, D.M.; Marquering, H.A. Deep learning for automatic Gleason pattern classification for grade group determination of prostate biopsies. Virchows Arch. 2019, 475, 77–83. [Google Scholar] [CrossRef]
  21. Jiménez del Toro, O.; Atzori, M.; Otálora, S.; Andersson, M.; Eurén, K.; Hedlund, M.; Rönnquist, P.; Müller, H. Convolutional neural networks for an automatic classification of prostate tissue slides with high-grade Gleason score. In Proceedings of the Medical Imaging 2017, Digital Pathology, SPIE, Orlando, FL, USA, 1 March 2017; p. 101400O. [Google Scholar] [CrossRef]
  22. Nagpal, K.; Foote, D.; Liu, Y.; Chen, P.H.C.; Wulczyn, E.; Tan, F.; Olson, N.; Smith, J.L.; Mohtashamian, A.; Wren, J.H.; et al. Development and validation of a deep learning algorithm for improving Gleason scoring of prostate cancer. NPJ Digit. Med. 2019, 2, 48. [Google Scholar] [CrossRef]
  23. Linkon, A.H.M.; Labib, M.M.; Hasan, T.; Hossain, M.; Jannat, M.E. Deep Learning in Prostate Cancer Diagnosis and Gleason Grading in Histopathology Images: An Extensive Study. Informatics in Medicine Unlocked; Elsevier: Amsterdam, The Netherlands, 2021; p. 100582. [Google Scholar] [CrossRef]
  24. Geng, X.; Yin, C.; Zhou, Z.H. Facial age estimation by learning from label distributions. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 2401–2412. [Google Scholar] [CrossRef] [PubMed]
  25. Luo, J.; He, B.; Ou, Y.; Li, B.; Wang, K. Topic-based label distribution learning to exploit label ambiguity for scene classification. Neural Comput. Appl. 2021, 33, 16181–16196. [Google Scholar] [CrossRef]
  26. Wu, X.; Wen, N.; Liang, J.; Lai, Y.K.; She, D.; Cheng, M.M.; Yang, J. Joint acne image grading and counting via label distribution learning. In Proceedings of the IEEE International Conference on Computer Vision, Seoul, Republic of Korea, 27 October 2019–2 November 2019; Institute of Electrical and Electronics Engineers Inc.: Piscataway, NJ, USA, 2019; pp. 10641–10650. [Google Scholar] [CrossRef]
  27. Arvaniti, E.; Fricker, K.S.; Moret, M.; Rupp, N.; Hermanns, T.; Fankhauser, C.; Wey, N.; Wild, P.J.; Rüschoff, J.H.; Claassen, M. Automated Gleason grading of prostate cancer tissue microarrays via deep learning. Sci. Rep. 2018, 8, 12054. [Google Scholar] [CrossRef] [PubMed]
  28. Bulten, W.; Litjens, G.; Pinckaers, H.; Ström, P.; Eklund, M.; Kartasalo, K.; Demkin, M.; Dane, S. The PANDA challenge: Prostate cANcer graDe Assessment using the Gleason grading system. In Proceedings of the 23rd International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI 2020), Lima, Peru, 19 March 2020. [Google Scholar] [CrossRef]
  29. Prostate cANcer graDe Assessment (PANDA) Challenge | Kaggle. Available online: https://www.kaggle.com/c/prostate-cancer-grade-assessment (accessed on 6 January 2023).
  30. GitHub—Kentaroy47/Kaggle-PANDA-1st-Place-Solution: 1st Place Solution for the Kaggle PANDA Challenge. Available online: https://github.com/kentaroy47/Kaggle-PANDA-1st-place-solution (accessed on 6 January 2023).
  31. RistKaggleWorkshop_20200924_PANDA_1st—Google Slide. Available online: https://docs.google.com/presentation/d/1Ies4vnyVtW5U3XNDr_fom43ZJDIodu1SV6DSK8di6fs/edit#slide=id.p (accessed on 6 January 2023).
  32. Tan, M.; Le, Q.V. EfficientNet: Rethinking model scaling for convolutional neural networks. In Proceedings of the 36th International Conference of Machine Learning PMLR 2019, Long Beach, CA, USA, 9–15 June 2019; Volume 97, pp. 10691–10700. [Google Scholar]
Figure 1. Original WSIs and their tiled images with different ISUP scores ((A) ISUP score = 0, (B) ISUP score = 4, and (C) ISUP score = 5). Note: For brevity, only five tiled images are shown. (A) ISUP score = 0 and Gleason score = 0 + 0 at the Karolinska Institute; (B) ISUP score = 4 and Gleason score = 4 + 4 at Radboud University Medical Center; (C) ISUP score = 5 and Gleason score = 4 + 5 at Radboud University Medical Center. Abbreviations: whole slide image (WSI).
Figure 1. Original WSIs and their tiled images with different ISUP scores ((A) ISUP score = 0, (B) ISUP score = 4, and (C) ISUP score = 5). Note: For brevity, only five tiled images are shown. (A) ISUP score = 0 and Gleason score = 0 + 0 at the Karolinska Institute; (B) ISUP score = 4 and Gleason score = 4 + 4 at Radboud University Medical Center; (C) ISUP score = 5 and Gleason score = 4 + 5 at Radboud University Medical Center. Abbreviations: whole slide image (WSI).
Cancers 15 01535 g001
Figure 2. Confusion matrix of the proposed CNN between the ground truth and predicted labels in the test set. Note: EfficientNet B3 with LDL (D = 60) was used.
Figure 2. Confusion matrix of the proposed CNN between the ground truth and predicted labels in the test set. Note: EfficientNet B3 with LDL (D = 60) was used.
Cancers 15 01535 g002
Table 2. Summary of the PANDA dataset.
Table 2. Summary of the PANDA dataset.
KERRYPNXRadboud University Medical CenterKarolinska Institute
Number of WSIsN = 5160N = 5456
Frequency of ISUP scoresISUP score 0, N = 967ISUP score 0, N = 1925
ISUP score 1, N = 852ISUP score 1, N = 1814
ISUP score 2, N = 675ISUP score 2, N = 668
ISUP score 3, N = 925ISUP score 3, N = 317
ISUP score 4, N = 768ISUP score 4, N = 481
ISUP score 5, N = 973ISUP score 5, N = 251
Annotatorstrained studentsa single experienced pathologist
Usage in this studydevelopment set (training/validation sets)unseen test set
Abbreviations: Prostate cANcer graDe Assessment (PANDA), whole slide image (WSI), and the International Society of Urological Pathology (ISUP).
Table 3. Results of the five-fold cross-validation of the baseline and proposed CNNs (EfficientNet B0–5).
Table 3. Results of the five-fold cross-validation of the baseline and proposed CNNs (EfficientNet B0–5).
CNNCross-Validated QWKCross-Validated Accuracy
Baseline CNN0.8200.545
Proposed CNN of EfficientNet B0 with LDL0.8170.646
Proposed CNN of EfficientNet B1 with LDL0.8360.663
Proposed CNN of EfficientNet B2 with LDL0.8400.667
Proposed CNN of EfficientNet B3 with LDL0.8500.680
Proposed CNN of EfficientNet B4 with LDL0.8400.663
Proposed CNN of EfficientNet B5 with LDL0.8320.654
Note: The development set (5160 WSIs from Radboud University Medical Center) was used for the five-fold cross validation. In the proposed CNN, the following parameters were used: the weight for LDL = 0.20, 𝜎 = 2.0, and D = 18. Abbreviations: WSI, whole slide image; CNN, convolutional neural network; QWK, quadratic weighted kappa; LDL, label distribution learning.
Table 4. Results of the five-fold cross validation of the proposed CNNs with different D values (EfficientNet B3 only).
Table 4. Results of the five-fold cross validation of the proposed CNNs with different D values (EfficientNet B3 only).
CNNCross-Validated QWKCross-Validated Accuracy
Proposed CNN with LDL (D = 18)0.8500.680
Proposed CNN with LDL (D = 12)0.8420.609
Proposed CNN with LDL (D = 30)0.8350.683
Proposed CNN with LDL (D = 60)0.8390.700
Note: The development set (5160 WSIs from Radboud University Medical Center) was used for the five-fold cross validation. Abbreviations: WSI, whole-slide image; CNN, convolutional neural network; QWK, quadratic weighted kappa.
Table 5. Performance of the baseline and proposed CNNs in the unseen test set.
Table 5. Performance of the baseline and proposed CNNs in the unseen test set.
CNNQWKAccuracy
Baseline CNN0.2400.247
Proposed CNN of EfficientNet B3 with LDL (D = 18)0.3010.249
Proposed CNN of EfficientNet B3 with LDL (D = 60)0.3640.407
Note: The two types of proposed CNNs were used in the evaluation of the unseen test set as their cross-validated QWK and accuracy were relatively high. The test set (5456 WSIs from the Karolinska Institute) is used in Table 5. Abbreviations: WSI, whole slide image; CNN, convolutional neural network; LDL, label distribution learning; QWK, quadratic weighted kappa.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Nishio, M.; Matsuo, H.; Kurata, Y.; Sugiyama, O.; Fujimoto, K. Label Distribution Learning for Automatic Cancer Grading of Histopathological Images of Prostate Cancer. Cancers 2023, 15, 1535. https://doi.org/10.3390/cancers15051535

AMA Style

Nishio M, Matsuo H, Kurata Y, Sugiyama O, Fujimoto K. Label Distribution Learning for Automatic Cancer Grading of Histopathological Images of Prostate Cancer. Cancers. 2023; 15(5):1535. https://doi.org/10.3390/cancers15051535

Chicago/Turabian Style

Nishio, Mizuho, Hidetoshi Matsuo, Yasuhisa Kurata, Osamu Sugiyama, and Koji Fujimoto. 2023. "Label Distribution Learning for Automatic Cancer Grading of Histopathological Images of Prostate Cancer" Cancers 15, no. 5: 1535. https://doi.org/10.3390/cancers15051535

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop