Next Article in Journal
Hybrid Quantum Neural Network for Drug Response Prediction
Next Article in Special Issue
Prostate Cancer and Its Mimics—A Pictorial Review
Previous Article in Journal
COL7A1 Expression Improves Prognosis Prediction for Patients with Clear Cell Renal Cell Carcinoma Atop of Stage
Previous Article in Special Issue
Diagnostic Accuracy of Magnetic Resonance Imaging in the Pre-Operative Staging of Cervical Cancer Patients Who Underwent Neoadjuvant Treatment: A Clinical–Surgical–Pathologic Comparison
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Attention-Based Deep Learning System for Classification of Breast Lesions—Multimodal, Weakly Supervised Approach

1
2nd Department of Radiology, Medical University of Gdansk, 80-214 Gdansk, Poland
2
Department of Intelligent Control Systems and Decision Support, Faculty of Electrical and Control Engineering, Gdansk University of Technology, 80-233 Gdansk, Poland
*
Authors to whom correspondence should be addressed.
Cancers 2023, 15(10), 2704; https://doi.org/10.3390/cancers15102704
Submission received: 24 March 2023 / Revised: 2 May 2023 / Accepted: 5 May 2023 / Published: 10 May 2023
(This article belongs to the Special Issue Application of Advanced Biomedical Imaging in Cancer Treatment)

Abstract

:

Simple Summary

Breast cancer affects millions of women worldwide. We aim to provide radiologists with automatic support for mammography review. We propose deploying deep learning models with Multiple Instance Learning algorithms for breast cancer diagnosis (cancer versus non-cancer classification) based on digital mammography images, taking advantage of data annotated at an image level without indicating where the lesion is. We employed algorithms to analyse original, high-resolution images with minimal reduction in size. We used graphical suggestions of attentional maps to verify the correctness of the algorithm and to indicate the areas of the images where cancer lesions are. Finally, we performed comparative and validation studies on external datasets differing in the number of images, pixel intensity levels, and subtypes of existing lesions, showing high accuracy and potential for generalisability of the algorithms.

Abstract

Breast cancer is the most frequent female cancer, with a considerable disease burden and high mortality. Early diagnosis with screening mammography might be facilitated by automated systems supported by deep learning artificial intelligence. We propose a model based on a weakly supervised Clustering-constrained Attention Multiple Instance Learning (CLAM) classifier able to train under data scarcity effectively. We used a private dataset with 1174 non-cancer and 794 cancer images labelled at the image level with pathological ground truth confirmation. We used feature extractors (ResNet-18, ResNet-34, ResNet-50 and EfficientNet-B0) pre-trained on ImageNet. The best results were achieved with multimodal-view classification using both CC and MLO images simultaneously, resized by half, with a patch size of 224 px and an overlap of 0.25. It resulted in AUC-ROC = 0.896 ± 0.017, F1-score 81.8 ± 3.2, accuracy 81.6 ± 3.2, precision 82.4 ± 3.3, and recall 81.6 ± 3.2. Evaluation with the Chinese Mammography Database, with 5-fold cross-validation, patient-wise breakdowns, and transfer learning, resulted in AUC-ROC 0.848 ± 0.015, F1-score 78.6 ± 2.0, accuracy 78.4 ± 1.9, precision 78.8 ± 2.0, and recall 78.4 ± 1.9. The CLAM algorithm’s attentional maps indicate the features most relevant to the algorithm in the images. Our approach was more effective than in many other studies, allowing for some explainability and identifying erroneous predictions based on the wrong premises.

1. Introduction

Breast cancer, with more than 2.2 million new cases in 2020, is the most common cancer in women worldwide [1]. It is responsible for around 680,000 cancer-related deaths, despite the rapid improvement in diagnostics and treatment methods [1]. Currently, the diagnosis is based on clinical examination, mammography, ultrasonography of the breast and loco-regional lymph nodes, and in some cases, the magnetic resonance of the breast. These examinations are followed by invasive breast tumour and lymph node biopsy procedures and pathological examination of specimens to verify the clinical diagnosis. There is no faster, less invasive, and cheaper diagnostic approach so far, which could reduce the number of necessary radiological, surgical, and pathology tests to confirm the presence of cancer and its potential metastases [2]. The most crucial task is to detect any suspicious lesion in the first examination, such as screening mammography, and then decide whether it might be malignant [3,4].
Mammography screening, widely used for the early detection of breast cancer, generates millions of film or digital mammograms requiring vast numbers of very well-trained radiologists to assess them. For many years, several attempts were made to create computer-aided diagnosis (CAD) support systems reading mammograms to facilitate more automated detection of breast lesions and their classification [5,6,7]. These systems were based on machine learning algorithms, mostly on deep learning. Our recent search with IEEE Xplore with terms ‘breast cancer’, AND ‘mammography’ AND ‘machine learning’, OR ‘deep learning’, OR ‘artificial intelligence’ OR ‘computer aided diagnosis’, OR ‘algorithms’ returned over 1500 records from years 2016 to 2022 with over 250 papers and conference proceedings addressing artificial intelligence in breast cancer diagnosis and classification based on film or digital mammography.
Existing deep learning algorithms can reach human-level or almost human-level performance in many tasks [8,9,10,11,12]. Despite the great success of AI-based computer-aided diagnosis systems, there are many problems related to using these algorithms. One of the significant difficulties is the insufficient number of adequately segmented and labelled data to train such large-scale models effectively. Another difficulty is selecting an appropriate structure for neural models and accompanying mechanisms supporting data analysis. Finally, and most importantly, clinicians have to deal with the complexity and frequently unclear interpretations of decisions taken by those data-driven, black-box models. The latter is essential considering various types of biases and artefacts in the datasets [13]. Until these issues are addressed, the broad adoption of deep learning-based decision support systems in clinical practice is limited.
In this paper, we would like to familiarise the readers with our research towards AI-based computer-aided diagnosis system for the early detection and diagnosis of breast cancer. To ensure the trustworthiness of the system’s performance, we would like the system to indicate the fundamental premises for a given decision. The assumption is that the data are annotated only at the image level, so it is impossible to take advantage of systems for segmentation. Furthermore, the data available for analysis are relatively limited. For this purpose, we designed a weakly supervised Multiple Instance Learning deep learning system exploring the information hidden in the mediolateral oblique (MLO) and craniocaudal (CC) full-field digital mammography images. Such an approach makes it possible to train the system under data scarcity effectively. It requires limited involvement of highly skilled breast radiologists in the annotation process but involves them in the explainability and information feedback loops. The designed system has been comprehensively validated for using different types and structures of analysed data, the impact of hyperparameters, and the detection and impact of artefacts. Our research used the MUG dataset from the Medical University of Gdansk, Poland, and the Chinese Mammography Database (CMMD). All the analyses were conducted on the MUG dataset, and finally, the optimised structure of the CAD system was validated on the CMMD dataset.
The contributions of the paper include the following:
  • A proposition for the deployment of deep learning models with Multiple Instance Learning algorithms for breast cancer diagnosis based on full-field digital mammography MLO and CC images, taking advantage of data annotated just at an image level;
  • A comparative analysis of breast cancer diagnosis performance based on just CC images, just MLO images, the analysis of combined CC and MLO images, and multimodal analysis of CC and MLO images;
  • An analysis of the potential of employing attentional maps to verify the correctness of the algorithm and to indicate the areas of the images where cancer foci are located (detection);
  • Comprehensive ablation studies showing the impact of the system’s hyperparameters on the effectiveness of the system’s performance, quality, and usability for the end users of the accompanying attentional maps. Most common neural models were also investigated;
  • Analysis and discussion of the results confirming the classification effectiveness of the proposed approach and showing the benefits of using attentional maps to detect biases in the data causing seemingly effective but completely wrong system operation;
  • Comparative and validation studies on sets differing in the number of images, pixel intensity levels, and subtypes of existing lesions;
  • An analysis of the benefits of transfer learning from one dataset to another.

2. Related Work

In recent years, we have witnessed substantial research on systems to automatically support the diagnosis and analysis of breast cancer lesions—computer-aided diagnosis systems. These include the classification, detection, and segmentation of breast tumours [14,15,16,17]. Such systems mainly exploit algorithms from the broad family of artificial intelligence, particularly deep learning algorithms.
Often, their effectiveness is on par with that of experienced radiologists, and almost always, such systems contribute to the effectiveness of clinicians if designed and used appropriately. According to [14], more than 70% of research work for breast cancer analysis is conducted on mammograms. The standard mammography examination consists of two projections for each breast—CC and MLO—and most of the research work considers these views separately.
In [18], Petrini et al., proposed a two-view classifier as a pair of single-view classifiers with the upper layers removed, followed by a concatenation of the outputs for both views. The single-view classifier was created by learning a patch classifier and modifying the top layers. The patch classifier was learned on ten patches selected from a given region of interest (ROI) and another ten from the background with a patch size of 224 × 224 px for each breast mammogram. Afterwards, the single-view model was trained on full-size mammograms (1152 × 896 px). The dual-view approach has improved performance, increasing AUC from 0.8033 ± 0.0183 to 0.8418 ± 0.0258.
Feature fusion of two-view mammograms was considered in [19] by Li et al. The authors performed a binary classification task (benign, malignant) using one patch cut from the annotation area. The size of the patch originally depended on the dimensions of the cut breast mass and was eventually changed to 512 × 512 px. The authors proposed two separate CNN models that extract CC and MLO patch features pre-trained with MLO and CC patches, respectively, and then fine-tuned with CC and MLO patches, respectively. All convolutional, pooling, and fully connected layers parameters were then frozen. Finally, two gate recurrent unit (GRU) modules that shared the same parameters in the fine-tuning process fused features from two-view mammograms. The fusion of the two views’ features improved accuracy from 91.0% and 90.8% for CC and MLO, respectively, considered separately, to 94.7% when fused.
Deep neural networks (DNN) are naturally predisposed to efficiently handle vast amounts of data, including medical data. However, many issues related to the effective and, importantly, trustworthy use of AI technologies must be tackled. The most common and effective DNNs are very large. For instance, the popular DNN architecture ResNet-50 consists of 48 convolutional layers along with 1 MaxPooling and 1 AveragePooling layer, which brings a total of 25.6 million parameters [20]. Another example is EfficientNet-B0, which is mainly built with mobile inverted bottleneck blocks, convolutional layers, one pooling layer, and a fully connected layer with about 5.3 million parameters as a whole [21]. We often take advantage of supervised learning to optimise the parameters of the neural models. It requires a considerable amount of good-quality and balanced training data.
However, in the case of many domains, including medicine, collecting data, especially annotated, is extremely costly, time-consuming, and thus, tricky. The main reasons are legal and privacy restrictions [22,23,24] and the need for manual annotations of the datasets, requiring much time from high-level medical experts.
In particular, the main challenge is not the availability of data but the availability of their annotations at the required precision level [25]. Manual data labelling by medical experts is laborious, time-consuming, and subjective [26,27]. Hence, in medical practice, we often experience situations where only a tiny part of data is precisely marked and labelled at the pixel level. At the same time, the rest is described just at the image level [26,28,29]. Different regularisation methods, such as data augmentation [30,31,32], or approaches to model training, such as unsupervised, weakly, or self-supervised learning, are used to cope with such conditions [33,34,35,36]. Given the nature of medical diagnostics, mainly the requirement for high performance and reliability, a supervised learning approach is still the best choice. Nevertheless, it is vital to seek strategies that minimise mentioned drawbacks and risks.
Another significant problem is that deep learning structures are black box models that map a given input to a target output [37,38,39]. Due to the complexity of the architecture, the large number of parameters and the specificity of the algorithms employed, they suffer from a lack of transparency during the training and decision-making processes. This lack of transparency during the decision-making process and unclear rules of reasoning concern end-users and make it difficult or sometimes even impossible to safely use it in many fields of our lives, especially in medicine.
The general lack of trust is additionally intensified by several cases in which DNNs made decisions based on wrong deduction or because it was fooled by an adversarial attack [40]. Even the EU General Data Protection Regulation states that every person has the right to an explanation, which means that DNN-based systems should provide explanations of decisions being undertaken [23,41].
The other important issue is that DNN training may be affected by irrelevant noisy areas or unwanted artefacts in the images [13]. Detailed analysis of the reasoning process can help to find and eliminate errors in deduction, to prove that DNN is working as expected, or even to extract new and essential features unnoticed by humans [41,42,43].
Most of the work on deep learning interpretability has focused on understanding the grounds of decisions [38,41,44]. In computer vision, interpretability is achieved mainly by visualisation techniques, verbal explanations, and analytical explanations [44,45]. The subcategory of visual explanations covers methods such as attention maps, visual saliency in heatmaps, and visualising class-related patterns [38,46]. An exciting branch of visual explanations is back-propagation techniques [47,48,49], which allow building attention maps that are self-consistent, consistent in the input domain, and consistent in the space of models [50] in contrast to methods such as sensitivity analysis [51] or class activation maps [52].
One of the emerging approaches is attention guidance [53,54], where the guidance provided, for example, by attention maps highlights relevant regions and suppresses unimportant ones, enabling a better classification. A similar method is based on self-erasing networks that prohibit attention from spreading to unexpected background regions by erasing unwanted areas [55,56]. Researchers developed methods incorporating human expert knowledge into training to avoid the negative influence of medically irrelevant features or artefactual data (Human in the Loop—HIL). HIL AI-based systems [57,58] are systems where a person (e.g., field expert) interacts with a learning system and manually answers questions, selects interesting regions, points out errors in decisions and premises, and simply acts as a guide. In the context of the issues mentioned above, both literature analysis and our research indicate that an effective and reliable approach addressing the problems identified is a form of weakly supervised learning, namely Multiple Instance Learning (MIL) [16,59,60,61,62,63].
MIL-based algorithms enable the generation of attention maps, analysing and indicating the relevance of input data areas for decision making by the neural models. These maps allow experts and end-users to quickly verify the relevance and correctness of the premises, thereby increasing the system’s transparency and reliability. Furthermore, they can guide clinicians on which parts of the analysed image are worth paying particular attention to when making a final diagnosis.
Features of medical datasets such as the relatively small number of annotated data, especially at pixel level (segmentation masks), the large size of the images (e.g., 3518 × 2800 px), and class imbalance make the direct use of commonly used neural models difficult or even impossible. Problems of insufficient or unbalanced data can be addressed to some extent through regularisation techniques, e.g., data augmentation and transfer learning. Due to the large size of the images, the straightforward deployment of the pre-trained deep neural networks such as ResNet or DenseNet [20,64] made available by other researchers is difficult and ineffective because they are designed to operate on a much smaller input size of 224 × 224 px. Recent research concerning the classification of breast cancer in mammograms with deep learning made use of high-resolution images resized to such dimensions as 64 × 64, 192 × 192, 224 × 224, and 256 × 256 px or utilised only a fragment of an image surrounding the region of interest (ROI) with a similar size as previously listed [65,66,67,68]. Reducing images to such a resolution causes the loss of valuable information about breast cancer symptoms [69,70]. In small breasts, it might lead to severely limited model performance caused by abnormalities becoming invisible.
On the other hand, leaving images at their original resolution is impractical and inefficient, as the extracted image features are diluted by a global averaging pooling mechanism that produces a fixed-size representation. Moreover, the images are characterised by a high noise-to-signal ratio—tissues affected by lesions covering a small portion of the examined body part, frequently smaller than 2% of breast volume [71,72]. The task of small object recognition is still a challenge in computer vision, not only in the medical domain.
An analysis of recent achievements in developing deep learning algorithms employed in the medical domain has shown a promising class of supervised learning algorithms called Multiple Instance Learning [60,73]. According to this approach, the whole image is divided into smaller fragments called instances or patches, which form a bag. The bag of instances is given a class based on the image-level annotations. The representations of the individual instances must be aggregated (e.g., using pooling) to obtain a prediction of the entire bag. Successful medical applications of MIL to date include the detection of malignant lesions on histopathology whole-slide imaging (WSI) [61] or the detection of diabetic retinopathy from the eye’s fundus images [74].

3. Proposed Methodology

In this paper, we describe and analyse an approach for breast lesion classification and coarse localisation of cancerous tumour lesions, taking advantage of a Clustering-constrained Attention Multiple Instance Learning deep learning classifier. We conducted several comparative analyses of breast lesions’ diagnostic performance based on solely CC images, purely MLO images, analysis of combined CC and MLO images, and multimodal analysis of CC and MLO images, with comprehensive ablation studies showing the influence of hyperparameters on system performance and interpretability. Assessment of mammography always includes two projections, MLO and CC, assessed by the radiologist simultaneously. Lesions should be visible in both projections to confirm the radiological diagnosis. Therefore, both images carry some vital information on cancer characteristics and features. Hence, pooling information from both projections should improve the predictive value of the algorithm.
During the research, we utilised and analysed full-field digital mammography MLO and CC data gathered and annotated at an image level by the Medical University of Gdansk, Poland (MUG dataset). Furthermore, we researched another dataset, the Chinese Mammography Database (CMMD) [75]. The experiments and the obtained results were comprehensively discussed and commented on from the point of view of both medical and ML specialists.

3.1. Datasets

The Medical University of Gdańsk (MUG) dataset consists of 1968 full-field digital mammography images (FFDM) from 789 patients (Table 1) undergoing mammography between 2012 and 2022 at the University Clinical Centre in Gdansk, Poland. All cancer images, classified as BI-RADS 4a–c or 5, had a pathological confirmation of the ground truth determined by the histopathological result of the biopsy or post-surgical specimens. The non-cancer group consisted of benign lesions, mainly BI-RADS 2 or 3, and images without any lesions (BI-RADS 1) with a negative follow-up with at least one additional negative mammography or ultrasound scan. In the non-cancer group, several patients were verified by the biopsy if BI-RADS = 3 and in all cases of BI-RADS 4a–c. We included female patients aged 18 or older who underwent mammography of both breasts with radiological reports available and histopathological assessment when needed, all performed in the mentioned clinical centre. Minors, patients with a previous history of breast cancer treatment, any other breast surgery and incomplete data were excluded. Another used dataset named Chinese Mammography Database (CMMD) is publicly available [75] and described in [76]. It consists of 3744 FFDM images acquired from 1775 patients undergoing mammography between 2012 and 2016. The type of benign or malignant tumours was confirmed by biopsy. For each dataset, the entire cohort was divided into two subsets: patients with breast cancer and patients without. There was no overlap of patients between the groups in the MUG dataset, while there were 30 overlapping patients in the CMMD dataset. The clinical details of the patient cohorts are shown in Table 1 and Figure 1.
The study protocol for data collection and storage complies with Polish GDPR. Approval for the study was sought from the Independent Bioethics Committee for Scientific Research at the Medical University of Gdańsk. Due to using anonymised, retrospective data, the requirement for patient-informed consent was waived. All images from the MUG dataset were obtained with a single scanner Siemens Mammomat Inspiration according to the national acquisition protocol. In the case of the CMMD dataset, two scanners were used, the SIEMENS Mammomat Inspiration and the GE Senographe DS, and the image acquisition protocol remains unknown to us. Additional information, including data specifications for both datasets, can be found in Table 2.
The images in MUG and CMMD datasets differ in pixel depth and pixel spacing. Figure 2 illustrates the difference in pixel intensities among images of these two datasets, and the DICOM image overlaying plane is enabled to show a ruler.

3.2. Multiple Instance Learning-Based Classification

In deep learning image classification tasks, Multiple Instance Learning aims to synthesise a model that can predict a bag label, e.g., suggestion of medical diagnosis. An image is sliced into small patches with a fixed size and overlapping factors to form a bag. The overlapping factor is a hyperparameter value between 0 and 1 that is selected beforehand, indicating the degree of overlap (0–100%) between adjacent square patches in the x and y axes. When an image is sliced into patches, the process starts from the top left corner of the image. Patches of a fixed square size are cropped with a stride in each axis direction equal to one minus the overlapping factor and then multiplied by the patch size. Due to varying breast sizes, some patches may contain just a tiny part of the breast or completely cover the image background. Thus, we discard those to save computational time. A set of those selected patches is merged into a bag that inherits the whole image label. The deep MIL pipeline consists of a feature extractor in the form of a convolutional neural network that encodes patches into features (low-dimensional soft embeddings), a pooling mechanism to obtain bag embedding, and a classifier [61]. The attention-pooling mechanism is an essential element of the pipeline (Figure 3).

3.2.1. Attention Pooling Mechanism

In many fields, including sensitive and socially relevant ones such as medicine, the explainability of the model is crucial for its trustworthy and, thus, safe use [37,77,78]. The MIL model prediction might be interpreted by skilfully visualising the critical instances discerned by an aggregation operator corresponding to the attention mechanism (Figure 4, Equation (1)). This mechanism functions as a weighted average of instances’ features, whose weights are determined by the dedicated attention neural network that consists of two fully connected layers. The hyperbolic tangent activation function is utilised in one of the layers for proper gradient flow during network training. This construction allows the discovery of (dis)similarities among instances. However, this is inefficient in cases where complex relations must be identified due to approximated linearity for values within the range <−1; 1>. A gated attention pooling involving a sigmoid activation function in the second layer of the neural network is used to alleviate this issue. Afterwards, both layers’ outputs are multiplied element-wise. Mammograms being analysed within the project differ in the number of valuable patches because some patches contain a small amount of breast tissue or negligible image background. Hence, in this case, the softmax function is applied to overcome susceptibility to variable bag size, ensuring that the network weights sum up to one. Obtained weights and extracted features for each instance are utilised in producing the final bag embedding in a weighted average manner. Such generated weights may be employed in attention map production corresponding with image regions that contributed the most to the final prediction. The resolution of such an attention map is directly related to the size of patches and their overlapping factor.
Firstly, each patch undergoes a feature extraction process using a convolutional neural network with the last fully connected layer removed. This results in a set of K patches being encoded in a lower-dimensional feature space H = { h 1 , , h K } . Afterwards, a pooling mechanism is applied to compute the attention scores. The bag embedding B is finally generated by decoding the patch features and their corresponding attention weights in a weighted sum manner by following MIL pooling:
B = k = 1 K a k h k
where a k is the k-th patch attention weight:
a k = e x p w T t a n h V h k T s i g m U h k T j = 1 K e x p w T t a n h V h j T s i g m U h j T
where w R L × 1 , V R L × M and U R L × M are learnable parameters, and is an element-wise multiplication.
Finally, bag embedding is processed by the classification layer, which outputs bag prediction [61]. The bag classifier is a fully connected neural network consisting of a single hidden layer with 256 neurons, followed by an output layer with a single neuron. The output layer applies a sigmoid activation function to produce a probability score between 0.0 and 1.0. A score above 0.5 corresponds to a positive prediction (cancer), while a score below 0.5 corresponds to a negative prediction (non-cancer). Bag level loss corresponds to the entire image prediction generated by the bag classifier by processing the bag embedding. The binary cross-entropy loss is computed by comparing the predicted image-level prediction score with the ground truth image-level label. The importance of each patch in the final prediction depends on computed attention scores, according to Equation (2).

3.2.2. Clustering-Constrained Attention

Due to the relatively small number of positive patches in the positively labelled bags, we took advantage of an exciting approach described in [74], which focuses on grouping and then distinguishing patches as opposite evidence. In this work, the authors employed the MIL system for whole-slide imaging in pathology classification problems, which involved constraining the attention module (Figure 5) by embedding a component that relies on identifying instances of high and low diagnostic values. The operation first sorts the attention scores and then selects the k instances with the largest scores and the k instances with the lowest scores. This process groups them into two clusters: positive and negative evidence, each containing k elements. As in the original CLAM paper, the parameter k was set to 8. The additional task of supervised learning in the network training process is applied for constraining and refining the instance-level feature space by more distinctly separating the features responsible for positive or negative evidence. The instance-level classifier is a fully connected neural network with one hidden layer containing 256 neurons. It is used to classify instance labels for the selected positive and negative evidence patches and, consequently, to compute the instance-level loss. The instance-level loss corresponds to 2000 patches selected during the instance-level clustering, in which pseudo labels were assigned by the positive or negative evidence cluster. The instance-level prediction scores for each selected patch are compared against their corresponding pseudo-cluster labels using the cross-entropy loss. In that fashion, we could exploit more concentrated attention weight values to produce attention maps focused to a greater extent.

3.3. Implementation Details

3.3.1. Data Preprocessing for Machine Learning Model

The pixel intensity values of each single-channel image were linearly scaled to the 0–1 range and then repeated to 3 channels owing to CNNs, which perform image processing on multi-channelled images. Afterwards, the image was divided into fixed-size, square-shaped tiles, called instances or patches, to compose a set called a bag. The composition of the bag was made by selecting instances containing at least 75% non-zero pixels to avoid including the background image and thus reducing computational time. Figure 6 shows an example of a patch-wise image division and selecting tiles that meet the condition of having more than 75% non-zero pixels. The sub-image on the far right illustrates an example of a single patch and its two neighbouring patches that overlap by 50% (overlapping factor: 0.5) in both the x and y directions.
The experiments tested three different patch sizes and three different overlap ratios (128 px, 224 px, 448 px for patch sizes and 0.00, 0.25, 0.50 for overlap ratios), and every combination is shown in Table 3. Furthermore, the study examined the full resolution of the images and images resized by half. Additionally, the information hidden within the full-field digital mammography MLO and CC images individually and jointly were explored through data fusion in utilising both views of the breast (CC and MLO)—multimodal view.
Initial works on the MUG dataset identified the need for an additional preprocessing technique since the attention module was, in some cases, overly attentive to the edges of the image. This was caused by a noise in the form of a high pixel-intensity line located at the edge of the image, which was related to collimator misalignment [79]. Each breast image was shifted in the chest wall direction by an a priori-selected number of pixels to eliminate this artefact.

3.3.2. Data Augmentation

Due to limited training data, several data augmentation techniques were applied, including rotating, flipping, and resized cropping. Each transformation was performed at the instance level. Augmentations were applied sequentially in a commonly used online approach—images were augmented on the fly. The angle of rotation was randomly selected from 4 predefined values. In the case of Resized Crop, the cropped portion of the instance was always resized back to the previous instance size. The amount of data generated depends on probabilities, possible choices, and the number of training epochs. Since we utilised a random oversampling method, the number of new samples generated for each class was approximately equal. The exact parameters are shown in Table 4. Data augmentation was applied only for the training data, while validation and test data remained untransformed apart from a pixel intensity scaling and an image channel expansion.

3.3.3. Models Used

The input images of the shape H × W × C (H—height; W—width; C—channels) are first converted into bag form of the shape N × H′ × W′ × C (N—number of patches; H′—patch height; W′—patch width; C—channels), and then patch features are extracted by shared CNN. We used feature extractors pre-trained on ImageNet [80] in the form of CNNs such as ResNet-18, ResNet-34, ResNet-50, and EfficientNet-B0 [20,21]. The classification head was removed from each network to access the extracted features. The feature extractors being utilised are listed and compared in Table 5.
In the attention-based module, the feature space undergoes dimensionality reduction using an FC layer with a ReLU activation function outputting 256 features describing each patch in the bag. Afterwards, an attention pooling mechanism consisting of 2 FC layers is applied to obtain bag embedding, as described in Section 3.2.1. In each FC layer contained in the attention-based model, a dropout regularisation technique was applied with a probability of zeroing a neuron equal to 0.25 during the training. At the last stage, produced bag embedding is classified by the FC layer with 256 neurons and sigmoid activation function, resulting in whole image prediction. Figure 7 illustrates the general classification pipeline.

3.3.4. Training Parameters

In each experiment, the models are trained for at least 20 epochs by introducing the early stopping method (20 epochs patience) and up to 100 epochs with the learning rate l r = 0.001 , linearly decayed every ten epochs by a constant factor g a m m a = 0.9 , L2 regularisation (weight decay) w d = 0.001 is used, and m o m e n t u m = 0.9 with stochastic gradient descent (SGD) as the optimiser. Due to GPU memory constraints, we set the batch size to 1, meaning that only one case at a time was processed. However, our shared feature extractor CNN perceives the input bag as a shape with a batch size equal to the number of instances in the bag. Therefore, we decided to freeze BatchNorm2d layers located in the pre-trained CNN. Additionally, we applied a gradient accumulation technique with 8 sample steps during the training. This solution imitated a larger batch size, and the sequentially accumulated results were used to update network weights once every eight samples. In the last network update during each epoch, this value may be different if the number of data samples is not divisible by eight. During the training, we used the binary cross-entropy (BCE) loss function for the bag classification task and the categorical cross-entropy (CCE) loss function for the constraining instance-level feature space task. The total loss was computed as the sum of both losses, with a scaling factor of 0.7 for the former and 0.3 for the latter.
The 5-fold cross-validation (CV) was used to estimate general model performance. The dataset was randomly divided into training, validation, and test sets (70%, 10%, and 20%, respectively) considering the uniqueness of the patient’s presence in the subsets. The validation set was obtained by further splitting the training set (4 folds) and used for monitoring the model’s performance during training and model selection. In contrast, the test set (1 fold) was held out to evaluate selected models. The number of patients for the training, validation, and test set in each fold from the MUG dataset is shown in Table 6.
Due to the class imbalance, we applied a weighted random oversampling method (ROS) during the training to enable the algorithm to provide predictions more reliably [81]. This approach works well due to assigning a greater probability of taking a sample from the minority class proportional to the imbalance. In combination with many variants of data augmentation applied to each instance separately, the diversity of cases alleviates the disadvantages of ROS.

3.3.5. Software and Hardware

All code was implemented in Python (v3.8) using PyTorch (v1.10.1) framework and Scikit-learn (v0.24.2) as the primary deep learning packages. Random number generators were initialised with a seed equal to 42. Experiments were conducted on two workstations, one working under Ubuntu 18.04 with Intel(R) i9-10920X (3.50 GHz) CPU and NVIDIA GeForce GTX 1080 Ti (12 GB) GPU, and the other working under Linux Mint 20.3 with AMD Ryzen 9 3950X (3.50 GHz) CPU and NVIDIA GeForce RTX 3090 (24 GB) GPU.

4. Results

Five metrics were employed to assess the performance of each model: Accuracy (3), Precision (4), Recall (5), F1-score (6), and receiver operating characteristic area under the curve (ROC AUC), where TP is true positive, TN is true negative, FP is false positive, and FN is false negative.
A c c u r a c y = T P + T N T P + T N + F P + F N
P r e c i s i o n = T P T P + F P
R e c a l l = T P T P + F N
F 1   s c o r e = 2 T P 2 T P + F P + F N

4.1. Ablation Study

We investigated the contribution of the undertaken regularisation techniques described in Section 3.3 to the system performance. The details of the experiments carried out are as follows: patient IDs were randomly split into the train, validation, and test sets (75%, 12.5%, and 12.5%, respectively), with ResNet-18 as the feature extractor, fixed base parameters for a patch size of 224 × 224 px, and an overlapping factor of 0.5. Table 7 summarises the detailed metrics results for the ablation study, allowing some conclusions to be drawn. Resizing an image by half its original size while keeping the patch size fixed significantly boosts all the metrics (about six p.p). As expected, data augmentation methods led to enhanced performance of the model. The use of dropout regularisation slightly worsened most metrics. The application of a weighted sampler resulted in minor refinement. However, by combining every single previously mentioned change with the accumulating gradient, allowing us to imitate a larger batch size, the outcome was significantly upgraded.

4.2. A Single-View Breast Cancer Classification

In the second set of experiments, we trained the system using a single-view breast image (CC or MLO). We used three different data pools during the training and evaluation process to comparatively analyse breast cancer diagnosis performance, and the pools are as follows:
  • Only CC view images (MLO images were not used);
  • Only MLO view images (CC images were not used);
  • Both CC and MLO view images (each view treated as a separate case).
For the a priori-selected patch size, overlapping factor, and CNN for extracting features, which allowed us to utilise available GPU memory fully, we obtained results gathered in Table 8 and shown in the boxplot (Figure A1Appendix A). The 5-fold cross-validation, as specified in Section 3.3.4, was performed to generate more objective results.

4.3. A Multimodal-View Breast Cancer Classification

As mentioned above, both CC and MLO projections contain valuable information indicating the presence and nature of cancerous lesions. We concatenated both views during the training to investigate the effectiveness of such an approach using the proposed CAD system. This method improved metric scores compared to the cases when using the image views separately during the training. As before, the following experiments were conducted with ResNet-18 backbone for two image resolutions (whole and halved), and more variants were tested for patch sizes and overlap ratios. The obtained results are presented in Table 9 and shown in the boxplot (Figure A2Appendix A).

4.4. Comparison with Other CNN Pre-Trained Models for Multimodal-View Breast Cancer Classification

The method described in Section 4.3 was assessed on popular CNNs other than ResNet-18, which include ResNet-34, ResNet-50, and EfficientNet-B0. The results of performed tests are shown in Table 10 and in the boxplot (Figure A3Appendix A). When comparing the AUC-ROC value, ResNet-34 was the best, while the performance of the ResNet-50 was noticeably worse.

4.5. Identification and Influence of Artefacts on Classification—Advantages of the MIL Explainability Mechanism

After training the model with attention weights, the breast regions that contributed the most to the model’s decision can be visualised. Received attention maps might be consulted with domain specialists. Consequently, it is possible to verify if the model works appropriately, not only through obtained metrics, which may be misleading. The first model trainings were conducted on multimodal-view, resized images with no regularisation mechanisms at all, the resulting metrics were surprisingly high, and the model required a small number of epochs to complete the training. Introducing more regularisation techniques into the training did not improve the results. It led us to analyse the results in each fold in great detail, including an analysis of the generated attentional maps. A comprehensive analysis of the attention maps revealed that the most attended regions were those at the edge of the image near the chest wall. As it turned out, the reason for that was the presence of “an edge” artefact in some images, examples of which are shown in Figure 8.
Furthermore, there was a high correlation in the data between images affected by artefacts and images classified as cancerous (ground truth). Therefore, the system erroneously inferred that the presence of an “edge” artefact was a premise for classifying such images as cancerous. To confirm this observation, we eliminated the “edge” artefact from all images within the MUG dataset during preprocessing. This rationalised the learning process, which could be observed both in the metrics and the learning time. Furthermore, the accompanying attentional maps analysis showed that the system started paying attention to patches with medical diagnostic value instead of those lying at the images’ edge. A comparison of the results obtained using biased and preprocessed images (by removing a vertical bar of 20 pixels width from each image) is shown in Table 11 and in the boxplot (Figure A4Appendix A).

4.6. Interpretable Decisions—Attention Heatmaps

Attention weights obtained as a result of the classification process might be used for visualisation purposes. First, an empty image (matrix) is created with the same shape as the original image. Then, the attention score of each patch is added to the corresponding square area of the image determined by the patch’s coordinates. The pixel values are then averaged based on the number of overlaid patches that contain that pixel. Finally, min–max scaling is applied to enable proper visualisation of shades of grey for human perception. The values in the heat maps depend on the patch’s diagnostic value. Namely, the brighter the area on the map is, the more critical it is. Higher heat map resolution can be achieved by increasing the overlap rate while using a patch size suitable for the trained model.
The usefulness of using interpretable methods in the form of heat maps can be seen by comparing images a–e with image f from Figure 9. It can be observed that the correct prediction of cancer in f is caused by dataset-specific image artefacts, which mainly occurred in cancer cases. Hence, the attention module learned to focus on patches taken near the edge of the image.

4.7. Summary of Results with Optimal Threshold Selection

The default value for the threshold in binary classification with class labels 0 (non-cancer) and 1 (cancer) is 0.5. This means the predicted class is cancer when the score exceeds 0.5. When lower than 0.5, the prediction is non-cancer. This section presents the results obtained after optimising the threshold by maximising the Youden index (Table 12 and Figure 10).

4.8. Evaluation with the CMMD Dataset

Since the CMMD dataset does not have a predetermined test set, only the general performance of the model might be compared to other research papers. For this reason, we employed a 5-fold cross-validation technique taking into account patient-wise breakdowns. Additionally, we used transfer learning of the knowledge gained by the model learned on the MUG dataset described in Section 4.3 to solve a closely related problem on the CMMD dataset, where the main difference is the image domain and presence of the lesion subtypes. Table 13 presents the obtained results next to the main MIL parameters and shows if the concatenated MLO and CC breast view image was resized and whether the transfer learning was applied. Sample attention maps of correct and incorrect predictions are presented in Figure 11.

5. Discussion

Acquisition of pixel-level annotated medical imaging data is very time-consuming and requires the involvement of specialist radiologists. Therefore, using weakly supervised learning methods to deal with data annotated just on the image level is practical. Multiple Instance Learning-based models are well suited for medical classification problems due to the possibility to efficiently handle large-resolution images by patching the whole image into smaller fragments. Additionally, due to the use of an attention-based pooling mechanism, it is possible to visualise the reasoning process of the model. Thus, the location of the lesion may be provided without the need to use pixel-level annotations during the training.
Furthermore, we showed a practical example that ensuring the availability of insight into the decision-making process of deep learning algorithms is essential. Especially in the analysis of medical images, model performance can be significantly biased by the presence of medically unimportant artefacts that may be invisible at first glance. This has a positive effect on the radiologists’ confidence in the decision taken by the system and allows machine learning specialists to search for biases and improve the system’s performance.
Model hyperparameters and employed regularisation techniques significantly affect the performance. In particular, imitating larger batch size through gradient accumulation and counteracting data imbalance with a weighted sampler combined with severe data augmentation applied on each patch help face the lack and imbalance of data problems. Parameters of MIL in the image domain, such as overlapping factor and patch size, seem crucial in tuning, but available GPU memory makes it hardware-dependent. We found that the best results were obtained for a patch size of 224 × 224 pixels, and this is probably caused by the fact that it is the standard input size of used feature extractor CNNs. On the other hand, such resolution also seems suitable for detecting and distinguishing relevant cancer features. With larger overlapping factors, more patches form a bag, which may affect the performance positively or negatively. Gathering more differing image fragments artificially increases available data size, and there is a bigger probability of, at best, covering tumour regions. However, for large amounts of patches, the attention weights might not be adequately learned and thus diluted. The best metrics were obtained for overlapping of 25% in each axis, which supports the above reasoning.
In light of the research we conducted, it is evident that combining CC and MLO views into the same bag substantially improves the model performance by taking advantage of both projections (about a 6 p.p. increase compared with the single CC or MLO). Using only single-view CC or MLO images resulted in similar metrics with a slight advantage in favour of the MLO view. This is consistent with clinical practice, where it provides more diagnostic value and better assessment of a larger portion of the mammary gland, especially the tail of Spence. However, the model might be harder to learn due to the presence of patches containing pectoral muscle regions, which disturb the bag with its high-intensity pixel patches.
Every used CNN has many learnable parameters sufficient for data scarcity problems. This might be why models performed slightly better when extracting lower-dimensional feature spaces because it was simpler to learn attention networks. Experiments using different CNNs as feature extractors performed somewhat better for ResNet18 and ResNet34, producing 512-dimensional feature space, less than ResNet50 and EfficientNet-B0.
In addition, we verified our approach on the Chinese Mammography Database (CMMD) dataset. It is a relatively new publicly available breast cancer dataset, and not many papers refer to it yet. Therefore, our work may be a valuable reference for the other researchers who will analyse this dataset in their studies. The CMMD does not have a predetermined test set, so due to the potential differences in test conditions, a direct comparison of methods is impossible, and only overall performance can be compared. Stadnick et al. [82] tested five open-sourced screening mammography classifiers on multiple public datasets, including CMMD. The AUC-ROC results presented by [82] range between 0.449 and 0.831, and our 5-fold cross-validation resulted in an average AUC-ROC of 0.848, outperforming each of the seven reported methods for CMMD. Walsh, Tardy in [83] trained and tested the CMMD dataset using different techniques for class imbalance. The reported AUC-ROC results obtained by [83] are in the range of 0.681–0.727. Our averaged results without transfer learning AUC-ROC of 0.772 and with transfer learning AUC-ROC of 0.848 are better than any of the four experimental procedures used.
Finally, as shown in this paper, assembling appropriate deep learning algorithms for cancer image analysis and classification requires a lot of domain knowledge and close cooperation between medical and AI teams. Approaches involving end-users in all stages, such as attention maps in the form of heatmaps, allow a better understanding of causes for system predictions. In some cases, it might enable the identification of erroneous predictions based on the wrong premises, which can cause many problems for patients and their doctors using the system.

6. Conclusions

This paper presented the approach for breast cancer diagnosis based on the Clustering-constrained Attention Multiple Instance Learning (CLAM) algorithm. The approach has the advantage of training on data labelled just at the image level. Moreover, it allows effective diagnosis even with a relatively small number of high-resolution images.
We have experimentally confirmed that, in line with medical knowledge, the best results can be obtained from a joint (multimodal) analysis of images from the CC and MLO projections.
We comprehensively analysed the impact of the system’s hyperparameters and selected regularisation methods on its efficiency. Ablation study results were carefully examined, and the CAD system was finally synthesised on this basis.
We also validated the system on a publicly available CMMD database, which enabled us to compare the results with those reported by other researchers. It should be emphasised that these databases differed significantly in terms of the number of images, pixel intensity levels, and subtypes of existing lesions. The results achieved surpass those reported in the literature.
Furthermore, we showed that a system trained on one dataset might be successfully employed to train a system on a terminal dataset more effectively in the transfer learning process.
It was confirmed that, with appropriately chosen hyperparameters, the CLAM algorithm’s attentional maps indicate the features most relevant to the algorithm in the images. Notably, the indicated spots are indeed medically relevant. Our valuable discovery is the identification of an artefact in the analysed dataset, previously unnoticed by the medical experts and clinically irrelevant, which significantly distorted the results of the CAD system.
The AI-based system developed is effective both on the database for which it was prepared and on another one—after additional training. Furthermore, the visualisation in the form of attentional maps of the features on which the diagnosis was founded makes the system more trustworthy for the end users, i.e., the medical specialists. It also allows ML specialists to discover biases and artefacts in the datasets and flaws in the algorithm itself more efficiently.
We are convinced that developing effective and trustworthy AI-based computer-aided diagnosis systems is only possible through the synergetic cooperation of specialists from different fields—AI, ML, and decision support specialists on the one hand and radiologists, oncologists, surgeons, and medical physicists on the other.

Author Contributions

Conceptualisation, M.B., J.B., R.B. and M.G.; methodology, M.B., J.B., R.B., M.F., A.K. and M.G.; software, J.B. and R.B.; validation, J.B., R.B., M.F., A.K. and M.G.; formal analysis, M.B., J.B., R.B. and M.G.; investigation, M.B., M.R., J.B., R.B., M.F., A.K. and M.G.; resources, M.B., M.R. and E.S.; data curation, M.R., J.B. and R.B.; writing—original draft preparation, M.B., M.R., J.B., R.B., M.F. and M.G.; writing—review and editing, all authors; visualisation, M.R., J.B. and R.B.; supervision, M.B. and M.G.; project administration, M.B., J.B., R.B. and M.G.; funding acquisition, M.B. and E.S. All authors have read and agreed to the published version of the manuscript.

Funding

This project was supported with funding from the European Union’s Horizon 2020 research and innovation programme under grant agreement no. 952103. This work was partly supported by the MINIATURA 5 grant 2021/05/X/NZ5/01776 of the National Science Center (NCN), Poland, GUT IDUB RADIUM grants 23/2022/IDUB/III.1/Ra and 24/2022/IDUB/III.1/Ra, GUT IDUB TECHNETIUM grant 26/1/2022/IDUB/III.4.1/Tc.

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki, and approved by the Independent Bioethics Committee for Scientific Research at the Medical University of Gdańsk, decision number NKBBN/355/2020, provided on 3 July 2020.

Informed Consent Statement

Patient consent was waived by the Independent Bioethics Committee for Scientific Research at the Medical University of Gdańsk due to the retrospective character of dataset collection, no active intervention, and multilevel anonymisation of data preventing the discovery of personal information.

Data Availability Statement

The Chinese Mammography Database (CMMD) is available at the Cancer Imaging Archive: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=70230508#70230508e1942f5ac79345328ce7713bd023e3cc (accessed on 15 March 2023). The private MUG dataset presented in this study is available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

Figure A1. Boxplots of 5-fold cross-validation metrics for single-view breast cancer classification.
Figure A1. Boxplots of 5-fold cross-validation metrics for single-view breast cancer classification.
Cancers 15 02704 g0a1
Figure A2. Boxplots of 5-fold cross-validation metrics for multi-view breast cancer classification.
Figure A2. Boxplots of 5-fold cross-validation metrics for multi-view breast cancer classification.
Cancers 15 02704 g0a2
Figure A3. Boxplots of 5-fold cross-validation metrics for multi-view breast cancer classification with different backbones as a feature extractor.
Figure A3. Boxplots of 5-fold cross-validation metrics for multi-view breast cancer classification with different backbones as a feature extractor.
Cancers 15 02704 g0a3
Figure A4. Boxplots of 5-fold cross-validation metrics for models trained with and without the presence of image artefacts.
Figure A4. Boxplots of 5-fold cross-validation metrics for models trained with and without the presence of image artefacts.
Cancers 15 02704 g0a4

References

  1. GLOBOCAN 2020: New Global Cancer Data|UICC. Available online: https://www.uicc.org/news/globocan-2020-new-global-cancer-data (accessed on 16 March 2023).
  2. McDonald, E.S.; Clark, A.S.; Tchou, J.; Zhang, P.; Freedman, G.M. Clinical Diagnosis and Management of Breast Cancer. J. Nucl. Med. Off. Publ. Soc. Nucl. Med. 2016, 57 (Suppl. 1), 9S–16S. [Google Scholar] [CrossRef] [PubMed]
  3. Zielonke, N.; Kregting, L.M.; Heijnsdijk, E.A.M.; Veerus, P.; Heinävaara, S.; McKee, M.; de Kok, I.M.C.M.; de Koning, H.J.; van Ravesteyn, N.T. The Potential of Breast Cancer Screening in Europe. Int. J. Cancer 2021, 148, 406–418. [Google Scholar] [CrossRef] [PubMed]
  4. Mathew, J.; Sibbering, M. Breast Cancer Screening. In Breast Cancer Management for Surgeons: A European Multidisciplinary Textbook; Wyld, L., Markopoulos, C., Leidenius, M., Senkus-Konefka, E., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 147–156. ISBN 9783319566733. [Google Scholar]
  5. Le, E.P.V.; Wang, Y.; Huang, Y.; Hickman, S.; Gilbert, F.J. Artificial Intelligence in Breast Imaging. Clin. Radiol. 2019, 74, 357–366. [Google Scholar] [CrossRef] [PubMed]
  6. Loizidou, K.; Elia, R.; Pitris, C. Computer-Aided Breast Cancer Detection and Classification in Mammography: A Comprehensive Review. Comput. Biol. Med. 2023, 153, 106554. [Google Scholar] [CrossRef]
  7. External Evaluation of 3 Commercial Artificial Intelligence Algorithms for Independent Assessment of Screening Mammograms|Breast Cancer|JAMA Oncology|JAMA Network. Available online: https://jamanetwork.com/journals/jamaoncology/article-abstract/2769894 (accessed on 16 March 2023).
  8. Zhao, J.; Chen, T.; Cai, B. A Computer-Aided Diagnostic System for Mammograms Based on YOLOv3. Multimed. Tools Appl. 2022, 81, 19257–19281. [Google Scholar] [CrossRef]
  9. Wu, N.; Phang, J.; Park, J.; Shen, Y.; Huang, Z.; Zorin, M.; Jastrzębski, S.; Févry, T.; Katsnelson, J.; Kim, E.; et al. Deep Neural Networks Improve Radiologists’ Performance in Breast Cancer Screening. IEEE Trans. Med. Imaging 2020, 39, 1184–1194. [Google Scholar] [CrossRef]
  10. Kim, H.-E.; Kim, H.H.; Han, B.-K.; Kim, K.H.; Han, K.; Nam, H.; Lee, E.H.; Kim, E.-K. Changes in Cancer Detection and False-Positive Recall in Mammography Using Artificial Intelligence: A Retrospective, Multireader Study. Lancet Digit. Health 2020, 2, e138–e148. [Google Scholar] [CrossRef]
  11. McKinney, S.M.; Sieniek, M.; Godbole, V.; Godwin, J.; Antropova, N.; Ashrafian, H.; Back, T.; Chesus, M.; Corrado, G.S.; Darzi, A.; et al. International Evaluation of an AI System for Breast Cancer Screening. Nature 2020, 577, 89–94. [Google Scholar] [CrossRef]
  12. Rodriguez-Ruiz, A.; Lång, K.; Gubern-Merida, A.; Broeders, M.; Gennaro, G.; Clauser, P.; Helbich, T.H.; Chevalier, M.; Tan, T.; Mertelmeier, T.; et al. Stand-Alone Artificial Intelligence for Breast Cancer Detection in Mammography: Comparison with 101 Radiologists. JNCI J. Natl. Cancer Inst. 2019, 111, 916–922. [Google Scholar] [CrossRef]
  13. Mikołajczyk, A.; Grochowski, M.; Kwasigroch, A. Towards Explainable Classifiers Using the Counterfactual Approach—Global Explanations for Discovering Bias in Data. J. Artif. Intell. Soft Comput. Res. 2021, 11, 51–67. [Google Scholar] [CrossRef]
  14. Madani, M.; Behzadi, M.M.; Nabavi, S. The Role of Deep Learning in Advancing Breast Cancer Detection Using Different Imaging Modalities: A Systematic Review. Cancers 2022, 14, 5334. [Google Scholar] [CrossRef] [PubMed]
  15. Fulton, L.; McLeod, A.; Dolezel, D.; Bastian, N.; Fulton, C.P. Deep Vision for Breast Cancer Classification and Segmentation. Cancers 2021, 13, 5384. [Google Scholar] [CrossRef] [PubMed]
  16. Elmoufidi, A. Deep Multiple Instance Learning for Automatic Breast Cancer Assessment Using Digital Mammography. IEEE Trans. Instrum. Meas. 2022, 71, 1–13. [Google Scholar] [CrossRef]
  17. Mahmood, T.; Li, J.; Pei, Y.; Akhtar, F.; Rehman, M.U.; Wasti, S.H. Breast Lesions Classifications of Mammographic Images Using a Deep Convolutional Neural Network-Based Approach. PLoS ONE 2022, 17, e0263126. [Google Scholar] [CrossRef] [PubMed]
  18. Petrini, D.G.P.; Shimizu, C.; Roela, R.A.; Valente, G.V.; Folgueira, M.A.A.K.; Kim, H.Y. Breast Cancer Diagnosis in Two-View Mammography Using End-to-End Trained EfficientNet-Based Convolutional Network. IEEE Access 2022, 10, 77723–77731. [Google Scholar] [CrossRef]
  19. Li, H.; Niu, J.; Li, D.; Zhang, C. Classification of Breast Mass in Two-view Mammograms via Deep Learning. IET Image Process. 2021, 15, 454–467. [Google Scholar] [CrossRef]
  20. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  21. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. In Proceedings of the 36th International Conference on Machine Learning, PMLR, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
  22. Kondylakis, H.; Ciarrocchi, E.; Cerda-Alberich, L.; Chouvarda, I.; Fromont, L.A.; Garcia-Aznar, J.M.; Kalokyri, V.; Kosvyra, A.; Walker, D.; Yang, G.; et al. Position of the AI for Health Imaging (AI4HI) Network on Metadata Models for Imaging Biobanks. Eur. Radiol. Exp. 2022, 6, 29. [Google Scholar] [CrossRef] [PubMed]
  23. EUR-Lex-32018R1725-EN-EUR-Lex. Available online: https://eur-lex.europa.eu/eli/reg/2018/1725/oj (accessed on 16 March 2023).
  24. Hamon, R.; Junklewitz, H.; Sanchez, M.J.I. Robustness and Explainability of Artificial Intelligence. Available online: https://publications.jrc.ec.europa.eu/repository/handle/JRC119336 (accessed on 16 March 2023).
  25. Zhou, Y.; He, X.; Huang, L.; Liu, L.; Zhu, F.; Cui, S.; Shao, L. Collaborative Learning of Semi-Supervised Segmentation and Classification for Medical Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: New York, NY, USA, 2019; pp. 2074–2083. [Google Scholar]
  26. Lee, R.S.; Gimenez, F.; Hoogi, A.; Miyake, K.K.; Gorovoy, M.; Rubin, D.L. A Curated Mammography Data Set for Use in Computer-Aided Detection and Diagnosis Research. Sci. Data 2017, 4, 170177. [Google Scholar] [CrossRef]
  27. Moreira, I.C.; Amaral, I.; Domingues, I.; Cardoso, A.; Cardoso, M.J.; Cardoso, J.S. INbreast: Toward a Full-Field Digital Mammographic Database. Acad. Radiol. 2012, 19, 236–248. [Google Scholar] [CrossRef]
  28. Heath, M.; Bowyer, K.; Kopans, D.; Moore, R.; Kegelmeyer, W.P. The Digital Database for Screening Mammography. In Proceedings of the 5th International Workshop on Digital Mammography, Toronto, ON, Canada, 11–14 June 2000; pp. 212–218. [Google Scholar]
  29. Sawyer-Lee, R.; Gimenez, F.; Hoogi, A.; Rubin, D. Curated Breast Imaging Subset of Digital Database for Screening Mammography (CBIS-DDSM). 2016. Available online: https://wiki.cancerimagingarchive.net/pages/viewpage.action?pageId=22516629 (accessed on 22 February 2023).
  30. Mikołajczyk, A.; Grochowski, M. Style Transfer-Based Image Synthesis as an Efficient Regularization Technique in Deep Learning. In Proceedings of the 2019 24th International Conference on Methods and Models in Automation and Robotics (MMAR), Międzyzdroje, Poland, 26–29 August 2019; pp. 42–47. [Google Scholar]
  31. Alomar, K.; Aysel, H.I.; Cai, X. Data Augmentation in Classification and Segmentation: A Survey and New Strategies. J. Imaging 2023, 9, 46. [Google Scholar] [CrossRef]
  32. Kaur, P.; Khehra, B.S.; Mavi, E.B.S. Data Augmentation for Object Detection: A Review. In Proceedings of the 2021 IEEE International Midwest Symposium on Circuits and Systems (MWSCAS), Lansing, MI, USA, 9–11 August 2021; pp. 537–543. [Google Scholar]
  33. Garrucho, L.; Kushibar, K.; Osuala, R.; Diaz, O.; Catanese, A.; del Riego, J.; Bobowicz, M.; Strand, F.; Igual, L.; Lekadir, K. High-Resolution Synthesis of High-Density Breast Mammograms: Application to Improved Fairness in Deep Learning Based Mass Detection. Front. Oncol. 2023, 12, 7424. [Google Scholar] [CrossRef] [PubMed]
  34. Kwasigroch, A.; Grochowski, M.; Mikołajczyk, A. Self-Supervised Learning to Increase the Performance of Skin Lesion Classification. Electronics 2020, 9, 1930. [Google Scholar] [CrossRef]
  35. Mikołajczyk, A.; Grochowski, M. Data Augmentation for Improving Deep Learning in Image Classification Problem. In Proceedings of the 2018 International Interdisciplinary PhD Workshop (IIPhDW), Świnoujście, Poland, 9–12 May 2018; pp. 117–122. [Google Scholar] [CrossRef]
  36. Mikołajczyk, A.; Majchrowska, S.; Carrasco Limeros, S. The (de) Biasing Effect of GAN-Based Augmentation Methods on Skin Lesion Images. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2022, Singapore, 18–22 September 2022; Wang, L., Dou, Q., Fletcher, P.T., Speidel, S., Li, S., Eds.; Springer Nature Switzerland: Cham, Switzerland, 2022; pp. 437–447. [Google Scholar]
  37. Marti-Bonmati, L.; Koh, D.-M.; Riklund, K.; Bobowicz, M.; Roussakis, Y.; Vilanova, J.C.; Fütterer, J.J.; Rimola, J.; Mallol, P.; Ribas, G.; et al. Considerations for Artificial Intelligence Clinical Impact in Oncologic Imaging: An AI4HI Position Paper. Insights Imaging 2022, 13, 89. [Google Scholar] [CrossRef] [PubMed]
  38. Chakraborty, S.; Tomsett, R.; Raghavendra, R.; Harborne, D.; Alzantot, M.; Cerutti, F.; Srivastava, M.; Preece, A.; Julier, S.; Rao, R.M.; et al. Interpretability of Deep Learning Models: A Survey of Results. In Proceedings of the 2017 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computed, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), San Francisco, CA, USA, 4–8 August 2017; pp. 1–6. [Google Scholar]
  39. Interpretability and Fairness Evaluation of Deep Learning Models on MIMIC-IV Dataset|Scientific Reports. Available online: https://www.nature.com/articles/s41598-022-11012-2 (accessed on 16 March 2023).
  40. Jin, J.; Dundar, A.; Culurciello, E. Robust Convolutional Neural Networks under Adversarial Noise. arXiv 2015, arXiv:1511.06306. [Google Scholar]
  41. Arrieta, A.B.; Díaz-Rodríguez, N.; Del Ser, J.; Bennetot, A.; Tabik, S.; Barbado, A.; Garcia, S.; Gil-Lopez, S.; Molina, D.; Benjamins, R.; et al. Explainable Artificial Intelligence (XAI): Concepts, Taxonomies, Opportunities and Challenges toward Responsible AI. Inf. Fusion 2020, 58, 82–115. [Google Scholar] [CrossRef]
  42. Tjoa, E.; Guan, C. A Survey on Explainable Artificial Intelligence (XAI): Toward Medical XAI. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 4793–4813. [Google Scholar] [CrossRef]
  43. Reyes, M.; Meier, R.; Pereira, S.; Silva, C.; Dahlweid, M.; Tengg-Kobligk, H.; Summers, R.; Wiest, R. On the Interpretability of Artificial Intelligence in Radiology: Challenges and Opportunities. Radiol. Artif. Intell. 2020, 2, e190043. [Google Scholar] [CrossRef]
  44. Mohseni, S.; Zarei, N.; Ragan, E.D. A Multidisciplinary Survey and Framework for Design and Evaluation of Explainable AI Systems. ACM Trans. Interact. Intell. Syst. 2021, 11, 1–45. [Google Scholar] [CrossRef]
  45. Phillips, P.J.; Hahn, C.A.; Fontana, P.C.; Yates, A.N.; Greene, K.; Broniatowski, D.A.; Przybocki, M.A. Four Principles of Explainable Artificial Intelligence; National Institute of Standards and Technology: Gaithersburg, MD, USA, 2021. [Google Scholar]
  46. Ouarti, N.; Carmona, D. Out of the Black Box: Properties of Deep Neural Networks and Their Applications. arXiv 2018, arXiv:1808.04433. [Google Scholar]
  47. Montavon, G.; Lapuschkin, S.; Binder, A.; Samek, W.; Müller, K.-R. Explaining Nonlinear Classification Decisions with Deep Taylor Decomposition. Pattern Recognit. 2017, 65, 211–222. [Google Scholar] [CrossRef]
  48. Bach, S.; Binder, A.; Montavon, G.; Klauschen, F.; Müller, K.-R.; Samek, W. On Pixel-Wise Explanations for Non-Linear Classifier Decisions by Layer-Wise Relevance Propagation. PLoS ONE 2015, 10, e0130140. [Google Scholar] [CrossRef] [PubMed]
  49. Zhang, J.; Bargal, S.A.; Lin, Z.; Brandt, J.; Shen, X.; Sclaroff, S. Top-Down Neural Attention by Excitation Backprop. Int. J. Comput. Vis. 2018, 126, 1084–1102. [Google Scholar] [CrossRef]
  50. Samek, W.; Binder, A.; Montavon, G.; Lapuschkin, S.; Müller, K.-R. Evaluating the Visualisation of What a Deep Neural Network Has Learned. IEEE Trans. Neural Netw. Learn. Syst. 2017, 28, 2660–2673. [Google Scholar] [CrossRef] [PubMed]
  51. Baehrens, D.; Schroeter, T.; Harmeling, S.; Kawanabe, M.; Hansen, K. How to Explain Individual Classification Decisions. J. Mach. Learn. Res. 2010, 11, 1803–1831. [Google Scholar]
  52. Zhou, B.; Khosla, A.; Lapedriza, A.; Oliva, A.; Torralba, A. Learning Deep Features for Discriminative Localization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; IEEE: New York, NY, USA, 2016; pp. 2921–2929. [Google Scholar]
  53. Li, K.; Wu, Z.; Peng, K.-C.; Ernst, J.; Fu, Y. Tell Me Where to Look: Guided Attention Inference Network. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 9215–9223. [Google Scholar]
  54. Guan, Q.; Huang, Y.; Zhong, Z.; Zheng, Z.; Zheng, L.; Yang, Y. Diagnose like a Radiologist: Attention Guided Convolutional Neural Network for Thorax Disease Classification. arXiv 2018, arXiv:1801.09927. [Google Scholar]
  55. Hou, Q.; Jiang, P.; Wei, Y.; Cheng, M.-M. Self-Erasing Network for Integral Object Attention. In Advances in Neural Information Processing Systems; Curran Associates, Inc.: Red Hook, NY, USA, 2018; Volume 31. [Google Scholar]
  56. Wei, Y.; Feng, J.; Liang, X.; Cheng, M.-M.; Zhao, Y.; Yan, S. Object Region Mining with Adversarial Erasing: A Simple Classification to Semantic Segmentation Approach. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; IEEE: New York, NY, USA, 2017; pp. 6488–6496. [Google Scholar]
  57. Akrout, M.; Farahmand, A.; Jarmain, T. Improving Skin Condition Classification with a Question Answering Model. In Proceedings of the 32nd Conference on Neural Information Processing Systems (NIPS 2018), Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
  58. Rupprecht, C.; Laina, I.; Navab, N.; Hager, G.D.; Tombari, F. Guide Me: Interacting with Deep Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8551–8561. [Google Scholar]
  59. Herrera, F.; Ventura, S.; Bello, R.; Cornelis, C.; Zafra, A.; Sánchez-Tarragó, D.; Vluymans, S. Multiple Instance Learning. In Multiple Instance Learning: Foundations and Algorithms; Herrera, F., Ventura, S., Bello, R., Cornelis, C., Zafra, A., Sánchez-Tarragó, D., Vluymans, S., Eds.; Springer International Publishing: Cham, Switzerland, 2016; pp. 17–33. ISBN 9783319477596. [Google Scholar]
  60. Ilse, M.; Tomczak, J.; Welling, M. Attention-Based Deep Multiple Instance Learning. In Proceedings of the 35th International Conference on Machine Learning, PMLR, Stockholm, Sweden, 10–15 July 2018; pp. 2127–2136. [Google Scholar]
  61. Li, B.; Li, Y.; Eliceiri, K.W. Dual-Stream Multiple Instance Learning Network for Whole Slide Image Classification with Self-Supervised Contrastive Learning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 14318–14328. [Google Scholar] [CrossRef]
  62. Quellec, G.; Lamard, M.; Cozic, M.; Coatrieux, G.; Cazuguel, G. Multiple-Instance Learning for Anomaly Detection in Digital Mammography. IEEE Trans. Med. Imaging 2016, 35, 1604–1614. [Google Scholar] [CrossRef]
  63. Xiong, D.; Zhang, Z.; Wang, T.; Wang, X. A Comparative Study of Multiple Instance Learning Methods for Cancer Detection Using T-Cell Receptor Sequences. Comput. Struct. Biotechnol. J. 2021, 19, 3255–3268. [Google Scholar] [CrossRef]
  64. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar] [CrossRef]
  65. Thilagaraj, M.; Arunkumar, N.; Govindan, P. Classification of Breast Cancer Images by Implementing Improved DCNN with Artificial Fish School Model. Comput. Intell. Neurosci. 2022, 2022, 6785707. [Google Scholar] [CrossRef]
  66. Elkorany, A.S.; Elsharkawy, Z.F. Efficient Breast Cancer Mammograms Diagnosis Using Three Deep Neural Networks and Term Variance. Sci. Rep. 2023, 13, 2663. [Google Scholar] [CrossRef]
  67. Zahoor, S.; Shoaib, U.; Lali, I.U. Breast Cancer Mammograms Classification Using Deep Neural Network and Entropy-Controlled Whale Optimisation Algorithm. Diagnostics 2022, 12, 557. [Google Scholar] [CrossRef]
  68. Castro-Tapia, S.; Castañeda-Miranda, C.L.; Olvera-Olvera, C.A.; Guerrero-Osuna, H.A.; Ortiz-Rodriguez, J.M.; Martínez-Blanco, M.d.R.; Díaz-Florez, G.; Mendiola-Santibañez, J.D.; Solís-Sánchez, L.O. Classification of Breast Cancer in Mammograms with Deep Learning Adding a Fifth Class. Appl. Sci. 2021, 11, 11398. [Google Scholar] [CrossRef]
  69. How Does Image Quality Affect Radiologists’ Perceived Ability for Image Interpretation and Lesion Detection in Digital Mammography?|SpringerLink. Available online: https://link.springer.com/article/10.1007/s00330-020-07679-8 (accessed on 19 March 2023).
  70. Warren, L.M.; Halling-Brown, M.D.; Looney, P.T.; Dance, D.R.; Wallis, M.G.; Given-Wilson, R.M.; Wilkinson, L.; McAvinchey, R.; Young, K.C. Image Processing Can Cause Some Malignant Soft-Tissue Lesions to Be Missed in Digital Mammography Images. Clin. Radiol. 2017, 72, 799.e1–799.e8. [Google Scholar] [CrossRef] [PubMed]
  71. Lagendijk, M.; Vos, E.L.; Ramlakhan, K.P.; Verhoef, C.; Koning, A.H.J.; van Lankeren, W.; Koppert, L.B. Breast and Tumour Volume Measurements in Breast Cancer Patients Using 3-D Automated Breast Volume Scanner Images. World J. Surg. 2018, 42, 2087–2093. [Google Scholar] [CrossRef] [PubMed]
  72. Clauser, P.; Londero, V.; Como, G.; Girometti, R.; Bazzocchi, M.; Zuiani, C. Comparison between Different Imaging Techniques in the Evaluation of Malignant Breast Lesions: Can 3D Ultrasound Be Useful? Radiol. Med. 2014, 119, 240–248. [Google Scholar] [CrossRef] [PubMed]
  73. Lu, M.Y.; Williamson, D.F.K.; Chen, T.Y.; Chen, R.J.; Barbieri, M.; Mahmood, F. Data-Efficient and Weakly Supervised Computational Pathology on Whole-Slide Images. Nat. Biomed. Eng. 2021, 5, 555–570. [Google Scholar] [CrossRef] [PubMed]
  74. Papadopoulos, A.; Topouzis, F.; Delopoulos, A. An Interpretable Multiple-Instance Approach for the Detection of Referable Diabetic Retinopathy in Fundus Images. Sci. Rep. 2021, 11, 14326. [Google Scholar] [CrossRef] [PubMed]
  75. Cui, C.; Li, L.; Cai, H.; Fan, Z.; Zhang, L.; Dan, T.; Li, J.; Wang, J. The Chinese Mammography Database (CMMD): An Online Mammography Database with Biopsy Confirmed Types for Machine Diagnosis of Breast. Data Cancer Imaging Arch 2021. [Google Scholar] [CrossRef]
  76. Cai, H.; Wang, J.; Dan, T.; Li, J.; Fan, Z.; Yi, W.; Cui, C.; Jiang, X.; Li, L. An Online Mammography Database with Biopsy Confirmed Types. Sci. Data 2023, 10, 123. [Google Scholar] [CrossRef]
  77. Ferlin, M.; Klawikowska, Z.; Niemierko, J.; Grzywinska, M.; Kwasigroch, A.; Szurowska, E.; Grochowski, M. Trustworthy Applications of ML Algorithms in Medicine—Discussion and Preliminary Results for a Problem of Small Vessels Disease Diagnosis. In Intelligent and Safe Computer Systems in Control and Diagnostics; Kowalczuk, Z., Ed.; Springer International Publishing: Cham, Switzerland, 2023; pp. 3–16. [Google Scholar]
  78. Responsible AI Practices. Available online: https://ai.google/responsibilities/responsible-ai-practices/ (accessed on 21 March 2023).
  79. Ayyala, R.S.; Chorlton, M.; Behrman, R.H.; Kornguth, P.J.; Slanetz, P.J. Digital Mammographic Artifacts on Full-Field Systems: What Are They and How Do I Fix Them? RadioGraphics 2008, 28, 1999–2008. [Google Scholar] [CrossRef]
  80. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  81. Wongvorachan, T.; He, S.; Bulut, O. A Comparison of Undersampling, Oversampling, and SMOTE Methods for Dealing with Imbalanced Classification in Educational Data Mining. Information 2023, 14, 54. [Google Scholar] [CrossRef]
  82. Stadnick, B.; Witowski, J.; Rajiv, V.; Chłędowski, J.; Shamout, F.E.; Cho, K.; Geras, K.J. Meta-Repository of Screening Mammography Classifiers. arXiv 2021, arXiv:2108.04800. [Google Scholar]
  83. Walsh, R.; Tardy, M. A Comparison of Techniques for Class Imbalance in Deep Learning Classification of Breast Cancer. Diagnostics 2023, 13, 67. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Patients class distribution in MUG and CMMD datasets.
Figure 1. Patients class distribution in MUG and CMMD datasets.
Cancers 15 02704 g001
Figure 2. Data sample from (a) MUG dataset and (b) CMMD dataset.
Figure 2. Data sample from (a) MUG dataset and (b) CMMD dataset.
Cancers 15 02704 g002
Figure 3. The general pipeline of the Clustering-constrained Attention Multiple Instance Learning-based classification scheme.
Figure 3. The general pipeline of the Clustering-constrained Attention Multiple Instance Learning-based classification scheme.
Cancers 15 02704 g003
Figure 4. Attention pooling mechanism.
Figure 4. Attention pooling mechanism.
Cancers 15 02704 g004
Figure 5. Attention-constraining module.
Figure 5. Attention-constraining module.
Cancers 15 02704 g005
Figure 6. Selection of image patches to form a bag.
Figure 6. Selection of image patches to form a bag.
Cancers 15 02704 g006
Figure 7. General attention-based MIL classification pipeline.
Figure 7. General attention-based MIL classification pipeline.
Cancers 15 02704 g007
Figure 8. Mammographic images of the breast with visible artefacts: (a,c) default view, (b,d) view after application of windowing.
Figure 8. Mammographic images of the breast with visible artefacts: (a,c) default view, (b,d) view after application of windowing.
Cancers 15 02704 g008
Figure 9. Left: whole FFDM image and ground truth label; middle: selected image tiles (patches); right: class prediction, score, and produced heatmap for (a) single-view approach using MLO view only, (b) single-view approach using CC view only, (c,d) single-view approach using MLO and CC views separately, respectively, (e) multimodal-view approach using MLO and CC views simultaneously, (f) same as (e) but with image artefacts not removed.
Figure 9. Left: whole FFDM image and ground truth label; middle: selected image tiles (patches); right: class prediction, score, and produced heatmap for (a) single-view approach using MLO view only, (b) single-view approach using CC view only, (c,d) single-view approach using MLO and CC views separately, respectively, (e) multimodal-view approach using MLO and CC views simultaneously, (f) same as (e) but with image artefacts not removed.
Cancers 15 02704 g009aCancers 15 02704 g009b
Figure 10. Boxplots of 5-fold cross-validation metrics comparison with optimised thresholds.
Figure 10. Boxplots of 5-fold cross-validation metrics comparison with optimised thresholds.
Cancers 15 02704 g010
Figure 11. Left: whole FFDM image and ground truth label; middle: selected image tiles (patches); right: class prediction, score, and produced heatmap for (ac) true positive and (d) false negative prediction.
Figure 11. Left: whole FFDM image and ground truth label; middle: selected image tiles (patches); right: class prediction, score, and produced heatmap for (ac) true positive and (d) false negative prediction.
Cancers 15 02704 g011aCancers 15 02704 g011bCancers 15 02704 g011c
Table 1. Clinical details of the patient cohort.
Table 1. Clinical details of the patient cohort.
MUG DatasetCMMD Dataset
Non-CancerCancerNon-CancerCancer
Number of patients392397465 + 30 *1280 + 30 *
Median Age
(Min–Max)
51
(21–86)
57
(29–88)
43
(17–84)
49
(21–87)
Number of images
of the left breast **
5824125621414
Number of images
of the right breast **
5923825461222
* Both groups contain the same 30 patients with bilateral images with benign findings in one breast and malignant in the other. ** CC images and MLO images in total.
Table 2. Specification of data.
Table 2. Specification of data.
MUG DatasetCMMD Dataset
Type of dataDICOM files—full-field digital mammography craniocaudal (CC) and mediolateral oblique (MLO) images
Excel file—age, pathological ground truth
DICOM files—full-field digital mammography craniocaudal (CC) and mediolateral oblique (MLO) images
Excel file—age, pathological ground truth
Image acquisition
protocol
National acquisition protocolUnknown
Image acquisitionSIEMENS Mammomat InspirationSIEMENS Mammomat Inspiration,
GE Senographe DS
Image resolution3518 × 2800 pixels2294 × 1914 pixels
Pixel spacing0.085 mm0.094 mm
Pixel depth12-bit (0–4095)8-bit (0–255)
Data source locationUniversity Clinical Center in Gdańsk, PolandSunYat-sen University Cancer Center in Guangzhou,
Nanhai Affiliated Hospital of Southern Medical University in Fushan, China
Table 3. Experimental parameters.
Table 3. Experimental parameters.
ExperimentImage Size
(px)
Patch Size
(px)
Overlap
(-)
Single view3518 × 2800 *224 **0.50
Single view with resize1759 × 14002240.50
Multimodal view7036 × 28001280.50
2240.00, 0.25, 0.50
4480.50
Multimodal view with resize3518 × 14001280.50
2240.00, 0.25, 0.50
4480.50
* Without resizing, one pixel corresponds to the breast area size of 0.085 × 0.085 mm. ** Thus, the patch in size of 224 × 224 px = 19.04 × 19.04 mm.
Table 4. Instance/patch-level augmentations.
Table 4. Instance/patch-level augmentations.
No.Augmentation MethodsParameters
1RotateAngle = (0°, 90°, 180°, 270°), Probability = 1
2Horizontal FlipProbability = 0.5
3Vertical FlipProbability = 0.5
4Resized CropScale range [0.8–1.0], Probability = 1
Table 5. Comparison of feature extractors.
Table 5. Comparison of feature extractors.
Feature ExtractorNumber of ParametersFeature Space Dimension
ResNet-1811.7 M512
ResNet-3421.8 M512
ResNet-5025.6 M2048
EfficientNet-B05.3 M1280
Table 6. The number of patients split between three sets within five subsequent folds from the MUG dataset.
Table 6. The number of patients split between three sets within five subsequent folds from the MUG dataset.
FoldTraining SetValidation SetTest Set
1(non-cancer)2804072
(cancer)2723986
2(non-cancer)2703884
(cancer)2824174
3(non-cancer)2743781
(cancer)2784277
4(non-cancer)2783975
(cancer)2744083
5(non-cancer)2813180
(cancer)2724877
Table 7. Ablation study results—the influence of employed regularisation techniques on system performance. (✔—applied; ✖—not applied).
Table 7. Ablation study results—the influence of employed regularisation techniques on system performance. (✔—applied; ✖—not applied).
ResizeTransformsDropoutWeighted
Sampler
Gradient
Accumulation
AUC-ROC
(-)
F1-Score
(%]
Accuracy
(%)
Precision
(%)
Recall
(%)
0.8171.072.074.072.0
0.8778.078.079.078.0
0.8781.081.081.081.0
0.8779.079.080.079.0
0.8980.080.080.080.0
0.9288.088.088.088.0
Table 8. Single-view classification results for 5-fold cross-validation.
Table 8. Single-view classification results for 5-fold cross-validation.
Feature
Extractor
ViewImage
Size (px)
Patch
Size (px)
Overlap
(-)
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
ResNet-18CC3518 × 28002240.50.792 ± 0.02672.4 ± 2.572.6 ± 2.872.8 ± 2.772.6 ± 2.8
1759 × 14000.838 ± 0.02576.2 ± 3.576.2 ± 3.777.2 ± 3.876.2 ± 3.7
MLO3518 × 28000.804 ± 0.05372.8 ± 4.972.8 ± 4.474.0 ± 4.872.8 ± 4.4
1759 × 14000.840 ± 0.00676.0 ± 1.476.0 ± 1.676.4 ± 1.476.0 ± 1.7
CC or MLO3518 × 28000.822 ± 0.02175.6 ± 2.875.8 ± 3.075.6 ± 2.975.8 ± 3.0
1759 × 14000.864 ± 0.01478.4 ± 0.478.4 ± 0.579.4 ± 0.878.6 ± 0.5
Table 9. Multimodal-view classification results for 5-fold cross-validation.
Table 9. Multimodal-view classification results for 5-fold cross-validation.
Feature
Extractor
ViewImage
Size (px)
Patch
Size (px)
Overlap
(-)
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
ResNet-18CC
and
MLO
7036 × 28001280.50.846 ± 0.02775.6 ± 4.876.0 ± 4.176.6 ± 4.476.0 ± 4.1
22400.828 ± 0.03375.8 ± 3.576.0 ± 3.575.8 ± 3.576.0 ± 3.5
2240.250.850 ± 0.03076.4 ± 4.176.6 ± 3.877.6 ± 4.376.6 ± 3.8
2240.50.844 ± 0.02677.0 ± 3.877.0 ± 3.877.4 ± 3.977.0 ± 3.8
4480.50.828 ± 0.04076.2 ± 5.776.6 ± 5.277.0 ± 5.076.6 ± 5.2
3518 × 14001280.50.894 ± 0.02279.8 ± 3.279.8 ± 2.680.2 ± 3.479.8 ± 2.6
22400.892 ± 0.01279.8 ± 1.780.0 ± 2.180.4 ± 2.180.0 ± 2.1
2240.250.896 ± 0.01781.8 ± 3.281.6 ± 3.282.4 ± 3.381.6 ± 3.2
2240.50.892 ± 0.01581.6 ± 1.981.6 ± 1.982.0 ± 1.781.6 ± 1.9
4480.50.866 ± 0.03977.2 ± 5.477.6 ± 4.778.2 ± 4.677.6 ± 4.7
Table 10. Multimodal-view classification results for 5-fold cross-validation with different backbone CNNs.
Table 10. Multimodal-view classification results for 5-fold cross-validation with different backbone CNNs.
Feature
Extractor
ViewImage
Size (px)
Patch
Size (px)
Overlap
(-)
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
ResNet-18CC
and
MLO
3518 × 14002240.50.892 ± 0.01581.6 ± 1.981.6 ± 1.982.0 ± 1.781.6 ± 1.9
ResNet-340.892 ± 0.02382.2 ± 2.682.2 ± 2.682.4 ± 2.482.2 ± 2.6
ResNet-500.878 ± 0.01379.8 ± 3.580.2 ± 3.380.8 ± 2.280.2 ± 3.3
EfficientNet-B00.890 ± 0.01981.4 ± 1.681.6 ± 1.482.2 ± 1.281.6 ± 1.4
Table 11. Metrics for 5-fold cross-validation with and without the presence of image artefacts.
Table 11. Metrics for 5-fold cross-validation with and without the presence of image artefacts.
Edge
Artefacts
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
Absent0.846 ± 0.02776.0 ± 3.776.4 ± 4.076.4 ± 4.076.4 ± 4.0
Present0.866 ± 0.02482.8 ± 2.082.8 ± 2.082.8 ± 2.082.8 ± 2.0
Table 12. Fivefold cross-validation results comparison with optimised thresholds.
Table 12. Fivefold cross-validation results comparison with optimised thresholds.
Feature
Extractor
ViewImage
Size (px)
Patch
Size (px)
Overlap
(-)
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
ResNet-18CC1759 × 14002240.50.838 ± 0.02577.6 ± 2.777.4 ± 2.778.4 ± 2.877.4 ± 2.7
MLO0.840 ± 0.00677.8 ± 0.777.8 ± 0.779.0 ± 0.677.8 ± 0.7
CC or MLO0.864 ± 0.01480.4 ± 1.780.4 ± 1.781.0 ± 1.480.4 ± 1.7
CC and MLO3518 × 14000.892 ± 0.01583.6 ± 1.283.6 ± 1.283.6 ± 1.283.6 ± 1.2
Table 13. Fivefold cross-validation results on the CMMD dataset.
Table 13. Fivefold cross-validation results on the CMMD dataset.
Transfer
Learning
ViewImage
Size (px)
Patch
Size (px)
Overlap
(-)
AUC-ROC
(-)
F1-Score
(%)
Accuracy
(%)
Precision
(%)
Recall
(%)
Not appliedCC and MLO2294 × 9572240.50.786 ± 0.05072.4 ± 3.372.0 ± 3.175.4 ± 3.972.0 ± 3.1
4588 × 19140.772 ± 0.01973.6 ± 0.573.0 ± 1.474.4 ± 1.073.0 ± 1.4
Applied2294 × 9570.790 ± 0.01373.6 ± 1.574.4 ± 1.273.8 ± 1.674.4 ± 1.2
4588 × 19140.848 ± 0.01578.6 ± 2.078.4 ± 1.978.8 ± 2.078.4 ± 1.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Bobowicz, M.; Rygusik, M.; Buler, J.; Buler, R.; Ferlin, M.; Kwasigroch, A.; Szurowska, E.; Grochowski, M. Attention-Based Deep Learning System for Classification of Breast Lesions—Multimodal, Weakly Supervised Approach. Cancers 2023, 15, 2704. https://doi.org/10.3390/cancers15102704

AMA Style

Bobowicz M, Rygusik M, Buler J, Buler R, Ferlin M, Kwasigroch A, Szurowska E, Grochowski M. Attention-Based Deep Learning System for Classification of Breast Lesions—Multimodal, Weakly Supervised Approach. Cancers. 2023; 15(10):2704. https://doi.org/10.3390/cancers15102704

Chicago/Turabian Style

Bobowicz, Maciej, Marlena Rygusik, Jakub Buler, Rafał Buler, Maria Ferlin, Arkadiusz Kwasigroch, Edyta Szurowska, and Michał Grochowski. 2023. "Attention-Based Deep Learning System for Classification of Breast Lesions—Multimodal, Weakly Supervised Approach" Cancers 15, no. 10: 2704. https://doi.org/10.3390/cancers15102704

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop