Next Article in Journal
Radiogenomics: Hunting Down Liver Metastasis in Colorectal Cancer Patients
Previous Article in Journal
The Immune Landscape of Colorectal Cancer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

State-of-the-Art Challenges and Perspectives in Multi-Organ Cancer Diagnosis via Deep Learning-Based Methods

1
Faculty of Information Technology, Beijing University of Technology, Beijing 100124, China
2
Computer Science Division, University of Aizu, Aizuwakamatsu 965-8580, Japan
3
Beijing Key Laboratory for Green Catalysis and Separation, Department of Chemistry and Chemical Engineering, Beijing University of Technology, Beijing 100124, China
4
Research Institute for Microwave and Millimeter-Wave (RIMMS), National University of Sciences and Technology (NUST), Islamabad 44000, Pakistan
*
Author to whom correspondence should be addressed.
Cancers 2021, 13(21), 5546; https://doi.org/10.3390/cancers13215546
Submission received: 26 August 2021 / Revised: 28 October 2021 / Accepted: 29 October 2021 / Published: 4 November 2021
(This article belongs to the Section Systematic Review or Meta-Analysis in Cancer Research)

Abstract

:

Simple Summary

Cancer is a deadly disease that needs to be diagnose at early stage to increase patient survival rate. Multi-organ (such as breast, brain, lung, and skin) cancer detection, segmentation and classification manually using medical imaging is time consuming and required high expertise. In this study, we summarize existing deep learning segmentation and classification methods for multi-organ cancer diagnosis and provide future challenges with possible solutions. This review may benefit researchers to design new robust approaches that could be useful for the medical specialists as a second view.

Abstract

Thus far, the most common cause of death in the world is cancer. It consists of abnormally expanding areas that are threatening to human survival. Hence, the timely detection of cancer is important to expanding the survival rate of patients. In this survey, we analyze the state-of-the-art approaches for multi-organ cancer detection, segmentation, and classification. This article promptly reviews the present-day works in the breast, brain, lung, and skin cancer domain. Afterwards, we analytically compared the existing approaches to provide insight into the ongoing trends and future challenges. This review also provides an objective description of widely employed imaging techniques, imaging modality, gold standard database, and related literature on each cancer in 2016–2021. The main goal is to systematically examine the cancer diagnosis systems for multi-organs of the human body as mentioned. Our critical survey analysis reveals that greater than 70% of deep learning researchers attain promising results with CNN-based approaches for the early diagnosis of multi-organ cancer. This survey includes the extensive discussion part along with current research challenges, possible solutions, and prospects. This research will endow novice researchers with valuable information to deepen their knowledge and also provide the room to develop new robust computer-aid diagnosis systems, which assist health professionals in bridging the gap between rapid diagnosis and treatment planning for cancer patients.

1. Introduction

Cancer diagnosis using different medical images plays a significant role in detecting various abnormalities, for instance, skin cancer [1], breast cancer [2], lung cancer [3], brain tumors [4,5], blood cancer [6], and so forth. Tumor-induced abnormalities are the leading source of universal demise [7]. The GLOBOCAN 2020 report illustrates that lung cancer (18%) is the leading cause of death; other cancers are also life-threatening for humans with different mortality rates, for example, breast cancer (6.9%) and brain cancer (2.5%) [8]. Many image modalities are utilized to analyze irregularities in different organs, such as Magnetic Resonance Imaging (MRI) [9], Positron Emission Tomography (PET) [10], Computed Tomography (CT) [11], and mammography [12].
The human brain is the most complex part of our body. The functioning of brain cells is highly influenced by the irregular mitosis mechanism. As a result, cancer cells are produced with distinct morphological properties, such as size, shape, boundaries, and so forth. Low-grade gliomas (grades I and II) and high-grade glioma (grades III and IV) are two major categories of brain tumors. Low-grade tumors grow slowly [13], while the high-grade are the most malignant primary brain tumors, which are more aggressive and disrupt the blood–brain supply [14]. Glioblastoma (GBM) [15], a grade IV glioma, is the most common, invasive, and lethal type of primary brain tumor. Cancerous cells are less contrasted than the nearby cells, making perfect brain tumor recognition challenging. Thus far, examining MR scans is one of the most effective techniques for detecting brain tumors owing to its non-invasive nature, painless test procedure and for manipulating the tumorous region from various angles [16].
A leading cause of death among women is breast cancer. Abnormal cells can be benign or malignant. Malignant breast cancer is more aggressive and threatening because it spreads to multiple body organs via the lymphatic system [17]. A benign tumor, a noncancerous tumor, is well recognized in type and has a large size, but malignant tumors are diffused and small. Due to narrow size and fatty tissue problems, the early detection of malignant tumors is challenging. The timely detection of breast cancer can help with the diagnosis procedure, which can alleviate the disease severity with more excellent recovery [18,19]. Therefore, state-of-the-art, fully automatic methods are needed for early breast tumor detection.
The formation of a certain nodule in the lungs is an indication of lung cancer. A round-shaped nodule in the lungs can be benign or malignant [20]. The malignant nodules develop swiftly, and their rapid progression might affect the other body parts. CT images are a commonly employed diagnostic technique for lung cancer detection [21].
In addition to the above-mentioned types of cancer, skin cancer is a fast dominant disease worldwide [22]. It has two categories, melanoma, and non-melanoma cancer. Melanoma cancer is considered aggressive than non-melanoma, a critical type of cancer that arises as a dark spot on the skin. Occasionally, these spots grow as a mole to progress in shape, uneven edges, and different skin colours. In the last year (2020), the death count for non-melanoma and melanoma skin cancer is 63,731 and 57,043, respectively [23,24]. To summarize, to cope with a fatal disease like cancer, early diagnosis is highly needed to guard the patients’ life, which could be done only by developing advanced CAD systems. Computer-aided procedures currently play an important role in medical image analysis. The CAD-aided segmentation and classification facilitate target separation, diagnosis, quantitative measurements, and treatment planning. Figure 1 depicts the commonly employed CAD approaches in the domain of medical image analysis. So far, many methods have been developed for various types of detection, segmentation, and classification. However, research in this area is still in its infancy. This study critically highlights the room for improvement in four major cancer types viz. lung, breast, brain, and skin in theoretical and technological ways. Moreover, this article covers cutting-edge methods for multi-organ cancer detection and diagnosis using medical images, which would be highly beneficial for novice researchers to propose the CAD system in a specific domain. Additionally, a comprehensive analysis of the most commonly used standard databases for the brain, breast, lung, and skin cancers is also elaborated. Majorly this review includes the brief discussion part where open research challenges and future directions are extensively discussed, which can assist physicians and radiologists in treatment planning.
The rest of the study is structured in several sections: Section 2 presents material and methods that include a brief explanation of the standard datasets available for multi-organ images and performance evaluation metrics. Section 3 contains the brain tumor detection methods. Section 4 is about breast cancer detection methods; Section 5 represents the lung cancer detection methods; Section 6 describes the skin lesion cancer detection methods. Section 7 is about the discussion and open research challenges along with the state-of-the-art solutions and prospects. In Section 8, the whole study is summarized under the conclusion heading.
The key motivation behind this conducted research is to look for the answer to these queries:
1.
What is the commonly employed imaging modality in each of the four cancers?
2.
What kind of databases is utilized for medical image analysis?
3.
Which kind of AI technology is in trend for the early diagnosis of these cancers?
4.
Why is CNN architecture is a trend in breast, brain, lung, and skin cancer diagnosis?
5.
What performance evaluation metrics are employed to evaluate the models’ efficiency?

2. Material and Methods

In this section, we have briefly discussed search strategy and selection criteria. Moreover, some standard datasets are described, which are implemented extensively for cancer segmentation, and classification approaches followed by performance measures are also briefly explained.

2.1. Search Strategy and Selection Criteria

In this review, we have utilized openly available search databases such as Google Scholar and PubMed to find most related articles using different queries. We have limited our search to manuscripts published between the years 2016–2021. We have used the following queries in different combinations: “brain cancer diagnosis”, “Brats dataset segmentation”, “breast cancer diagnosis”, “lung cancer diagnosis”, “skin cancer diagnosis”, “LIDC/IDRI database segmentation and classification”, “skin lesions”, “brain tumor segmentation and classification”, “breast cancer segmentation and classification”, “WBCD dataset segmentation”, “lung cancer segmentation and classification”, “skin cancer segmentation and classification”, “PH2 dataset segmentation”, “brain tumor detection using machine learning and deep learning classifiers”, “breast tumor diagnosis using machine learning and deep learning classifiers”, “lung cancer detection using machine learning and deep learning classifiers”, “skin cancer diagnosis using machine learning and deep learning classifiers”, “brain tumor MRI and deep learning”, “DDSM classification”, “artificial intelligence and breast cancer”, “artificial intelligence and brain tumor”, “artificial intelligence and lung cancer”, and so forth. More than 300 related papers are reviewed, among them 156 papers are selected for the current study, and 111 manuscripts out of 156 were most relevant to brain, breast, lung, and skin cancer diseases.

2.2. Most Popular Publicly Available Datasets

2.2.1. Multimodal Brain Tumor Image Segmentation Benchmark(BraTS) Database

BraTS dataset consisted of multi-institutional routine clinically acquired pre-operative multimodal MRI scans of High-Grade Glioma, that is, Glioblastoma (GBM/HGG) and Lower Grade Glioma (LGG), with a pathologically confirmed diagnosis and available overall survival (OS), are provided as the training, validation and testing data. In the MRI study, BraTS focuses on evaluating advanced techniques of brain tumors segmentation and classification. MRT multi-institutional pre-operative images are utilized to segment brain tumors, specifically gliomas that vary in appearance, shape, size, and histology. The patients’ OS is also brought into focus by integrative radiometric features analysis. It has eight popular dataset collections from 2012 to 2020. The datasets used in the BraTS challenge are acquired from 3T multi-modal scanners with ground truth annotated and confirmed expert board-certified neuroradiologists [25].

2.2.2. Lung Image Database Consortium image collection(LIDC/IDRI) Database

The dataset, named LIDC and IDRI, are commonly employed datasets for implementing recognition about the lung nodule. LIDC-IDRI is a freely available databank of CT lung images. It contains scans of nodule outlines and subjective nodule properties rankings. It is established to help lung nodules study and comprises 1018 cases with 244,617 CT scans and XML report files generated by experienced thoracic radiologists [26].

2.2.3. Digital Database for Screening Mammography(DDSM) Database

It is a standard reference of mammographic scans generated in Massachusetts General Hospital. Digital database for screening mammography contains 2620 cases, 1935 images composed of tumors. Each case has four-view mammography screenings images of breast and subject information. Suspicious areas and relevant pixel-level ground truth information annotated by the radiologist (normal, benign, and malignant image) are included in images [27].

2.2.4. Wisconsin Breast Cancer Database(WBCD) Database

A commonly employed multivariate Wisconsin Breast Cancer Database contains 683 patients’ information from the biopsy test of female breast cancer. There are 11 attributes on each record. The first ten columns comprise the attributes column, and the 11th column holds class attributes. WBCD has 444 benign examples and 239 malignant examples for training and testing tasks [28].

2.2.5. International Skin Imaging Collaboration(ISIC) Database

ISIC has a baseline dataset of 25,331 JPEG dermoscopic images for training and 8238 for testing skin lesions that are openly accessible. It describes outstanding problems in the segmentation and classification of skin lesions, incorporated with a high-resolution image validated by experienced experts. The dimensions of the scans are unbalanced because they use various types of visual sensors [29].

2.2.6. PH2 Database

PH2 has been developed for the detection of skin lesion cancer. It is a dermoscopic image database acquired at the Dermatology Service of Hospital Pedro Hispano, Matosinhos, Portugal. The database contains 200 melanocytic skin lesion images for research purposes, in which 80 images are benign cancer, 40 images belong to malignant, and the remaining 80 images are associated with the suspicious lesion. These are 8-bit RGB color images with having dimensions of 768 × 560 pixels [30].
The summary of the above-mentioned and other related datasets for various cancer diseases for the example database name, modality, number of images or patients, link to the source, is presented in Table 1.

2.3. Performance Evaluation Metrics

The major evaluation measures to check the performance of segmentation and classification tasks of DL models such as Accuracy (Acc), Specificity (SP), Sensitivity (SN), area under the curve (AUC), and true positive (TP) are reviewed in this research. SN, also entitled as a recall, is a probability of recognizing the segmented image’s true pixels. SP depict the ability to recognize negative pixel. Acc is defined as the ratio among the correctly identified pixels (TP + TN) and the total pixel in an image (TP + FP + TN + FN). The mathematical formulas to compute these performance measures are given in Equations (1)–(4).
S N = T P T P + F N
S P = T N T N + F P
A c c = T B + T N T P + F P + T N + F N
A U C = 1 2 ( T P T P + F N + T N T N + F P )

3. Brain Tumor

A major contributing factor towards the universal death rate is a brain tumor. The World Health Organization (WHO) broadly categorized brain tumors as benign and malignant tumors as from Grades I–IV (lowest–extreme aggressive) [31]. Grades I and II belong to the Low-grade (LG) tumors category and are also termed benign tumors. Grade III and IV belong to the High-grade (HG) tumors category and are also termed as malignant tumors [32]. LG and HG tumors also differ in terms of their growth and the number of years of life anticipation. Therefore, timely diagnosis of brain tumor is necessary to minimize the universal death rate. However, distinguishing healthy tissue from tumor is not an easy task, owing to varying shapes and sizes, poor contrast and variable locations [31]. These factors influence the complexity of tumor growth and predict the extent of resection at the time of surgical planning, this has repercussions in terms of patient management [33]. Therefore, reliable tumor classification and segmentation are important tasks for the determination of tumor size, exact position, and type. For that purpose, computed tomography (CT), biopsy, cerebral angiography, positron emission tomography (PET) and MRI are significant medical imaging modalities [34]. Among them, MRI scans have attracted greater attention owing to their non-invasive nature and their in-depth as well as objective analysis. MRI is a potent and sensitive modality for detecting brain tumors and their boundary delineation [35]. Initially, the precise analysis and comprehensive monitoring of brain tumors were dependent on the radiologist experts. However, the radiologist’s dependent process is tedious and time-consuming. Therefore, the development of advanced CAD systems truly helps radiologists for the improved and timely diagnosis of brain tumor. Many articles have been published on brain tumor detection, classification, and segmentation to date. Some researchers apply conventional ML-based feature extraction methods for brain tumor detection.
Nilesh Bhaskarrao et al. [36] proposed a brain tumor segmentation technique by applying Berkeley wavelet transform (BWT) with support vector machine SVM. BWT was used for the feature extraction task, followed by the SVM classifier to perform the classification task. The authors reveal that the results obtained 96.51%, 94.2%, and 97.72% for accuracy, specificity, and sensitivity. Alfonse et al. [37] used the SVM method for automated brain tumor segmentation and classification using MR images. Firstly, brain images are segmented, employing adaptive thresholding. Secondly, features are extracted using Fast Fourier Transform (FFT), then Minimal Redundancy Maximal Relevance methods are used for feature selection. This technique achieved 98.9% classification accuracy. In SVM, the classification of different points based on proximity accompanied by splitting hyper-plane required more execution time to calculate linear or quadratic complications. Wu et al. [38] introduced a multi-level Gabor wavelet method to reduce the linear or quadratic calculations by using image superpixels instead of image voxels for the segmentation of GBM. Extracted features are fused to SVM for classification purposes. Recently, an effort reducing the classification error using SVM is presented by Soltaninejad et al. [39] proposed to classify tumor grades (for example, II, III, and IV) using statistical features extraction. ROI was segmented manually or by a superpixel-based method. The better version of brain tumor grading for MRI image analysis is developed in [40].
Tianbao Ren et al. [41] developed an automated brain tumor segmentation approach. Initially, the authors used histogram equalization to acquire the related information. Then they implemented an improved Kernel-based Fuzzy C mean (KFCOM) with Weighted fuzzy kernel clustering (WKFCOM) model that enhances brain image segmentation performance. The results illustrate that the proposed combined algorithm achieves an improved misclassification rate which was less than 2.36%. In comparison to FCM methods, a state-of-the-art technique [42] is introduced that categorized White Matter, Grey Matter, and cerebrospinal fluid spaces using Adaptive Fuzzy K-mean (AFKM) Clustering. Researchers declare that the implemented AFKM algorithm obtains superior results in contrast to FCM both qualitatively and quantitatively.
Li et al. [43] suggested a multi-modality deep learning network for brain tumor segmentation that extracts multi-scale features of brain tissue from the MR scans. A data mining system via the combination of FCM and SVM was presented for the MR image segmentation in [44]. A combined technique with k-mean and FCM was introduced [45]. This technique implemented a median filter for MR brain images denoising and a brain surface extractor for features extraction. Then, clustering is done through the proposed method. Anitha Vishnuvarthanan et al. [46] suggested a hybrid algorithm that involves Bacteria Foraging Optimization (BFO) and a Modified FKM clustering approach for the segmentation and classification of brain MR images.
Deep learning is an emerging approach and extensively applied in many object detection applications to automatically perform feature extraction for complex patterns. The most well-known method is a convolutional neural network (CNN); an example of a 3D architecture for brain tumor detection is presented in Figure 2. Multiple hidden layers of CNN with batch normalization architecture was implemented for brain tumor classification [47].
The developed classification model was evaluated on BRATS 2013 and achieved 0.99% accuracy. Recently transfer learning approach is implemented in several medical domains. It employs pre-trained networks that are previously trained on massive datasets like ImageNet. Resent 50, GoogleNet, and VGG 19, are examples of pre-trained learning networks broadly used to resolve classification issues [48]. S. Deepak and P.M. Ameer [49] developed a CNN-based GoogleNet transfer learning classification model to classify brain tumors, including glioma, meningioma, and pituitary. The proposed algorithm attained better accuracy 92.3%, which was more enhanced to 97.8% by applying multiclass SVM. To improve the CNN-based model performance in terms of accuracy, researchers suggested a combined framework using Stationary Wavelet Transform (SWT) and Growing Convolution Neural Network (GCNN) for brain tumor segmentation [50]. SWT technique was applied for feature extraction rather than Fourier transform that provides improved results for discontinuous data followed by the Random Forest method for the classification task. The suggested technique contributes a 2% improvement compared with traditional CNN. The in-depth features are learned by applying transfer learning models like AlexNet and computing the scores of each feature matrix. The obtained feature vector scores are merged to generate a final feature vector. The combined features vector is given to the classifier to examine the performance of the proposed method [51]. The hand-crafted and machine-learned-based features fusion like semantic segmentation network (SegNet) is also used to segment brain tumors [52]. A summary of brain tumor cancer diagnosis CAD systems is illustrated in Table 2.
The systematic literature analysis showed that deep learning technology resulted in great realistic performances in brain tumor image analysis. It has been observed that the most commonly employed ML method is SVM. However, deep learning algorithms are the top performers, especially DCNN. However, the main limitation of DCNN is a dependency on massive training data with expert radiologists’ annotations from different institutions. It is a pretty tricky task. We believe the hybrid intelligent systems designed by integrating machine learning approaches with other methodologies like deeply learned approaches offer a highly proficient, accurate classification system. It appears to give higher classification accuracy in the range of 95–100%. Moreover, CNN can be known as an archetypical classifier owing to immense usage in the prognosis of various diseases such as brain tumor classification, segmentation, and detection. Most of the publications were exploited BRATS data sets for tumor diagnosis tasks.

4. Breast Cancer

Several image processing-based architectures in collaboration with artificial intelligence and ML are reported for the performance enhancement of medical detection and diagnostics processes. CAD systems are considered as a robust approach in the modern diagnosis and detection of breast cancers using medical imaging [68]. Robust CAD systems increase the characteristic of images and improve the diagnostic capability of healthcare professionals. The researchers apply five imaging modalities such as ultrasound, MRI, mammography, thermography, histology, and so forth, for breast cancer diagnosis. Generally, the human breast is observed as a supersensitive organ of the body; thus, few of these discussed medical modalities are suggested. Across all modalities, mammography is recommended because it is a reliable way to detect breast cancer in the early stages. Eight benchmark imaging databases of the breast exist and are freely available on the internet for breast cancer diagnosis, termed INbreast, Mammographic Image Analysis Society (MIAS), Wisconsin Breast Cancer Dataset (WBCD), Image Retrieval in Medical Applications (IRMA), Database for Screening Mammography (DDSM), Wisconsin Diagnosis Breast Cancer (WDBC), breast cancer data repository (BCDR), and Breast Cancer Histopathological Image (BreakHis) [69]. During pre-processing some necessary operations are applied to better image quality, like contrast improvement, noise reduction, and artifact removal.
After the preprocessing, breast mass segmentation is the next crucial stage for increasing the accuracy of detection systems with decreasing false results about existence of abnormality [70]. Segmentation of breast tumor scans is challenging due to numerous obstacles like rough or lobulated corners, breast lumps or tissues, pectoral muscle, and mutual values of mass intensities. These problems complicate the procedure of diagnosis systems in facilitating the health care experts. Usually, segmentation provides localization of breast tumors or lesions and detection within two-dimensional or three-dimensional images. A prominent feature of breast cancer is silhouette and contour, as they provide essential information about the metastatic nature of the breast mass.
From the literature survey, segmentation methods of breast mass are classified into color-based [71], contour-based, morphological-based, threshold-oriented-based [72], region-oriented, DL-based network [73,74], and hybrid segmentation methods [75]. At present, DL-based applications developed via CNN are gaining more attention for breast cancer detection and segmentation through CAD systems [74,76], as presented in Figure 3.
Breast cancer prognosis is highly dependent on the classification results of morphological samples and cell surface receptors including hormone receptor status [77]. Breast cancer classification is generally done by three methods: supervised classification, unsupervised classification, and semi-supervised classification. The perfect classification performance of morphological data depends on extracting important features such as statistical, shape, and textural [78]. These extracted features are forwarded to different ML/DL classification algorithms for training, validation, and testing model. Neural networks (NN) [79], Support vector machines (SVM) [80], and k-nearest-neighbor (KNN) [81] are few robust and powerful computational algorithms that are helpful for complex classification challenges.The use of AI systems is an additional development in the medical imaging domain. In particular, the practice of CNN-based classification algorithms and hybrid frameworks are adopted that have produced encouraging outcomes in breast cancer detection. The complete summary of current approaches is illustrated in Table 3.
The following observations are made from the aforementioned literature analysis: (1) most popular utilized classifiers are deep learning, SVM, CNN, and KNN; (2) Mammogram is the widely employed modality for breast cancer diagnosis; (3) For the forthcoming robust design of CAD schemes, the development of 3D mammogram-based CAD schemes could be the new trend; (4) For the effective and in-time diagnosis of breast cancer instead of sole mammogram, other modalities like CT, ultrasound, thermal and histological images must be considered. Boundary images should be labeled for the classification of multi-class breast cancer because they allow researchers to analyze the usefulness of the recently established multiple classes breast cancer model.

5. Lung Cancer

Advanced CAD systems help in timely lung cancer diagnosis and/or prognosis by working on the CT images using AI approaches. These CAD-based decision support systems examine the input CT scans, employing various methods to segment and classify medical images as presented in Figure 4. Various databases have been employed so far for lung cancer diagnosis for instance Automated Nodule Detection Database (ANODE09), the Early Lung Cancer Action Program (ELCAP) database, The LUNA16 dataset, the Lung Image Database Consortium, and Image Database Resource Initiative (LIDC-IDRI), and so forth. Among them, LIDC-IDRI is used as the gold standard database for the evaluation of lung cancer detection techniques [89]. The polygon approximation technique is proposed to detect nodule shape properties [90]. The intensity and geometric features vector then fused to the SVM classifier to identify actual nodules. The proposed algorithm evaluated on the LIDC dataset achieved promising results of 98.8%, 97.7%, 96.2% in terms of accuracy, sensitivity, and specificity. Currently, 3D-based segmentation is a robust method to accurately detect nodules from lung images. Many approaches have integrated these characteristics. Paing et al. [91] developed a fully automated and optimized random forest approach to classifying pulmonary nodules using tomography scans. A 3D chain code algorithm is applied to improve the borders. The public dataset contains 888 images evaluated on the proposed algorithm that obtained satisfactory results of 93.11%, 94.86%, and 91.37% accuracy, specificity, and sensitivity. The false positives per scan were only 0.086.
Deep learning makes breakthroughs owing to the latest developments in image processing, particularly medical image examination. The attention of CAD systems has shifted from traditional hand-craft features to automated deep-learned features. Its performance is superior in terms of segmentation and classification of nodular items through CT scan images. Perez, G., and Arbelaez, P. proposed a 3D CNN approach for the diagnosis of lung cancer automatically and obtained effective results for recall of 99.6% and AUC of 0.913% [92]. The model was trained on the LIDC-IDRI standard dataset to evaluate the performance. In DL, autoencoder and softmax are well-known and suitable methods for feature selection and classification. Apart from this, a different variant of the 3D CNN method is also producing superior results for extracting features automatically [93]. CNN architecture contains several hidden layers such as convolution, max-pooling, and a fully connected layer with a softmax function for image classification [94,95]. In recent years, DL-based CAD algorithms are developed, for example, 3D U-Net [96], patch-based 3D U-Net [97], and hybrid CNN algorithm [98]. Moreover, integrating ML and DL techniques also enhance the performance for segmentation as well as classification [99]. Table 4 shows existing literature on different method’s performance for lung nodules detection.
From the literature analysis of lung cancer diagnosis, it was found that the CT scan is a widely employed modality in the CAD system. The researchers believe that 3D CNNs would provide more promising results than 2D CNN owing to the ability of the model to manipulate spatial information via 3D convolution and pooling operation. Up till now few works is reported in literature where 3D CNNs are employed in medical image analysis for lung cancer diagnosis [101,105,113,114]. Overall, some reports exhibit good performances in segmentation and classification. However, still, some methods have some limitations like low sensitivity, high FP, and more time consumption. Therefore, developing robust CAD meeting the aforementioned demand is the challenge. These days using a pre-trained model is in trend, which can surpass these limitations.

6. Skin Cancer

The diagnosis of melanoma skin cancer illnesses has been conventionally detected by manual analysis and visual examination. These visual examination methods and analyses of skin lesion scans by expert dermatologists are lengthy, complicated, biased, expensive [115]. Melanoma skin cancer is a lethal illness that leads to death when not diagnosed in time. Traditionally, skin cancer was detected based on hand-crafted feature techniques that restrict the high performance of CAD approaches. Current improvements in deep learning methods in medical image analysis and computer vision have led to excessive development of CAD and detection schemes to recognize fatal cancerous skin infections [116].
DL systems automatically extract essential features and have the additional benefit of extracting features directly from the un-processed image. Recently, DCNNs achieved substantial performance in medical imaging [117] and the segmentation of skin lesion images. MAR Ratul et al. [118] proposed an automated CAD system using four CNN-based networks, namely VGG-16, VGG-19, MobileNet, and InceptionV3, to identify malignant skin lesions. The developed architectures were evaluated on the HAM10000 dataset and attained 87.42% for VGG-16, 85.02% for VGG-19, 88.22% for MobileNet, and 89.81% for InceptionV3. In another report, an improved U-Net approach named skinNet was applied using dilated convolutions in the final layer of the encoder section. Its shortcoming includes skin lesion image feature lost by the U-Net shortcut skip connection; thus, a weaker decoder branch in recuperating feature vectors [119]. Esteva et al. [120] implemented a CNN model, trained from scratch, applying dermoscopic images for evaluation that indicated the promising results of the designed CNN network. However, to train an entire architecture from scratch to identify skin lesion cancer utilizing small datasets is usually not realistic. Thus, most scientists typically use fine-tuned methods or pre-trained algorithms through transfer learning.
In another investigation [121], researchers developed a pre-trained architecture using an ensemble of deep learning models including EfficientNets, SENet, and ResNeXt WSL to address skin lesion classification. To attain state-of-the-art results, ISIC 2019 dataset was employed on the proposed model, and certain scaling rules were applied, making it flexible for each image dimension. In another study, [122] developed an effective CNN-based transfer learning SqueezeNet model for Mobile-based skin lesions classification. The networks were trained and tested on 1856 images. The squeezeNet model achieved state-of-the-art results in terms of overall accuracy of 97.21%, sensitivity 94.42%, and specificity 98.14%. CNN-based SqueezeNet architecture is presented in Figure 5.
A fully convolutional residual network (FCRN) based CAD system was suggested by [123], by employing an index calculation unit to refine final results. The mean possibilities of each map are applied for the classification of melanoma skin lesions. Dash et al. [124] introduced an improved U-Net approach called PsLSNet to extract spatial information and used 29 hidden layers of CNN. A novel CNN-based segmentation system was introduced without any preprocessing step [125]. The system learns full resolution features using CNNs. The algorithm was tested on dual datasets called PH2 and ISBI 2017, and the achieved results were compared with other methods, for instance, FCN, SegNet, and U-Net. Although DL-based learning algorithms attain outstanding acknowledgment achievements in the medical imaging domain, more investigation is still needed to improve the performance of CAD systems. The reported CAD systems can never replace doctors but can only assist doctors in the planning of skin cancer diagnosis. Table 5 shows the state-of-the-art CAD approaches of skin cancer.
The systematic review analysis reveals the fact that skin lesion diagnosis is undoubtedly a complex task. It might be attributed to the tremendous variation in image type as well as human skin color and appearance, as illustrated in Figure 6. From the analytical analysis of skin cancer literature data, we determined that ensemble methods with advanced architectures including Resnext, DenseNet121, PsLSNet, InceptionResNetV2, SqueezeNet architecture, Xception, FRCN, ResNet-50 based CNN appears to perform tremendously well. As is assessed from the literature that the utilization of appropriate pre-processing techniques like grab-cut can overcome the skin cancer image analysis challenge. Because deep learning models produce far better classification results when the well-pre-processed images are used as the input.

7. Discussion

7.1. Primary Observations

In this manuscript, we extensively analyze manuscripts published between the years 2016–2021. The comprehensive analysis is depicted in the form of a column graph, as illustrated in Figure 7. One can see that researchers employed more DL-based supervised classifiers (NN, CNN, U-Net, VGG-16, ResNet-50) as compared with ML-based (SVM, DT, RF, K-means) and other classifiers for the detection of various cancers. Deep learning-based unsupervised methods are not considered extensively.
Compared to traditional ML approaches, DL-based CAD systems have been practiced significantly in many applications. In the last few years, it has been found that most researchers have imitated a robust trend towards the development of DL, particularly CNN-based CAD systems. CNN has proven itself a better choice in the medical image analysis field since its emergence in 1989 [140] owing to accurate analysis and time-saving features. Moreover, it is the key solution for many learning problems including feature extraction and object recognition. Summarizing all, CNN has marked great success for detection, classification, and segmentation of multi-organ body cancer attributed to the well-organization and the construction of the layered structure, elements employed in each consecutive layer, and so forth [141,142]. The evident benefit of DL over traditional ML techniques is its less dependency on feature extraction methods as it can automatically develop related features by its hidden layers, that is, convolution and max-pooling [143,144]. Furthermore, DL empowers to directly extract features from input medical image data through a back-propagation operation, which automatically modifies each layer’s weights as stated parameters of supervised learning produced by validated data [144]. In contrast, the execution of a conventional CAD system usually demands regular human participation and constant engineering to confirm its smooth process. CNN-based models can increase the global detection rate of different cancer medical images whenever the inaccurate output is notified through repetitive self-learning skills. It is worth observing that additional repetitions occur between the learning phases; thus, the network’s superior is its sensitivity. Moreover, as CNN extracts and learns features from input images directly thereby, a tumor segmentation step is unnecessary. Nevertheless, in conventional approaches, the segmentation of tumors for multi-organ cancer images is essential for learning appropriate features. Therefore, essential details might have vanished mainly when a wrong segmentation appears [145]. Table 2, Table 3, Table 4 and Table 5 clearly depict determined that the CNN-based techniques [62,67,83,84,107,133] for different body organs-based CAD systems achieved efficient and state-of-the-art performances for the detection, segmentation, and classification using multi-modalities medical images. Additionally, we have highlighted a few loopholes in DL-based approaches discussed in the next section and suggested some recommendations, which can bridge the gap between research studies and early cancer diagnosis.

7.2. Open Research Challenges, Possible Solutions and Future Prospects

DL-based methods have significantly enhanced segmentation accuracy, obliged to their ability to handle complicated tasks. Firstly, DL-based methodologies still have not gained full acknowledgment among pathologists for everyday clinical exercise. One of the primary causes might be the lack of standardized medical image acquisition protocols. The unification of the acquisition protocols could minimize it. Secondly, deep learning algorithms usually need massive annotated medical images by expert radiologists to accomplish the training task. Gathering such an enormous annotated dataset of medical images is often a challenging task. Even performing the labeling of a new dataset through experienced experts will also be very time-consuming and costly. Many techniques have been commonly applied to conquer the scarcity of annotated data. Thirdly, the researchers who worked on machine learning algorithms have limited knowledge about the radiology of medical images that is also a barrier in achieving state-of-the-art performance. However, we have discussed other problems and their possible solutions below. This study also provides some recommendations, future directions, and a viewpoint for novice researchers to excel in the medical image analysis domain:
  • The most generally implemented technique to extend the range of the training dataset is named data augmentation. It is the application in which different offline changes are done, including affine transformation, cropping, flip, rotation, padding, saturation to the examples [146], and colour augmentation [147];
  • Transfer learning from the popular networks [86] employed in the same field or even another area is considered another solution to surpass limited data. It has been established that transfer learning by pre-trained networks produced superior results even when the source and target networks are not the same, transferring the weights of different tasks [148];
  • The morphological variation in the cancerous cell is one of the significant issues in medical cancer image detection. The cancerous organ/lesion may differ significantly in dimension, outline, and position from patient to patient [149]. Using deeper architectures can be an effective solution to this issue, as reported in [115]. The unclear border with an imperfect contrast among targeting organs and the nearby tissues in tumor images is an inherent challenge typically produced via attenuation coefficient [150,151]. The use of multi-modality-based methods can solve this issue [152,153];
  • The computational complexity of the network is another challenge in DL-based techniques, owing to variability in image dimensions, network construction, or the heavily over-parameterized networks. To evade the powerful GPU hardware constraint and accelerate the segmentation task, one can decrease the number of hidden layers or parameters of the proposed network and emphasize algorithms that artificially generate training data for example GAN [154,155] rather than altering the network;
  • The appearance of mostly AI-based architectures seems like a black box. Thus, researchers have no idea about the internal representations of the network and the perfect approach to realize the system completely. Hence, DL approaches are greatly affected by the inherent snags of medical images, that is, noise and illumination. Complete knowledge and understanding of such black box issues in the future would be a revolution in the DL field [156];
  • During training time, the ground truth outlines are manually delineated by expert physicians. If manual delineation would be done by a different individual or even the same one at the distinct circumstance, there must be a possibility that the proposed model can be biased and favor expert ground truths as a system error. However, this drawback could be expected to occur in all supervised learning CAD techniques;
  • The amalgamation of the robust individual approaches by utilizing their benefits is suitable in further improving the CAD performance. Develop novel CAD systems using hybrid ML-based approaches like SegNet [74], U-Net-Vnet-Fast-R-CNN [157], AgileNet [158] to overcome the complication of overfitting that happens in the training time; this could help in the early diagnosis of multi-organ cancers;
  • It is observed that the DL-based unsupervised clustering techniques include; deep auto-encoders, regularized information maximization (RIM), Deep InfoMax (DIM), deep adaptive clustering (DAC), and so forth, have not been engaged widely in comparison with supervised learning techniques [159]. It could avoid the costly training process. These techniques could also be employed to improve the performance of CAD systems in the medical imaging domain.

8. Conclusions

This survey critically analyzes four-organ cancer (brain, lung, breast, and skin) diagnosis techniques. The various algorithms were analyzed and discussed in terms of their performances, which were published from 2016 to date. However, the main focus is on recent DL-based approaches. This review work reveals the fact that MRI, CT, dermoscopic images, and mammograms are the gold standard for a brain tumor, lung cancer, skin cancer, and breast cancer diagnosis, respectively. From the comparative analysis, it was found that BRAT, LIDC-IDRI, ISIC and PH2, WDSM, and DDSM are widely employed databases for brain, lung, skin, and breast cancer, respectively. This study provides deep knowledge about the significant positive impact of deep learning on medical image analysis and strengthens the knowledge of the trend and the latest techniques over the past few years. Despite extensive research conducted in this domain so far, there still exists room for improvement. This review has elaborated on the open research challenges and their possible solutions point by point. Moreover, the performance evaluation parameters and publicly available standard datasets for each cancer disease have been briefly discussed. This study will prove highly beneficial for the researchers practicing the deep learning-based cancer diagnosis to design the robust CAD architecture that could be helpful for the medical experts as a second opinion. Some useful recommendations would prove fruitful for researchers as well. In the future, further studies will be conducted to analyze other cancer diseases such as liver and stomach cancers and so forth.

Author Contributions

S.A. conceived this study. Y.P. and R.K. contributed to the design of this study. J.L. reviewed, drafted, and revised the study. K.u.R. and A.B.R. proofread this study. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the National Key R&D Program of China with project No. 2020YFB2104402.

Acknowledgments

Authors would like to thank National Key R&D Program of China for providing experimental facilities to conduct this study.

Conflicts of Interest

The authors declare that they have no conflict of interest.

References

  1. Zhang, N.; Cai, Y.X.; Wang, Y.Y.; Tian, Y.T.; Wang, X.L.; Badami, B. Skin cancer diagnosis based on optimized convolutional neural network. Artif. Intell. Med. 2020, 102, 1–7. [Google Scholar] [CrossRef] [PubMed]
  2. Mambou, S.J.; Maresova, P.; Krejcar, O.; Selamat, A.; Kuca, K. Breast cancer detection using infrared thermal imaging and a deep learning model. Sensors 2018, 18, 2799. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Priya, S.S.; Ramamurthy, B. Lung cancer detection using image processing techniques. Res. J. Pharm. Technol. 2018, 11, 2045–2049. [Google Scholar] [CrossRef]
  4. Mohsen, H.; El-Dahshan, E.S.A.; El-Horbaty, E.S.M.; Salem, A.B.M. Classification using deep learning neural networks for brain tumors. Future Comput. Informatics J. 2018, 3, 68–71. [Google Scholar] [CrossRef]
  5. Kapoor, L.; Thakur, S. A survey on brain tumor detection using image processing techniques. In Proceedings of the 2017 7th International Conference on Cloud Computing, Data Science & Engineering-Confluence, Noida, India, 12–13 January 2017; pp. 582–585. [Google Scholar]
  6. Latha, T.S. Recognition of Blood Cancer Using Different Classification Techniques. Int. Trans. Electr. Eng. Comput. Sci. 2020, 1, 33–41. [Google Scholar]
  7. Fahad, H.; Ghani Khan, M.U.; Saba, T.; Rehman, A.; Iqbal, S. Microscopic abnormality classification of cardiac murmurs using ANFIS and HMM. Microsc. Res. Tech. 2018, 81, 449–457. [Google Scholar] [CrossRef]
  8. Sung, H.; Ferlay, J.; Siegel, R.L.; Laversanne, M.; Soerjomataram, I.; Jemal, A.; Bray, F. Global cancer statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J. Clin. 2021, 71, 209–249. [Google Scholar] [CrossRef]
  9. Sulzer, D.; Cassidy, C.; Horga, G.; Kang, U.J.; Fahn, S.; Casella, L.; Pezzoli, G.; Langley, J.; Hu, X.P.; Zucca, F.A.; et al. Neuromelanin detection by magnetic resonance imaging (MRI) and its promise as a biomarker for Parkinsons disease. NPJ Park. Dis. 2018, 4, 1–13. [Google Scholar]
  10. Lameka, K.; Farwell, M.D.; Ichise, M. Positron emission tomography. Handb. Clin. Neurol. 2016, 135, 209–227. [Google Scholar]
  11. Christe, A.; Peters, A.A.; Drakopoulos, D.; Heverhagen, J.T.; Geiser, T.; Stathopoulou, T.; Christodoulidis, S.; Anthimopoulos, M.; Mougiakakou, S.G.; Ebner, L. Computer-aided diagnosis of pulmonary fibrosis using deep learning and CT images. Investig. Radiol. 2019, 54, 627–632. [Google Scholar] [CrossRef] [Green Version]
  12. Katzen, J.; Dodelzon, K. A review of computer aided detection in mammography. Clin. Imaging 2018, 52, 305–309. [Google Scholar] [CrossRef]
  13. Brown, T.J.; Bota, D.A.; van Den Bent, M.J.; Brown, P.D.; Maher, E.; Aregawi, D.; Liau, L.M.; Buckner, J.C.; Weller, M.; Berger, M.S.; et al. Management of low-grade glioma: A systematic review and meta-analysis. Neuro-Oncol. Pract. 2019, 6, 249–258. [Google Scholar] [CrossRef] [Green Version]
  14. Cho, H.h.; Park, H. Classification of low-grade and high-grade glioma using multi-modal image radiomics features. In Proceedings of the 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Jeju Island, Korea, 11–15 July 2017; pp. 3081–3084. [Google Scholar]
  15. Ali, S.; Xia, Q.; Muhammad, T.; Liu, L.; Meng, X.; Bars-Cortina, D.; Khan, A.A.; Huang, Y.; Dong, L. Glioblastoma Therapy: Rationale for a Mesenchymal Stem Cell-based Vehicle to Carry Recombinant Viruses. Stem Cell Rev. Rep. 2021, in press. [Google Scholar] [CrossRef]
  16. Anaraki, A.K.; Ayati, M.; Kazemi, F. Magnetic resonance imaging-based brain tumor grades classification and grading via convolutional neural networks and genetic algorithms. Biocybern. Biomed. Eng. 2019, 39, 63–74. [Google Scholar] [CrossRef]
  17. Collins, O. Last Year’s Virus, This Year’s Cancer Treatment. BU Well 2018, 3, 1–4. [Google Scholar]
  18. Spanhol, F.A.; Oliveira, L.S.; Petitjean, C.; Heutte, L. Breast cancer histopathological image classification using convolutional neural networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 2560–2567. [Google Scholar]
  19. Li, J.; Pei, Y.; Yasin, A.; Ali, S.; Mahmood, T. Computer Vision-Based Microcalcification Detection in Digital Mammograms Using Fully Connected Depthwise Separable Convolutional Neural Network. Sensors 2021, 21, 4854. [Google Scholar]
  20. Husham, A.; Hazim Alkawaz, M.; Saba, T.; Rehman, A.; Saleh Alghamdi, J. Automated nuclei segmentation of malignant using level sets. Microsc. Res. Tech. 2016, 79, 993–997. [Google Scholar] [CrossRef]
  21. Shakeel, P.M.; Burhanuddin, M.; Desa, M.I. Automatic lung cancer detection from CT image using improved deep neural network and ensemble classifier. Neural Comput. Appl. 2020, in press. [Google Scholar] [CrossRef]
  22. Leiter, U.; Keim, U.; Garbe, C. Epidemiology of skin cancer: Update 2019. In Sunlight, Vitamin D and Skin Cancer; Springer: Berlin/Heidelberg, Germany, 2020; pp. 123–139. [Google Scholar]
  23. Pacheco, A.G.; Krohling, R.A. Recent advances in deep learning applied to skin cancer detection. arXiv 2019, arXiv:1912.03280. [Google Scholar]
  24. Ashraf, R.; Afzal, S.; Rehman, A.U.; Gul, S.; Baber, J.; Bakhtyar, M.; Mehmood, I.; Song, O.Y.; Maqsood, M. Region-of-Interest Based Transfer Learning Assisted Framework for Skin Cancer Detection. IEEE Access 2020, 8, 147858–147871. [Google Scholar] [CrossRef]
  25. Menze, B.H.; Jakab, A.; Bauer, S.; Kalpathy-Cramer, J.; Farahani, K.; Kirby, J.; Burren, Y.; Porz, N.; Slotboom, J.; Wiest, R.; et al. The multimodal brain tumor image segmentation benchmark (BRATS). IEEE Trans. Med. Imaging 2014, 34, 1993–2024. [Google Scholar] [CrossRef] [PubMed]
  26. Armato, S.G., III; McLennan, G.; Bidaut, L.; McNitt-Gray, M.F.; Meyer, C.R.; Reeves, A.P.; Zhao, B.; Aberle, D.R.; Henschke, C.I.; Hoffman, E.A.; et al. The lung image database consortium (LIDC) and image database resource initiative (IDRI): A completed reference database of lung nodules on CT scans. Med. Phys. 2011, 38, 915–931. [Google Scholar] [CrossRef]
  27. Heath, M.; Bowyer, K.; Kopans, D.; Kegelmeyer, P.; Moore, R.; Chang, K.; Munishkumaran, S. Current status of the digital database for screening mammography. In Digital mammography; Springer: Berlin/Heidelberg, Germany, 1998; pp. 457–460. [Google Scholar]
  28. Showrov, M.I.H.; Islam, M.T.; Hossain, M.D.; Ahmed, M.S. Performance Comparison of Three Classifiers for the Classification of Breast Cancer Dataset. In Proceedings of the 2019 4th International Conference on Electrical Information and Communication Technology (EICT), Khulna, Bangladesh, 20–22 December 2019; pp. 1–5. [Google Scholar]
  29. Milton, M.A.A. Automated skin lesion classification using ensemble of deep neural networks in ISIC 2018: Skin lesion analysis towards melanoma detection challenge. arXiv 2019, arXiv:1901.10802. [Google Scholar]
  30. Mendonça, T.; Ferreira, P.M.; Marques, J.S.; Marcal, A.R.; Rozeira, J. PH 2-A dermoscopic image database for research and benchmarking. In Proceedings of the 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC), Osaka, Japan, 3–7 July 2013; pp. 5437–5440. [Google Scholar]
  31. Louis, D.N.; Perry, A.; Reifenberger, G.; Von Deimling, A.; Figarella-Branger, D.; Cavenee, W.K.; Ohgaki, H.; Wiestler, O.D.; Kleihues, P.; Ellison, D.W. The 2016 World Health Organization classification of tumors of the central nervous system: A summary. Acta Neuropathol. 2016, 131, 803–820. [Google Scholar] [CrossRef] [Green Version]
  32. Tandel, G.S.; Biswas, M.; Kakde, O.G.; Tiwari, A.; Suri, H.S.; Turk, M.; Laird, J.R.; Asare, C.K.; Ankrah, A.A.; Khanna, N.; et al. A review on a deep learning perspective in brain cancer classification. Cancers 2019, 11, 111. [Google Scholar] [CrossRef] [Green Version]
  33. Ganau, L.; Paris, M.; Ligarotti, G.; Ganau, M. Management of gliomas: Overview of the latest technological advancements and related behavioral drawbacks. Behav. Neurol. 2015, 2015, 862634. [Google Scholar] [CrossRef]
  34. Jayadevappa, D.; Srinivas Kumar, S.; Murty, D. Medical image segmentation algorithms using deformable models: A review. IETE Tech. Rev. 2011, 28, 248–255. [Google Scholar] [CrossRef]
  35. Yazdani, S.; Yusof, R.; Karimian, A.; Pashna, M.; Hematian, A. Image segmentation methods and applications in MRI brain images. IETE Tech. Rev. 2015, 32, 413–427. [Google Scholar] [CrossRef]
  36. Bahadure, N.B.; Ray, A.K.; Thethi, H.P. Image analysis for MRI based brain tumor detection and feature extraction using biologically inspired BWT and SVM. Int. J. Biomed. Imaging 2017, 2017, 9749108. [Google Scholar] [CrossRef] [Green Version]
  37. Alfonse, M.; Salem, A.B.M. An automatic classification of brain tumors through MRI using support vector machine. Egy. Comp. Sci. J 2016, 40, 11–22. [Google Scholar]
  38. Wu, W.; Chen, A.Y.; Zhao, L.; Corso, J.J. Brain tumor detection and segmentation in a CRF (conditional random fields) framework with pixel-pairwise affinity and superpixel-level features. Int. J. Comput. Assist. Radiol. Surg. 2014, 9, 241–253. [Google Scholar] [CrossRef]
  39. Soltaninejad, M.; Ye, X.; Yang, G.; Allinson, N.; Lambrou, T. Brain tumor grading in different MRI protocols using SVM on statistical features. In Proceedings of the MIUA 2014 18th Annual Conference, Medical Image Understanding and Analysis, Egham, UK, 9–11 July 2014; pp. 259–264. [Google Scholar]
  40. Soltaninejad, M.; Ye, X.; Yang, G.; Allinson, N.; Lambrou, T. An image analysis approach to MRI brain tumor grading. Oncol. News 2015, 9, 204–207. [Google Scholar]
  41. Ren, T.; Wang, H.; Feng, H.; Xu, C.; Liu, G.; Ding, P. Study on the improved fuzzy clustering algorithm and its application in brain image segmentation. Appl. Soft Comput. 2019, 81, 1–9. [Google Scholar] [CrossRef]
  42. Sulaiman, S.N.; Non, N.A.; Isa, I.S.; Hamzah, N. Segmentation of brain MRI image based on clustering algorithm. In Proceedings of the 2014 IEEE Symposium on Industrial Electronics & Applications (ISIEA), Hangzhou, China, 9–11 June 2014; pp. 60–65. [Google Scholar]
  43. Li, J.; Yu, Z.L.; Gu, Z.; Liu, H.; Li, Y. MMAN: Multi-modality aggregation network for brain segmentation from MR images. Neurocomputing 2019, 358, 10–19. [Google Scholar] [CrossRef]
  44. Singh, A. Detection of brain tumor in MRI images, using combination of fuzzy c-means and SVM. In Proceedings of the 2015 2nd International Conference on Signal Processing and Integrated Networks (SPIN), New Delhi, India, 19–20 February 2015; pp. 98–102. [Google Scholar]
  45. Sujan, M.; Alam, N.; Noman, S.A.; Islam, M.J. A segmentation based automated system for brain tumor detection. Int. J. Comput. Appl. 2016, 153, 41–49. [Google Scholar] [CrossRef]
  46. Vishnuvarthanan, A.; Rajasekaran, M.P.; Govindaraj, V.; Zhang, Y.; Thiyagarajan, A. An automated hybrid approach using clustering and nature inspired optimization technique for improved tumor and tissue segmentation in magnetic resonance brain images. Applied Soft Computing 2017, 57, 399–426. [Google Scholar] [CrossRef]
  47. Gajhede, N.; Beck, O.; Purwins, H. Convolutional neural networks with batch normalization for classifying hi-hat, snare, and bass percussion sound samples. In Proceedings of the Audio Mostly (ACM), Norrköping, Sweden, 4–6 October 2016; pp. 111–115. [Google Scholar]
  48. Kaur, T.; Gandhi, T.K. Automated brain image classification based on VGG-16 and transfer learning. In Proceedings of the 2019 International Conference on Information Technology (ICIT), Bhubaneswar, India, 19–21 December 2019; pp. 94–98. [Google Scholar]
  49. Deepak, S.; Ameer, P. Brain tumor classification using deep CNN features via transfer learning. Comput. Biol. Med. 2019, 111, 103345. [Google Scholar] [CrossRef]
  50. Mittal, M.; Goyal, L.M.; Kaur, S.; Kaur, I.; Verma, A.; Hemanth, D.J. Deep learning based enhanced tumor segmentation approach for MR brain images. Appl. Soft Comput. 2019, 78, 346–354. [Google Scholar] [CrossRef]
  51. Amin, J.; Sharif, M.; Yasmin, M.; Saba, T.; Anjum, M.A.; Fernandes, S.L. A new approach for brain tumor segmentation and classification based on score level fusion using transfer learning. J. Med. Syst. 2019, 43, 1–16. [Google Scholar] [CrossRef]
  52. Rehman, M.U.; Cho, S.; Kim, J.; Chong, K.T. BrainSeg-Net: Brain Tumor MR Image Segmentation via Enhanced Encoder–Decoder Network. Diagnostics 2021, 11, 169. [Google Scholar] [CrossRef]
  53. Chandra, G.R.; Rao, K.R.H. Tumor detection in brain using genetic algorithm. Procedia Comput. Sci. 2016, 79, 449–457. [Google Scholar] [CrossRef] [Green Version]
  54. Arunachalam, M.; Royappan Savarimuthu, S. An efficient and automatic glioblastoma brain tumor detection using shift-invariant shearlet transform and neural networks. Int. J. Imaging Syst. Technol. 2017, 27, 216–226. [Google Scholar] [CrossRef]
  55. Pereira, S.; Pinto, A.; Alves, V.; Silva, C.A. Brain tumor segmentation using convolutional neural networks in MRI images. IEEE Trans. Med. Imaging 2016, 35, 1240–1251. [Google Scholar] [CrossRef] [PubMed]
  56. Perkuhn, M.; Stavrinou, P.; Thiele, F.; Shakirin, G.; Mohan, M.; Garmpis, D.; Kabbasch, C.; Borggrefe, J. Clinical evaluation of a multiparametric deep learning model for glioblastoma segmentation using heterogeneous magnetic resonance imaging data from clinical routine. Investig. Radiol. 2018, 53, 647–654. [Google Scholar] [CrossRef]
  57. Zhao, X.; Wu, Y.; Song, G.; Li, Z.; Zhang, Y.; Fan, Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med. Image Anal. 2018, 43, 98–111. [Google Scholar] [CrossRef]
  58. Vishnuvarthanan, G.; Rajasekaran, M.P.; Subbaraj, P.; Vishnuvarthanan, A. An unsupervised learning method with a clustering approach for tumor identification and tissue segmentation in magnetic resonance brain images. Appl. Soft Comput. 2016, 38, 190–212. [Google Scholar] [CrossRef]
  59. Amin, J.; Sharif, M.; Yasmin, M.; Saba, T.; Raza, M. Use of machine intelligence to conduct analysis of human brain data for detection of abnormalities in its cognitive functions. Multimed. Tools Appl. 2020, 79, 10955–10973. [Google Scholar] [CrossRef]
  60. Zhao, J.; Meng, Z.; Wei, L.; Sun, C.; Zou, Q.; Su, R. Supervised brain tumor segmentation based on gradient and context-sensitive features. Front. Neurosci. 2019, 13, 1–11. [Google Scholar] [CrossRef]
  61. Khan, M.A.; Lali, I.U.; Rehman, A.; Ishaq, M.; Sharif, M.; Saba, T.; Zahoor, S.; Akram, T. Brain tumor detection and classification: A framework of marker-based watershed algorithm and multilevel priority features selection. Microsc. Res. Tech. 2019, 82, 909–922. [Google Scholar] [CrossRef]
  62. Peng, S.; Chen, W.; Sun, J.; Liu, B. Multi-Scale 3D U-Nets: An approach to automatic segmentation of brain tumor. Int. J. Imaging Syst. Technol. 2020, 30, 5–17. [Google Scholar] [CrossRef]
  63. Li, Q.; Yu, Z.; Wang, Y.; Zheng, H. TumorGAN: A multi-modal data augmentation framework for brain tumor segmentation. Sensors 2020, 20, 4203. [Google Scholar] [CrossRef]
  64. Alqazzaz, S.; Sun, X.; Yang, X.; Nokes, L. Automated brain tumor segmentation on multi-modal MR image using SegNet. Comput. Vis. Media 2019, 5, 209–219. [Google Scholar] [CrossRef] [Green Version]
  65. Bodapati, J.D.; Shaik, N.S.; Naralasetti, V.; Mundukur, N.B. Joint training of two-channel deep neural network for brain tumor classification. Signal Image Video Process. 2021, 15, 753–760. [Google Scholar] [CrossRef]
  66. Abd El Kader, I.; Xu, G.; Shuai, Z.; Saminu, S.; Javaid, I.; Salim Ahmad, I. Differential deep convolutional neural network model for brain tumor classification. Brain Sci. 2021, 11, 352. [Google Scholar] [CrossRef]
  67. Shivaprasad, B. Bidirectional ConvLSTMXNet for Brain Tumor Segmentation of MR Images. Teh. Glas. 2021, 15, 37–42. [Google Scholar]
  68. Raja, N.; Rajinikanth, V.; Fernandes, S.L.; Satapathy, S.C. Segmentation of breast thermal images using Kapur’s entropy and hidden Markov random field. J. Med. Imaging Health Inf. 2017, 7, 1825–1829. [Google Scholar] [CrossRef]
  69. Mahmood, T.; Li, J.; Pei, Y.; Akhtar, F.; Imran, A.; Rehman, K.U. A Brief Survey on Breast Cancer Diagnostic With Deep Learning Schemes Using Multi-Image Modalities. IEEE Access 2020, 8, 165779–165809. [Google Scholar] [CrossRef]
  70. Zerouaoui, H.; Idri, A. Reviewing Machine Learning and Image Processing Based Decision-Making Systems for Breast Cancer Imaging. J. Med. Syst. 2021, 45, 1–20. [Google Scholar] [CrossRef]
  71. Nie, K.; Baltzer, P.; Preim, B.; Mistelbauer, G. Knowledge-Assisted Comparative Assessment of Breast Cancer using Dynamic Contrast-Enhanced Magnetic Resonance Imaging. Comput. Graph. Forum. 2020, 39, 13–23. [Google Scholar] [CrossRef]
  72. Makandar, A.; Halalli, B. Threshold based segmentation technique for mass detection in mammography. J. Comput. 2016, 11, 472–478. [Google Scholar] [CrossRef]
  73. Li, S.; Dong, M.; Du, G.; Mu, X. Attention dense-u-net for automatic breast mass segmentation in digital mammogram. IEEE Access 2019, 7, 59037–59047. [Google Scholar] [CrossRef]
  74. El Adoui, M.; Mahmoudi, S.A.; Larhmam, M.A.; Benjelloun, M. MRI breast tumor segmentation using different encoder and decoder CNN architectures. Computers 2019, 8, 52. [Google Scholar] [CrossRef] [Green Version]
  75. Sridhar, B. A quality representation of tumor in breast using hybrid model watershed transform and Markov random fields. In Proceedings of the 2020 International Conference on Computer Communication and Informatics (ICCCI), Coimbatore, India, 22–24 January 2020; pp. 1–5. [Google Scholar]
  76. Benzebouchi, N.E.; Azizi, N.; Ayadi, K. A computer-aided diagnosis system for breast cancer using deep convolutional neural networks. In Computational Intelligence in Data Mining; Springer: Berlin/Heidelberg, Germany, 2019; pp. 583–593. [Google Scholar]
  77. Wang, S.; Hu, J.; Zhang, Y.; Shen, J.; Dong, F.; Zhang, X.; Lu, C.; Shang, D. Presentation and survival by hormonal receptor status in metaplastic breast cancer: A propensity score-matched analysis. Breast 2021, 60, 168–176. [Google Scholar] [CrossRef]
  78. Duarte, M.A.; Pereira, W.C.; Alvarenga, A.V. Calculating texture features from mammograms and evaluating their performance in classifying clusters of microcalcifications. In Proceedings of the Mediterranean Conference on Medical and Biological Engineering and Computing, Seville, Spain, 25–28 September 2019; pp. 322–332. [Google Scholar]
  79. Byra, M.; Jarosik, P.; Szubert, A.; Galperin, M.; Ojeda-Fournier, H.; Olson, L.; O Boyle, M.; Comstock, C.; Andre, M. Breast mass segmentation in ultrasound with selective kernel U-Net convolutional neural network. Biomed. Signal Process. Control. 2020, 61, 1–10. [Google Scholar] [CrossRef]
  80. Ragab, D.A.; Sharkas, M.; Marshall, S.; Ren, J. Breast cancer detection using deep convolutional neural networks and support vector machines. PeerJ 2019, 7, 1–23. [Google Scholar] [CrossRef]
  81. Htay, T.T.; Maung, S.S. Early stage breast cancer detection system using glcm feature extraction and k-nearest neighbor (k-NN) on mammography image. In Proceedings of the 2018 18th International Symposium on Communications and Information Technologies (ISCIT), Bangkok, Thailand, 26–28 September 2018; pp. 171–175. [Google Scholar]
  82. Sun, W.; Tseng, T.L.B.; Zhang, J.; Qian, W. Enhancing deep convolutional neural network scheme for breast cancer diagnosis with unlabeled data. Comput. Med. Imaging Graph. 2017, 57, 4–9. [Google Scholar] [CrossRef] [Green Version]
  83. Zhang, Q.; Xiao, Y.; Dai, W.; Suo, J.; Wang, C.; Shi, J.; Zheng, H. Deep learning based classification of breast tumors with shear-wave elastography. Ultrasonics 2016, 72, 150–157. [Google Scholar] [CrossRef] [PubMed]
  84. Debelee, T.G.; Amirian, M.; Ibenthal, A.; Palm, G.; Schwenker, F. Classification of mammograms using convolutional neural network based feature extraction. In Proceedings of the International Conference on Information and Communication Technology for Develoment for Africa, Bahir Dar, Ethiopia, 22–24 November 2017; pp. 89–98. [Google Scholar]
  85. Han, S.; Kang, H.K.; Jeong, J.Y.; Park, M.H.; Kim, W.; Bang, W.C.; Seong, Y.K. A deep learning framework for supporting the classification of breast lesions in ultrasound images. Phys. Med. Biol. 2017, 62, 7714–7728. [Google Scholar] [CrossRef]
  86. Khamparia, A.; Bharati, S.; Podder, P.; Gupta, D.; Khanna, A.; Phung, T.K.; Thanh, D.N. Diagnosis of breast cancer based on modern mammography using hybrid transfer learning. Multidimens. Syst. Signal Process. 2021, 32, 747–765. [Google Scholar] [CrossRef]
  87. Jahangeer, G.S.B.; Rajkumar, T.D. Early detection of breast cancer using hybrid of series network and VGG-16. Multimed. Tools Appl. 2021, 80, 7853–7886. [Google Scholar] [CrossRef]
  88. Dora, L.; Agrawal, S.; Panda, R.; Abraham, A. Optimal breast cancer classification using Gauss–Newton representation based algorithm. Expert Syst. Appl. 2017, 85, 134–145. [Google Scholar] [CrossRef]
  89. Setio, A.A.A.; Ciompi, F.; Litjens, G.; Gerke, P.; Jacobs, C.; Van Riel, S.J.; Wille, M.M.W.; Naqibullah, M.; Sánchez, C.I.; Van Ginneken, B. Pulmonary nodule detection in CT images: False positive reduction using multi-view convolutional networks. IEEE Trans. Med. Imaging 2016, 35, 1160–1169. [Google Scholar] [CrossRef] [PubMed]
  90. Naqi, S.; Sharif, M.; Yasmin, M.; Fernandes, S.L. Lung nodule detection using polygon approximation and hybrid features from CT images. Curr. Med. Imaging 2018, 14, 108–117. [Google Scholar] [CrossRef]
  91. Paing, M.P.; Hamamoto, K.; Tungjitkusolmun, S.; Visitsattapongse, S.; Pintavirooj, C. Automatic detection of pulmonary nodules using three-dimensional chain coding and optimized random forest. Appl. Sci. 2020, 10, 2346. [Google Scholar] [CrossRef] [Green Version]
  92. Perez, G.; Arbelaez, P. Automated lung cancer diagnosis using three-dimensional convolutional neural networks. Med. Biol. Eng. Comput. 2020, 58, 1803–1815. [Google Scholar] [CrossRef]
  93. Hu, Z.; Tang, J.; Wang, Z.; Zhang, K.; Zhang, L.; Sun, Q. Deep learning for image-based cancer detection and diagnosis- A survey. Pattern Recognit. 2018, 83, 134–149. [Google Scholar] [CrossRef]
  94. Ali, S.; Shaukat, Z.; Azeem, M.; Sakhawat, Z.; Mahmood, T.; ur Rehman, K. An efficient and improved scheme for handwritten digit recognition based on convolutional neural network. SN Appl. Sci. 2019, 1, 1–9. [Google Scholar] [CrossRef] [Green Version]
  95. Ali, S.; Li, J.; Pei, Y.; Aslam, M.S.; Shaukat, Z.; Azeem, M. An Effective and Improved CNN-ELM Classifier for Handwritten Digits Recognition and Classification. Symmetry 2020, 12, 1742. [Google Scholar] [CrossRef]
  96. Suzuki, K.; Otsuka, Y.; Nomura, Y.; Kumamaru, K.K.; Kuwatsuru, R.; Aoki, S. Development and Validation of a Modified Three-Dimensional U-Net Deep-Learning Model for Automated Detection of Lung Nodules on Chest CT Images from the Lung Image Database Consortium and Japanese Datasets. Acad. Radiol. 2020. In Press. [Google Scholar] [CrossRef]
  97. Zhao, C.; Han, J.; Jia, Y.; Gou, F. Lung nodule detection via 3D U-Net and contextual convolutional neural network. In Proceedings of the 2018 International Conference on Networking and Network Applications (NaNA), Xi’an, China, 12–18 October 2018; pp. 356–361. [Google Scholar]
  98. Xie, H.; Yang, D.; Sun, N.; Chen, Z.; Zhang, Y. Automated pulmonary nodule detection in CT images using deep convolutional neural networks. Pattern Recognition 2019, 85, 109–119. [Google Scholar] [CrossRef]
  99. Kasinathan, G.; Jayakumar, S.; Gandomi, A.H.; Ramachandran, M.; Fong, S.J.; Patan, R. Automated 3-D lung tumor detection and classification by an active contour model and CNN classifier. Expert Syst. Appl. 2019, 134, 112–119. [Google Scholar] [CrossRef]
  100. Asuntha, A.; Brindha, A.; Indirani, S.; Srinivasan, A. Lung cancer detection using SVM algorithm and optimization techniques. J. Chem. Pharm. Sci 2016, 9, 3198–3203. [Google Scholar]
  101. Anirudh, R.; Thiagarajan, J.J.; Bremer, T.; Kim, H. Lung nodule detection using 3D convolutional neural networks trained on weakly labeled data. In Proceedings of the Medical Imaging 2016: Computer-Aided Diagnosis, International Society for Optics and Photonics, San Diego, CA, USA, 27 February–3 March 2016; Volume 9785, pp. 1–6. [Google Scholar]
  102. Chon, A.; Balachandar, N.; Lu, P. Deep Convolutional Neural Networks for Lung Cancer Detection; Standford University: Standford, CA, USA, 2017; pp. 1–9. [Google Scholar]
  103. Nibali, A.; He, Z.; Wollersheim, D. Pulmonary nodule classification with deep residual networks. Int. J. Comput. Assist. Radiol. Surg. 2017, 12, 1799–1808. [Google Scholar] [CrossRef]
  104. Alakwaa, W.; Nassef, M.; Badr, A. Lung cancer detection and classification with 3D convolutional neural network (3D-CNN). Lung Cancer 2017, 8, 409–417. [Google Scholar] [CrossRef] [Green Version]
  105. Naqi, S.M.; Sharif, M.; Jaffar, A. Lung nodule detection and classification based on geometric fit in parametric form and deep learning. Neural Comput. Appl. 2020, 32, 4629–4647. [Google Scholar] [CrossRef]
  106. Ali, I.; Hart, G.R.; Gunabushanam, G.; Liang, Y.; Muhammad, W.; Nartowt, B.; Kane, M.; Ma, X.; Deng, J. Lung nodule detection via deep reinforcement learning. Frontiers in oncology 2018, 8, 1–7. [Google Scholar] [CrossRef] [Green Version]
  107. Naqi, S.M.; Sharif, M.; Lali, I.U. A 3D nodule candidate detection method supported by hybrid features to reduce false positives in lung nodule detection. Multimed. Tools Appl. 2019, 78, 26287–26311. [Google Scholar] [CrossRef]
  108. El-Regaily, S.A.; Salem, M.A.M.; Aziz, M.H.A.; Roushdy, M.I. Multi-view Convolutional Neural Network for lung nodule false positive reduction. Expert Syst. Appl. 2020, 162, 1–17. [Google Scholar] [CrossRef]
  109. Onishi, Y.; Teramoto, A.; Tsujimoto, M.; Tsukamoto, T.; Saito, K.; Toyama, H.; Imaizumi, K.; Fujita, H. Multiplanar analysis for pulmonary nodule classification in CT images using deep convolutional neural network and generative adversarial networks. Int. J. Comput. Assist. Radiol. Surg. 2020, 15, 173–178. [Google Scholar] [CrossRef]
  110. Wang, Q.; Shen, F.; Shen, L.; Huang, J.; Sheng, W. Lung nodule detection in CT images using a raw patch-based convolutional neural network. J. Digit. Imaging 2019, 32, 971–979. [Google Scholar] [CrossRef]
  111. Chen, L.; Gu, D.; Chen, Y.; Shao, Y.; Cao, X.; Liu, G.; Gao, Y.; Wang, Q.; Shen, D. An artificial-intelligence lung imaging analysis system (ALIAS) for population-based nodule computing in CT scans. Comput. Med. Imaging Graph. 2021, 89, 1–9. [Google Scholar] [CrossRef] [PubMed]
  112. Chaunzwa, T.L.; Hosny, A.; Xu, Y.; Shafer, A.; Diao, N.; Lanuti, M.; Christiani, D.C.; Mak, R.H.; Aerts, H.J. Deep learning classification of lung cancer histology using CT images. Sci. Rep. 2021, 11, 1–12. [Google Scholar] [CrossRef] [PubMed]
  113. Masood, A.; Sheng, B.; Yang, P.; Li, P.; Li, H.; Kim, J.; Feng, D.D. Automated decision support system for lung cancer detection and classification via enhanced RFCN with multilayer fusion RPN. IEEE Trans. Ind. Inform. 2020, 16, 7791–7801. [Google Scholar] [CrossRef]
  114. Ding, J.; Li, A.; Hu, Z.; Wang, L. Accurate pulmonary nodule detection in computed tomography images using deep convolutional neural networks. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Cambridge, UK, 27 September–1 October 2017; pp. 559–567. [Google Scholar]
  115. Yu, L.; Chen, H.; Dou, Q.; Qin, J.; Heng, P.A. Automated melanoma recognition in dermoscopy images via very deep residual networks. IEEE Trans. Med. Imaging 2016, 36, 994–1004. [Google Scholar] [CrossRef] [PubMed]
  116. Adeyinka, A.A.; Viriri, S. Skin lesion images segmentation: A survey of the state-of-the-art. In Proceedings of the International Conference on Mining Intelligence and Knowledge Exploration, Cluj-Napoca, Romania, 20–22 December 2018; pp. 321–330. [Google Scholar]
  117. Naylor, P.; Laé, M.; Reyal, F.; Walter, T. Nuclei segmentation in histopathology images using deep neural networks. In Proceedings of the 2017 IEEE 14th International Symposium on Biomedical Imaging (ISBI 2017), Melbourne, Australia, 18–21 April 2017; pp. 933–936. [Google Scholar]
  118. Ratul, M.A.R.; Mozaffari, M.H.; Lee, W.; Parimbelli, E. Skin lesions classification using deep learning based on dilated convolution. BioRxiv 2020. [Google Scholar] [CrossRef] [Green Version]
  119. Vesal, S.; Ravikumar, N.; Maier, A. SkinNet: A deep learning framework for skin lesion segmentation. In Proceedings of the 2018 IEEE Nuclear Science Symposium and Medical Imaging Conference Proceedings (NSS/MIC), Sydney, NSW, Australia, 10–17 November 2018; pp. 1–3. [Google Scholar]
  120. Esteva, A.; Kuprel, B.; Novoa, R.A.; Ko, J.; Swetter, S.M.; Blau, H.M.; Thrun, S. Dermatologist-level classification of skin cancer with deep neural networks. Nature 2017, 542, 115–118. [Google Scholar] [CrossRef]
  121. Gessert, N.; Nielsen, M.; Shaikh, M.; Werner, R.; Schlaefer, A. Skin Lesion Classification Using Ensembles of Multi-Resolution EfficientNets with Meta Data. In MethodsX; 2020; pp. 1–10. Available online: https://arxiv.org/pdf/1910.03910.pdf (accessed on 30 June 2021).
  122. Hameed, N.; Shabut, A.; Hameed, F.; Cirstea, S.; Harriet, S.; Hossain, A. Mobile based Skin Lesions Classification Using Convolution Neural Network. Ann. Emerg. Technol. Comput. 2020, 4, 26–37. [Google Scholar] [CrossRef]
  123. Li, Y.; Shen, L. Skin lesion analysis towards melanoma detection using deep learning network. Sensors 2018, 18, 556. [Google Scholar] [CrossRef] [Green Version]
  124. Dash, M.; Londhe, N.D.; Ghosh, S.; Semwal, A.; Sonawane, R.S. PsLSNet: Automated psoriasis skin lesion segmentation using modified U-Net-based fully convolutional network. Biomed. Signal Process. Control. 2019, 52, 226–237. [Google Scholar] [CrossRef]
  125. Al-Masni, M.A.; Al-Antari, M.A.; Choi, M.T.; Han, S.M.; Kim, T.S. Skin lesion segmentation in dermoscopy images via deep full resolution convolutional networks. Comput. Methods Programs Biomed. 2018, 162, 221–231. [Google Scholar] [CrossRef]
  126. Suganya, R. An automated computer aided diagnosis of skin lesions detection and classification for dermoscopy images. In Proceedings of the 2016 International Conference on Recent Trends in Information Technology (ICRTIT), Chennai, India, 8–9 April 2016; pp. 1–5. [Google Scholar]
  127. Shoieb, D.A.; Youssef, S.M.; Aly, W.M. Computer-aided model for skin diagnosis using deep learning. J. Image Graph. 2016, 4, 122–129. [Google Scholar] [CrossRef] [Green Version]
  128. Jafari, M.H.; Karimi, N.; Nasr-Esfahani, E.; Samavi, S.; Soroushmehr, S.M.R.; Ward, K.; Najarian, K. Skin lesion segmentation in clinical images using deep learning. In Proceedings of the 2016 23rd International Conference on Pattern Recognition (ICPR), Cancun, Mexico, 4–8 December 2016; pp. 337–342. [Google Scholar]
  129. Bakheet, S. An svm framework for malignant melanoma detection based on optimized hog features. Computation 2017, 5, 4. [Google Scholar] [CrossRef] [Green Version]
  130. Qian, C.; Liu, T.; Jiang, H.; Wang, Z.; Wang, P.; Guan, M.; Sun, B. A detection and segmentation architecture for skin lesion segmentation on dermoscopy images. arXiv 2018, arXiv:1809.03917. [Google Scholar]
  131. Zhuang, J.; Li, W.; Manivannan, S.; Wang, R.; Zhang, J.J.G.; Pan, J.; Jiang, G.; Yin, Z. Skin lesion analysis towards melanoma detection using deep neural network ensemble. ISIC Challenge 2018 2018, 2, 1–6. [Google Scholar]
  132. Lee, Y.C.; Jung, S.H.; Won, H.H. WonDerM: Skin lesion classification with fine-tuned neural networks. arXiv 2018, arXiv:1808.03426. [Google Scholar]
  133. Zhang, G.; Shen, X.; Chen, S.; Liang, L.; Luo, Y.; Yu, J.; Lu, J. DSM: A deep supervised multi-scale network learning for skin cancer segmentation. IEEE Access 2019, 7, 140936–140945. [Google Scholar] [CrossRef]
  134. Seeja, R.; Suresh, A. Deep learning based skin lesion segmentation and classification of melanoma using support vector machine (SVM). Asian Pac. J. Cancer Prev. 2019, 20, 1555–1561. [Google Scholar]
  135. Jianu, S.R.S.; Ichim, L.; Popescu, D. Automatic diagnosis of skin cancer using neural networks. In Proceedings of the 2019 11th International Symposium on Advanced Topics in Electrical Engineering (ATEE), Bucharest, Romania, 28–30 March 2019; pp. 1–4. [Google Scholar]
  136. Srividhya, V.; Sujatha, K.; Ponmagal, R.; Durgadevi, G.; Madheshwaran, L. Vision based Detection and Categorization of Skin lesions using Deep Learning Neural networks. Procedia Comput. Sci. 2020, 171, 1726–1735. [Google Scholar] [CrossRef]
  137. Thomas, S.M.; Lefevre, J.G.; Baxter, G.; Hamilton, N.A. Interpretable deep learning systems for multi-class segmentation and classification of non-melanoma skin cancer. Med. Image Anal. 2021, 68, 1–26. [Google Scholar] [CrossRef]
  138. Afza, F.; Sharif, M.; Mittal, M.; Khan, M.A.; Hemanth, D.J. A hierarchical three-step superpixels and deep learning framework for skin lesion classification. Methods 2021, in press. [Google Scholar] [CrossRef]
  139. Adegun, A.; Viriri, S. Deep learning techniques for skin lesion analysis and melanoma cancer detection: A survey of state-of-the-art. Artif. Intell. Rev. 2021, 54, 811–841. [Google Scholar] [CrossRef]
  140. LeCun, Y.; Boser, B.; Denker, J.S.; Henderson, D.; Howard, R.E.; Hubbard, W.; Jackel, L.D. Backpropagation applied to handwritten zip code recognition. Neural Comput. 1989, 1, 541–551. [Google Scholar] [CrossRef]
  141. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  142. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef] [Green Version]
  143. Murphy, A.; Skalski, M.; Gaillard, F. The utilisation of convolutional neural networks in detecting pulmonary nodules: A review. Br. J. Radiol. 2018, 91, 1–6. [Google Scholar] [CrossRef] [PubMed]
  144. Khan, A.; Sohail, A.; Zahoora, U.; Qureshi, A.S. A survey of the recent architectures of deep convolutional neural networks. Artificial Intelligence Review 2020, 53, 5455–5516. [Google Scholar] [CrossRef] [Green Version]
  145. Suzuki, K. Overview of deep learning in medical imaging. Radiol. Phys. Technol. 2017, 10, 257–273. [Google Scholar] [CrossRef]
  146. Milletari, F.; Ahmadi, S.A.; Kroll, C.; Plate, A.; Rozanski, V.; Maiostre, J.; Levin, J.; Dietrich, O.; Ertl-Wagner, B.; Bötzel, K.; et al. Hough-CNN: Deep learning for segmentation of deep brain regions in MRI and ultrasound. Comput. Vis. Image Underst. 2017, 164, 92–102. [Google Scholar] [CrossRef] [Green Version]
  147. Golan, R.; Jacob, C.; Denzinger, J. Lung nodule detection in CT images using deep convolutional neural networks. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 243–250. [Google Scholar]
  148. Yosinski, J.; Clune, J.; Bengio, Y.; Lipson, H. How transferable are features in deep neural networks? arXiv 2014, arXiv:1411.1792. [Google Scholar]
  149. Kamnitsas, K.; Ledig, C.; Newcombe, V.F.; Simpson, J.P.; Kane, A.D.; Menon, D.K.; Rueckert, D.; Glocker, B. Efficient multi-scale 3D CNN with fully connected CRF for accurate brain lesion segmentation. Med. Image Anal. 2017, 36, 61–78. [Google Scholar] [CrossRef] [PubMed]
  150. Dou, Q.; Yu, L.; Chen, H.; Jin, Y.; Yang, X.; Qin, J.; Heng, P.A. 3D deeply supervised network for automated segmentation of volumetric medical images. Med. Image Anal. 2017, 41, 40–54. [Google Scholar] [CrossRef]
  151. Kronman, A.; Joskowicz, L. A geometric method for the detection and correction of segmentation leaks of anatomical structures in volumetric medical images. Int. J. Comput. Assist. Radiol. Surg. 2016, 11, 369–380. [Google Scholar] [CrossRef]
  152. Zeng, G.; Zheng, G. Multi-stream 3D FCN with multi-scale deep supervision for multi-modality isointense infant brain MR image segmentation. In Proceedings of the 2018 IEEE 15th International Symposium on Biomedical Imaging (ISBI 2018), Washington, DC, USA, 4–8 April 2018; pp. 136–140. [Google Scholar]
  153. Zhang, W.; Li, R.; Deng, H.; Wang, L.; Lin, W.; Ji, S.; Shen, D. Deep convolutional neural networks for multi-modality isointense infant brain image segmentation. NeuroImage 2015, 108, 214–224. [Google Scholar] [CrossRef] [Green Version]
  154. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  155. Yi, X.; Walia, E.; Babyn, P. Generative adversarial network in medical imaging: A review. Med. Image Anal. 2019, 58, 1–24. [Google Scholar] [CrossRef] [Green Version]
  156. Hussain, J. Deep Learning Black Box Problem. Master’s Thesis, Uppsala University, Uppsala, Sweden, 2019. [Google Scholar]
  157. Monkam, P.; Qi, S.; Ma, H.; Gao, W.; Yao, Y.; Qian, W. Detection and classification of pulmonary nodules using convolutional neural networks: A survey. IEEE Access 2019, 7, 78075–78091. [Google Scholar] [CrossRef]
  158. Zhao, X.; Qi, S.; Zhang, B.; Ma, H.; Qian, W.; Yao, Y.; Sun, J. Deep CNN models for pulmonary nodule classification: Model modification, model integration, and transfer learning. J. X-ray Sci. Technol. 2019, 27, 615–629. [Google Scholar] [CrossRef]
  159. Karim, M.R.; Beyan, O.; Zappa, A.; Costa, I.G.; Rebholz-Schuhmann, D.; Cochez, M.; Decker, S. Deep learning-based clustering approaches for bioinformatics. Briefings Bioinform. 2021, 22, 393–415. [Google Scholar] [CrossRef] [Green Version]
Figure 1. A flowchart of a standard CAD system for diagnosing multi-organ of human body cancer. This flowchart illustrates all steps of a CAD system.
Figure 1. A flowchart of a standard CAD system for diagnosing multi-organ of human body cancer. This flowchart illustrates all steps of a CAD system.
Cancers 13 05546 g001
Figure 2. Overview of 3D CNN architecture for brain tumor prediction.
Figure 2. Overview of 3D CNN architecture for brain tumor prediction.
Cancers 13 05546 g002
Figure 3. DL-based framework for breast tumor classification and segmentation.
Figure 3. DL-based framework for breast tumor classification and segmentation.
Cancers 13 05546 g003
Figure 4. CAD system for lung cancer detection.
Figure 4. CAD system for lung cancer detection.
Cancers 13 05546 g004
Figure 5. CNN-based SqueezeNet architecture for skin cancer detection.
Figure 5. CNN-based SqueezeNet architecture for skin cancer detection.
Cancers 13 05546 g005
Figure 6. Variation in skin lesion [139].
Figure 6. Variation in skin lesion [139].
Cancers 13 05546 g006
Figure 7. Distribution of publications between different DL-based, ML-based and other classifiers for brain, breast, lung and skin cancer detection.
Figure 7. Distribution of publications between different DL-based, ML-based and other classifiers for brain, breast, lung and skin cancer detection.
Cancers 13 05546 g007
Table 1. Commonly utilized publicly available databases with different modalities for brain, breast, lung and skin cancer detection.
Table 1. Commonly utilized publicly available databases with different modalities for brain, breast, lung and skin cancer detection.
DatabaseModalityImages/PatientsLink to the Source
BRATS2012MRI-scans45 Patientshttps://www.smir.ch/BRATS/Start2012 accessed on 30 June 2021
BRATS2015MRI-scans274 Patientshttps://www.smir.ch/BRATS/Start2015 accessed on 30 June 2021
BRATS2017MRI-scans285 Patientshttps://www.med.upenn.edu/sbia/brats2017/registration.html accessed on 30 June 2021
BrainWebMRI-scans20http://www.bic.mni.mcgill.ca/brainweb/ accessed on 30 June 2021
HarvardMRI-scans13,000 brain MRIshttp://www.med.harvard.edu/aanlib/ accessed on 30 June 2021
Mini-MIASMammograms322http://peipa.essex.ac.uk/info/mias.html accessed on 30 June 2021
DDSMMammograms2620 caseshttp://www.eng.usf.edu/cvprg/Mammography/Database.html accessed on 30 June 2021
WBCDBiopsy683 Patientshttps://archive.ics.uci.edu/ml/datasets/Breast+Cancer+Wisconsin+(Diagnostic) accessed on 30 June 2021
LIDC/IDRICT-scans1018 caseshttps://wiki.cancerimagingarchive.net/display/Public/LIDC-IDRI accessed on 30 June 2021
ISIC-2016Dermoscopic1279https://challenge.isic-archive.com/data accessed on 30 June 2021
ISIC-2017Dermoscopic2750https://challenge.isic-archive.com/data accessed on 30 June 2021
HAM10000Dermoscopic10,015https://challenge.isic-archive.com/data accessed on 30 June 2021
PH2Dermoscopic200https://www.fc.up.pt/addi/ph2%20database.html accessed on 30 June 2021
SD-198Clinical6584http://xiaopingwu.cn/assets/projects/sd-198/ accessed on 30 June 2021
SD-260Clinical20,660http://xiaopingwu.cn/assets/projects/sd-198/ accessed on 30 June 2021
Table 2. Methodologies of brain tumor cancer diagnosis.
Table 2. Methodologies of brain tumor cancer diagnosis.
MethodsTask PerformedUser InterventionDatasetEvaluation Matrix (%)YearRef.
PCA+DNNSegmentationFully-automaticHarvardSN = 0.97, Acc = 96.9, AUC = 0.982017[4]
GLCM + Logistic regression (LR)SegmentationFully-automaticBrats15SN = 0.88, SP = 0.90, Acc = 0.89, AUC = 0.882017[14]
DWT + Genetic algorithmsDetectionSemi-automaticPrivateAcc = 95.62016[53]
BWT + SVMDetectionFully-automaticBrainWebSN = 97.7, SP = 94.2, Acc = 96.52017[36]
GLCM + Gabor + DWT + K-meansDetectionFully-automaticBrats15SN = 89.7, SP = 99.9, Acc = 99.82017[54]
CNNSegmentationFully-automaticBrats13WT = 0.78, TC = 0.65, ET = 0.752016[55]
DeepMedicSegmentationFully-automaticPublicWT = 0.86, ET = 0.78, TC = 0.622018[56]
Integration of FCNNs and CRFsSegmentationFully-automaticBrats15WT = 0.84, TC = 0.67, ET = 0.622018[57]
SOM + FKMSegmentationFully-automaticHarvardAcc = 96.1, SN = 87.12016[58]
CNNSegmentationFully-automaticTCGA-GBMAcc = 90.92019[16]
KNNSegmentationFully-automaticBrats15SN = 100, SP = 87.7, Acc = 96.6, AUC = 0.982020[59]
Random forestSegmentationFully-automaticBrats15SN = 0.84, SP = 0.71, Acc = 0.872019[60]
Watershed, Gamma Contrast stretchingClassification-HarvardAcc = 0.982019[61]
Multi-Scale 3D U-NetsSegmentationFully-automaticBrats15SN = 0.86, SP = 0.86, Acc = 0.852020[62]
TumorGANSegmentationFully-automaticBrats17WT = 0.85, TC = 0.792020[63]
SegNetSegmentationFully-automaticBrats17WT = 0.85, TC = 0.81, ET = 0.792019[64]
Two-Channel DNNClassificationFully-automaticBrats18Acc = 93.692021[65]
DCNNClassificationFully-automaticPrivateAcc =  99.252021[66]
Convolutional LSTM XNetSegmentationFully-automaticBrats19SN =  0.91, SP =  0.98, Acc =  0.992021[67]
BrainSeg-NetSegmentationFully-automaticBrats18WT = 0.89, TC = 0.82, ET = 0.772021[52]
Table 3. State-of-the-art CAD systems of breast cancer diagnosis.
Table 3. State-of-the-art CAD systems of breast cancer diagnosis.
MethodsTask PerformedUser InterventionDatasetEvaluation Matrix (%)YearRef.
Morphological thresholdMass detectionAutomaticMini-MIASAcc = 94.542016[72]
SSL scheme using CNNMass detectionAutomaticPrivateAcc = 0.822017[82]
DLClassificationAutomaticPrivateAcc = 93.4, SN = 88.6, SP = 97.12016[83]
CNNClassificationAutomaticDDSMAcc = 98.902018[84]
DCNNLesions classificationAutomatic-Acc = 90, SN = 90, SP = 962017[85]
Attention Dense-U-NetSegmentationAutomaticDDSMAcc = 78.3, SN = 77.8, SP = 84.62019[73]
SegNet and U-NetTumor SegmentationSami-automaticPrivate instituteAcc = 68.88, 76.142019[80]
DCNN-SVM-AlexNetCancer detectionSami-automaticCBIS-DDSMAcc = 87.22019[81]
CNN based selective kernel U-NetSegmentationAutomaticMedical centersDice score = 0.8262020[79]
OPTICS clusteringLesion classificationAutomaticDCE-MRIAcc = 71.42020[71]
Hybrid transfer learningCancer detectionAutomaticDDSMMVGG + ImageNet =  94.3, MVGG =  89.82021[86]
Hybrid VGG-16 and series network, GDDTClassificationAutomatic-VGG-16 = 96.45, GDDT = 95.152021[87]
GNRBABreast classificationAutomaticWDBCAcc = 0.982017[88]
Table 4. State-of-the-art CAD systems for lung cancer diagnosis.
Table 4. State-of-the-art CAD systems for lung cancer diagnosis.
MethodsTask PerformedDatasetEvaluation Matrix (%)YearRef.
SVM algorithmSegmentationPrivateAcc = 89.52016[100]
3D CNN trained on weakly labeled dataNodule DetectionSPIE-LUNGxSN = 802016[101]
DCNNLung cancer detectionKaggle, LUNA16Acc = 0.75, SN = 0.77, SP = 0.742017[102]
Deep residual networksNodule classificationLIDC/IDRIAcc = 89.9, SN = 91, SP = 88.62017[103]
3D-CNNDetection and ClassificationBowl 2017Acc = 86.62017[104]
Polygon approximation with SVMNodule detectionLIDCAcc = 98.8, SN = 97.7, SP = 96.22018[90]
Deep residual networksNodule classificationLIDC-IDRIAcc = 0.89, SN = 0.91, SP = 0.882017[103]
Deep learningNodule detectionLIDC-IDRIAcc = 0.96, SN = 0.95, SP = 0.972020[105]
Deep reinforcement learningNodule detectionLIDC-IDRIAcc = 0.64, SN = 0.58, SP = 0.552018[106]
3D nodule candidateNodule detectionLIDCAcc = 0.99, SN = 0.98, SP = 0.982019[107]
Optimized Random ForestAutomatic detectionLIDC-IDRIAcc = 93.1, SN = 94.8, SP = 91.3, FP = 0.0862020[91]
CNNSegments nodulesLIDCAcc = 89.8, SN = 85.2, SP = 90.62020[108]
2D DCNNNodule detectionLUNA16SN = 86.42, FP = 73.42019[98]
Generative adversarial networks with DCNNNodule classificationPrivateSN = 93.9, SP = 77.82020[109]
Patch-Based CNNNodule detectionLIDC-IDRISN = 92.82019[110]
SVMDetection and segmentationPrivateSN = 90.6, SP = 73.62021[111]
VGG-16 based CNNClassifcationMassachusetts General Hospital (MGH)Acc = 68.6, SN = 37.5, SP = 82.9, AUC = 0.702021[112]
Table 5. State-of-the-art CAD systems of skin cancer/lesion diagnosis.
Table 5. State-of-the-art CAD systems of skin cancer/lesion diagnosis.
MethodsTask PerformedDatasetEvaluation Matrix (%)YearRef.
K-means clustering and SVMSkin Lesions DetectionDermwebAcc = 95.4, SN = 96.8, SP = 89.32016[126]
CNN and SVMMelanoma classificationDERMISAcc = 93.7, SN = 87.5, SP = 1002016[127]
CNNMelanoma lesion segmentationDermquestAcc = 98.5, SN = 95.0, SP = 98.92016[128]
SVM FrameworkMelanoma DetectionPublicAcc = 97.32, SN = 98.21, SP = 96.432017[129]
CNNsClassificationClinical ImagesAcc = 72.02017[120]
Encoder-Decoder with DeepLab and PSPNetSkin lesion segmentationISIC 2018Acc = 94.2, SN = 90.6, SP = 96.3, Dice = 89.82018[130]
Ensemble ClassifiersClassificationISIC 2018Acc = 97.4, SN = 74.7, SP = 95.1, Dice = 97.42018[131]
Fine-tuned neural neworksClassificationISIC 2018Acc = 97.4, SN = 75.7, SP = 95.9, Dice = 97.22018[132]
Deeep Supervised Multi-Scale NetworkSkin Cancer SegmentationISBI 2017 and PH2Acc = 94.3, SN = 85.9, Dice = 87.52019[133]
SVMSkin lesion classificationISBI 2016Dice = 77.5, Acc = 85.12019[134]
Neural NetworksMelanoma detectionPH2Acc = 0.81, SN = 0.72, SP = 0.892019[135]
Full resolution convolutional network (FRCN)SegmentationISBI 2017 and PH2Acc = 942018[125]
CNNDetection and CategorizationDermISAcc = 95, SN = 93.32020[136]
Deep LearningSegmentation and classificationMyLab PathologyAcc = 97.92021[137]
ResNet-50 based CNN and Naive Bayes classifierSkin lesion classificationPh2, ISBI2016, and HAM1000Acc  =  95.40, 91.1, 85.502021[138]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ali, S.; Li, J.; Pei, Y.; Khurram, R.; Rehman, K.u.; Rasool, A.B. State-of-the-Art Challenges and Perspectives in Multi-Organ Cancer Diagnosis via Deep Learning-Based Methods. Cancers 2021, 13, 5546. https://doi.org/10.3390/cancers13215546

AMA Style

Ali S, Li J, Pei Y, Khurram R, Rehman Ku, Rasool AB. State-of-the-Art Challenges and Perspectives in Multi-Organ Cancer Diagnosis via Deep Learning-Based Methods. Cancers. 2021; 13(21):5546. https://doi.org/10.3390/cancers13215546

Chicago/Turabian Style

Ali, Saqib, Jianqiang Li, Yan Pei, Rooha Khurram, Khalil ur Rehman, and Abdul Basit Rasool. 2021. "State-of-the-Art Challenges and Perspectives in Multi-Organ Cancer Diagnosis via Deep Learning-Based Methods" Cancers 13, no. 21: 5546. https://doi.org/10.3390/cancers13215546

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop