Next Article in Journal
“Listen to Your Immune System When It’s Calling for You”: Monitoring Autoimmune Diseases Using the iShU App
Next Article in Special Issue
IoMT-Based Osteosarcoma Cancer Detection in Histopathology Images Using Transfer Learning Empowered with Blockchain, Fog Computing, and Edge Computing
Previous Article in Journal
Designing an AI-Based Virtual Try-On Web Application
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Histopathologic Oral Cancer Prediction Using Oral Squamous Cell Carcinoma Biopsy Empowered with Transfer Learning

by
Atta-ur Rahman
1,†,
Abdullah Alqahtani
2,
Nahier Aldhafferi
2,
Muhammad Umar Nasir
3,†,
Muhammad Farhan Khan
4,†,
Muhammad Adnan Khan
5,* and
Amir Mosavi
6,7,8
1
Department of Computer Science (CS), College of Computer Science and Information Technology (CCSIT), Imam Abdulrahman Bin Faisal University (IAU), P.O. Box 1982, Dammam 31441, Saudi Arabia
2
Department of Computer Information Systems (CIS), College of Computer Science and Information Technology (CCSIT), Imam Abdulrahman Bin Faisal University (IAU), P.O. Box 1982, Dammam 31441, Saudi Arabia
3
Faculty of Computing, Riphah School of Computing and Innovation, Riphah International University, Lahore Campus, Lahore 54000, Pakistan
4
Department of Forensic Sciences, University of Health Sciences, Lahore 54000, Pakistan
5
Department of Software, Gachon University, Seongnam 13120, Korea
6
John von Neumann Faculty of Informatics, Obuda University, 1034 Budapest, Hungary
7
Institute of Information Engineering, Automation and Mathematics, Slovak University of Technology in Bratislava, 81107 Bratislava, Slovakia
8
Faculty of Civil Engineering, TU-Dresden, 01062 Dresden, Germany
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2022, 22(10), 3833; https://doi.org/10.3390/s22103833
Submission received: 3 April 2022 / Revised: 6 May 2022 / Accepted: 17 May 2022 / Published: 18 May 2022
(This article belongs to the Special Issue Machine Learning for Medical Data Analysis)

Abstract

:
Oral cancer is a dangerous and extensive cancer with a high death ratio. Oral cancer is the most usual cancer in the world, with more than 300,335 deaths every year. The cancerous tumor appears in the neck, oral glands, face, and mouth. To overcome this dangerous cancer, there are many ways to detect like a biopsy, in which small chunks of tissues are taken from the mouth and tested under a secure and hygienic microscope. However, microscope results of tissues to detect oral cancer are not up to the mark, a microscope cannot easily identify the cancerous cells and normal cells. Detection of cancerous cells using microscopic biopsy images helps in allaying and predicting the issues and gives better results if biologically approaches apply accurately for the prediction of cancerous cells, but during the physical examinations microscopic biopsy images for cancer detection there are major chances for human error and mistake. So, with the development of technology deep learning algorithms plays a major role in medical image diagnosing. Deep learning algorithms are efficiently developed to predict breast cancer, oral cancer, lung cancer, or any other type of medical image. In this study, the proposed model of transfer learning model using AlexNet in the convolutional neural network to extract rank features from oral squamous cell carcinoma (OSCC) biopsy images to train the model. Simulation results have shown that the proposed model achieved higher classification accuracy 97.66% and 90.06% of training and testing, respectively.

1. Introduction

Oral cancer is one of the most frequent deadly diseases and has long been a serious public health concern across the world. Oral cancer is a subset of head and neck malignancies, with 475,000 new cases diagnosed each year globally. Early-stage sickness has a survival rate of about 80%, whereas late-stage sickness has a survival rate of less than 20% [1,2]. Squamous cell carcinoma of the oral cavity is the most prevalent form of oral cancer, accounting for more than 85% of cases. Although early detection of oral cancer is critical, most patients are identified at the last stage of the illness, resulting in a dismal prognosis.
Due to the clinical appearance of oral cancer being insufficient to identify the dysplastic state, analysis, or degree, therapy selection based on the clinical appearance of the illness is insufficient.
Oral cancer is linked to a variety of risk factors, and the post-treatment survival rate is similarly unexpected [3,4]. Biological models, as well as medical forms of related and lesion-free tumor models, may be recognized in many regions of the body using appearance models and stereotypes without labeling. Potentially cancerous lesions like leukoplakia, erythroplasia, and oral submucosal fibrosis are also common in the at-risk group. It is also critical to distinguish between benign and malignant tumors. Age, gender, and smoking history can all have an impact on the prognosis of oral cancer [5]. Understanding the advancements of technology such as artificial intelligence may help to resolve any healthcare snags [6,7]. A sore or ulcer that does not heal and may cause discomfort or bleeding is the most prevalent indication of cancer. White or red sores in the mouth, lips, gums, or tongue that are not healing, a lump or mass in the mouth, loose teeth, chewing or swallowing difficulties, jaw swelling, trouble talking, and persistent painful throat [8]. The use of Artificial Intelligence in the treatment of oral malignant growths has the potential to address existing problems in disease detection and prognosis prediction. Artificial intelligence, which replicates human cognitive capabilities, is a technological achievement that has captured the imaginations of scientists all around the world [9]. Its application in dentistry has only just begun, resulting in amazing results. The tale begins about 300 BC. C. Plato depicted an important model of brain activity. The artificial intelligence system is a framework that accepts information, discovers designs, trains using data, and generates outcomes [10,11]. Artificial intelligence operates in two stages: the first, which involves training, and the second, which includes testing. The parameters of the model set are determined by the training data. Data from prior instances, such as patient data or data from other examples, are used retrospectively by the model. The test phase is subsequently subjected to these criteria. Various biomarkers identified by artificial intelligence in various research have established prognostic variables for oral cancer. Early detection of a malignant tumor improves patient survival and treatment options [12]. Much research on medical image analysis for smartphone-based oral cancer detectors based on artificial intelligence algorithms has been done. Artificial intelligence technology makes it easier to diagnose, treat, and manage patients with oral cancer. To aid diagnosis, artificial intelligence decreases physician burden, complicated data, and exhaustion [13]. Given the practicality and alleged benefits of deep learning techniques in cancer prediction, their use in this field has gotten a lot of interest in recent years. This is because it is designed to assist doctors in making educated decisions, therefore enhancing and encouraging better patient health management. Surprisingly, technological advancements have resulted in the transition from neural networks to deep neural networks. This deep learning method has also been lauded for its potential to enhance cancer management.

2. Literature Review

The researchers’ objective was to diagnose the early stages of oral cancer utilizing the less accurate output of Naive Bayes, Multilayer Perceptron, KNN, and Support Vector Mechanical techniques. Oral cavities and analysis improve classification precision [14]. The researcher’s objective is to develop a model for doctors. Use tree-based decision-making approaches, artificial neural network vector maintenance methods, and DATA, NN, and HDM high precision analyses. The representation of the ADM’s score is the best suited for detecting breast cancer recurrence since it has the highest precision and the lowest error level. When both ANN and DT are evaluated, the results suggest that SVM is the best technique for diagnosis [15]. Madhura V, Meghana Nagaraju, and their colleague review several reports to investigate the diagnosis of oral cancer using machine learning [16]. They then utilize categorization rules to anticipate and association rules to demonstrate attribute co-dependency [17]. It then employs the a priori technique to pick frequent item sets and construct the association rule from the bottom up, employing a breadth search and hash to efficiently count the items. In their study, researchers utilize an adaptive fuzzy system based on deep neurons to get exact findings in data mining techniques. The procedure starts with data processing and grouping with Fuzzy C-Means. The architecture of an adaptive fuzzy system based on deep neurons has been presented, and analysis methods have been used to obtain correct findings such as precision, accuracy, and so on [18]. Their study made use of data sets including 251 X-rays from the equator, which were then subdivided into experimental data testing and training methods such as in-depth ANA investigations, transfer studies, and Convolutional Neural networks [19].
The researcher’s objective was to test novel automation ways for Oral Squamous Cell Carcinoma diagnoses on clear pictures utilizing in-depth training and Convolutional Neural Network methods. The focus of this Convolutional Neural Network is on the search for quotations, pictures, training, data, and evaluation [20]. Oral cancer can exhibit a wide range of patterns and behaviors [21]. In recent years, researchers used numerous machine learning techniques to overcome cancer [22], and the machine learning models can detect cancer. Machine learning predicts oral cancer is way better than previous prediction techniques [23]. Oral cancer is a fatal disease, and the major root of this fatal disease comes from the genome [24] and a variety of pathogenic changes [25]. Early prediction of oral cancer [26] and its treatment can increase the patient survival chances. Oral cancer is a progressive and very complicated disease [27] that can only predict using numerous machine and deep learning algorithms. So, the researchers provide the combining machine learning strategies [28] to predict oral cancer in its early stages. Researchers can use different histopathological machine learning [29] approaches to overcome cancer. Some researchers provide the fused machine learning-based solution [30] to predict cancer using real-time data and achieved good accuracy using different neural algorithms. Previous research used cloud-based deep learning approaches [31] to overcome cancerous diseases and gave better treatment to decrease the high mortality rate in females. Machine learning approaches help in genes association to overcome the cancerous empowered with deep learning approaches [32]. Researchers used digital images [33] to predict cancer using artificial neural networks and deep CNN [34] approaches to get highly efficient results. During the COVID phase, various researchers apply machine learning approaches to COVID patients to predict cancer [35] and get highly efficient results using different preprocessing techniques. Deep learning radionics-based detections [36] on cancerous patients give high feature results with the help of chemoradiotherapy.
The researcher’s objective was to test novel automation ways for Oral Squamous Cell Carcinoma diagnoses on clear pictures utilizing in-depth training and Convolutional Neural Network methods. The focus of this Convolutional Neural Network is on the search for citations, pictures, training, data, and classification. The researchers used machine learning techniques and genetic data to predict oral cancer development in OPL patients. To examine the course of oral cancer in individuals with a history of OPL, the researchers employed a Support Vector Machine, Multi-Layer Perceptron, a minimally invasive procedure, and a DNN [37]. The researcher utilized ordered electrical machines to classify the usage of hyperspectral to identify lung cancer in the amplification attempt. They divided the data into pictures using a Convolutional Neural Network. In this regard, in-depth research approaches were used to address the lack of an independent design of the cancer detection system [38].
The majority of the approaches need complex system configuration, resulting in significant operational expenses. The researchers assessed several cancer detection approaches and clarified the benefits of symptomatic simulations. As a result, HSI can be used to classify data. They use a vector support machine with a self-mapping structure to categorize the data [39].
Table 1 shows most of the previous research used machine and deep learning techniques to predict oral cancer using OSCC biopsies and other datasets, but they did not achieve the highest accuracy due to their highlighted weakness. As observed from previous studies oral cancer prediction is an important mission to save many lives. So, in this study transfer learning empowered model is proposed, which extract features and train on OSCC biopsy data to predict oral cancer.

3. Materials and Methods

With the development of machine learning and deep learning, every field of medical image diagnosing is very easy to analyze and predict with the help of medical images. There are various artificial intelligence techniques to diagnose medical diseases, but deep learning has efficient results with a very effective approach. So, in this study deep learning has a major role in the first stage prediction of medical images with a high prediction efficiency and authentic results. The proposed model used a MacBook Pro 2017 with 16 GB RAM used for simulation and used transfer learning to simulate the cancerous images. The proposed model of oral cancer prediction using Oral Squamous Cell Carcinoma biopsy empowered with transfer learning Figure 1 consists of three phases. In the first phase, the proposed model takes binary input (Oral Squamous Cell Carcinoma images and Normal Images) and transfers these inputs to the image pre-processing section. The image pre-processing section performs image dimension analysis to transform input images into 227 × 227 resolution of height and width, respectively, because the AlexNet model input layer requirement is 227 × 227-dimension images for the training model.
In the second phase, the proposed model gets a pre-trained model AlexNet of the neural network then the proposed model modifies AlexNet according to the research requirements using transfer and train model with pre-processed Oral Squamous Cell Carcinoma biopsy images and then stores this trained model into the cloud for easy access in testing at any time. In the third phase, which is known as the testing phase, the trained model retrieves from the cloud and apply to the neural network, and the neural network gets the input of pre-processed Oral Squamous Cell Carcinoma biopsy images to predict oral cancer. If the cancer is diagnosed then the patient will consult with paramedical staff otherwise, the patient has to be free to leave for home.
Table 2 shows the pseudocode of the proposed model for oral cancer prediction. After starting the pseudocode first, the proposed model gets input from the data cloud and preprocessed the data for customized AlexNet, and the code starts the training phase for training the model, and then in image testing phase pseudocode of the proposed model used trained model for the prediction of cancerous images. At the end, the code calculates the prediction accuracy and misclassification rate using various performance matrices.

4. Data Set

The Oral Squamous Cell Carcinoma (OSCC) biopsy data retrieve from Kaggle, which has open access to everyone [41]. This oral cancer dataset contains binary class one is sick with oral cancer, and the second class contains normal patients. The proposed model uses this dataset to analyze and predict oral cancer. Table 3 shows the detailed dataset classes, and Figure 2 shows the samples of both classes of the dataset.

5. Customized AlexNet

In this period, deep learning algorithms are very efficient to analyze medical images with the help of many pre-trained models for prediction. In this study, the proposed model used the pre-trained convolutional neural network model AlexNet for transfer learning to analyze and predict oral cancer using Oral Squamous Cell Carcinoma biopsy images. Customized AlexNet is known as the transfer learning technique and this technique is very primary and contiguous to analyzing and predicting medical image diagnosing. So, we customized the AlexNet algorithm according to input and prediction requirements. Customized AlexNet contains two fully connected layers, five max pool layers, and five convolutional layers. Each layer of AlexNet contains ReLU as an activation function. For the input layer of AlexNet, the firstly proposed model processed images into 227 × 227 resolution because the input layer of AlexNet only read this height and width, respectively. Figure 3 shows samples of pre-processed images in 227 × 227 resolution.
Figure 4 shows the customized AlexNet model using transfer learning, in this customized AlexNet the last three layers of AlexNet were customized for analysis and prediction purposes. Input instances of all classes are fully connected. The output size of the fully connected layer is equal to the length of the total labels of the prediction class. The purpose of the Softmax layer is to detect the boundaries of input images that are represented in the convolutional layer, while the fully connected layer learns from all class features.
The purpose of the convolutional layer is to extract the features from input class images by using different image extract feature filters i-e gabber filter and secure the connection between image pixels. Each layer contains a ReLU activation function to activate neurons to weighted sum or not. So, the fully connected layer is set as binary for prediction purposes. So, the proposed model used this customized AlexNet model to train OSCC biopsy images to predict oral cancer.

6. Simulation and Results

In this study, the proposed model applied a customized pre-trained AlexNet model to predict oral cancer. MATLAB 2020a is used to simulate results on MAC Book Pro 2016 i5 with 16 GB RAM containing a high GPU. The proposed model split the dataset of oral cancer prediction into 70% for training and 30% for testing. After the testing phase, the proposed model used numerous performance statistical parameters [30,42,43,44,45,46,47] like Classification Accuracy (CA), Classification Miss Rate (CMR), Specificity, Sensitivity, F1-Score, Positive Predicted Value (PPV), Negative Predicted Value (NPV), False Positive Ratio (FPR), False Negative Ratio (FNR), Likelihood Positive Ratio (LPR), Likelihood Negative Ratio (LNR) and Fowlkes-Mallows Index to evaluate the results of oral cancer prediction. The proposed model represents Ç for predicted true positive values, ø for predicted true negative values, ∂ for predicted false positive values, and µ for predicted false negative values.
CA   =   Ç   +   ø   Ç   +   ø   + +   µ   × 100
CMR   =   100 Ç   +   ø   Ç   +   ø   + +   µ   × 100
Sensitivity   =   Ç Ç   + µ   × 100
Specificity   =   ø ø   +   × 100
F 1   score   =   2 2 + +   µ   × 100
PPV   =   Ç Ç   +   × 100
NPV   =   ø ø   +   µ × 100
FPR   =   100 ø ø   +   × 100
FNR   =   100 Ç Ç   + µ   × 100
LPR   =   Ç Ç   + µ   × 100 100 ø ø   +   × 100
LNR   =   100 Ç Ç   + µ   × 100 ø ø   +   100
Fowlkes - Mallows   Index   =   Ç Ç   +   × 100 × Ç Ç   + µ   × 100
Table 4 shows the proposed model training simulation parameters results on the various number of tested epochs with the same layers settings 227 × 227 × 3, MAX of image dimension, and pooling method, respectively. The learning rate of every layer is set at 0.001. So, the proposed model shows that mini-batch loss at the final stage of the 70 epoch is 0.7336. Table 5 shows proposed model training simulation accuracies on every 10th epoch, and every epoch consists of 38 iterations. So, the proposed model achieved the highest training accuracy 97.66%, with 00:08:34 elapsed time on the 70th epoch.
Figure 5 shows the final predicted results of Normal and oral cancer biopsies by the proposed model for oral cancer prediction. There are thousands of images that are not possible to portray in this study. So, the proposed model shows the first 36 results as a sample of tested results. These predicted images are generated using the proposed model for the prediction of oral cancer. Figure 5 shows the accurately predicted images of a cancerous cell labeled as “OSCC” and normal predicted cell images labeled as “Normal”.
Figure 6 shows the proposed model training progress on 70 epochs, as in the training progress shown proposed model converges after the 59th epoch, and at the 70th epoch proposed model achieved 97.66% accuracy with a 2.34% loss rate. Figure 7 shows the proposed model testing confusion matrix on oral cancer prediction. So, the confusion matrix consists of a total of 1483 instances for testing of class 1 “Normal” and class 2 “Oral Cancer”. As Figure 7 shows, the proposed model truly predicted 677 patients that are normal and 658 patients that have oral cancer. Furthermore, the proposed model wrong predicted 95 normal patients and 53 cancerous patients.
Table 6 shows the proposed performance parameter results empowered with transfer learning to analyze the performance of the model. Performance results of the proposed model are 90.02%, 9.08%, 92.74%, 87.38%, 90.15%, 87.69%, 92.55%, 12.62%, 7.26%, 7.35%, 0.08%, 90.18% of Classification Accuracy, Classification Miss Rate, Sensitivity, Specificity, F1-score, Positive Predicted Value, Negative Predicted Value, False Positive Ratio, False Negative Ratio, Likelihood Positive ratio, Likelihood Negative Ratio and Fowlkes-Mallows Index respectively. So, the proposed model of oral cancer prediction using Oral Squamous Cell Carcinoma biopsy empowered with transfer learning achieved the best prediction accuracy as compared with the previous studies because the proposed model covers the previous limitation e.g., used data image preprocessing layers and learning criteria accurately to achieve high prediction results with the at most reliability. With the help of data preprocessing, the proposed model achieved the best predicted results with the help of a customized AlexNet algorithm, customized layers of AlexNet play a pivotal role to get high prediction accuracy.
Table 7 shows the comparative analysis of the proposed model for oral cancer prediction using transfer learning with previous research. The proposed model performance in terms of classification accuracy during the testing phase. During the testing phase, the proposed model gives 90.06% and 9.94%, classification accuracy, and miss-Classification rate, respectively, and during the training phase the proposed model gives 97.6% and 2.34%, classification accuracy, and miss-rate respectively. The overall summary of the analysis shows that the classification accuracy is improved by around 12% as compared with the previously proposed models of machine learning [26,27] and deep learning [28]. The proposed model gives better classification accuracy as compared with the previously published research [26,27,28].

7. Conclusions

This research used customized AlexNet of CNN to predict cancerous oral and normal oral using Oral Squamous Cell Carcinoma biopsy images. The performance parameters like Classification Accuracy, Classification Miss Rate, Sensitivity, Specificity, F1-score, Positive Predicted Value, Negative Predicted Value, False Positive Ratio, False Negative Ratio, Likelihood Positive ratio, Likelihood Negative Ratio, and Fowlkes-Mallows Index were used to analyze the proposed model optimally and efficiently. With the help of statistical performance parameters, we can easily measure the reliability of the proposed model for the prediction of oral cancer. So, the proposed model achieved 90.06% and 9.08% of prediction accuracy and loss rate, respectively. The reason behind the highest accuracy achievement of the proposed model is its customized layer techniques, dataset preprocessing techniques, and training epochs. This study and the proposed model can be used in the medical field to prevent unnecessary treatments and tests, also paramedic staff can easily do the treatment of cancerous patients to prevent unnecessary deaths. The major contribution of this study in health 5.0 for early prediction and treatment of oral cancer. In the future, this proposal can be extended to fuse datasets and also can apply fuzzy techniques to get more precise results empowered with federated machine learning and blockchain for data security and fast prediction results.
The method is simple to use, affordable, and appropriate for oral cancer detection in impoverished nations. In a future study, scientists can utilize real-time data and a large database to enhance the transfer learning and machine learning structure. Using histopathological images, this research can also be used for the early diagnosis of different stages of oral squamous cancer detection. This will help doctors provide proper medicine and save lives.

Author Contributions

A.-u.R., M.F.K. and A.A. have collected data from different resources; M.U.N., M.A.K. and N.A. performed formal analysis and Simulation; M.F.K., A.-u.R. and A.M. contributed to writing—original draft preparation, M.F.K., M.U.N., A.A. and A.M. writing—review and editing, M.A.K. and N.A. performed supervision; M.U.N., A.-u.R. and A.M. drafted pictures and tables, A.A. and M.U.N. performed revisions and improve the quality of the draft. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The simulation files/data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. World Health Organization. WHO Report on Cancer: Setting Priorities, Investing Wisely and Providing Care for All; Technical Report; World Health Organization: Geneva, Switzerland, 2020.
  2. Sinevici, N.; O’sullivan, J. Oral cancer: Deregulated molecular events and their use as biomarkers. Oral Oncol. 2016, 61, 12–18. [Google Scholar] [CrossRef] [PubMed]
  3. Ilhan, B.; Lin, K.; Guneri, P.; Wilder-Smith, P. Improving Oral Cancer Outcomes with Imaging and Artificial Intelligence. J. Dent. Res. 2020, 99, 241–248. [Google Scholar] [CrossRef] [PubMed]
  4. Dhanuthai, K.; Rojanawatsirivej, S.; Thosaporn, W.; Kintarak, S.; Subarnbhesaj, A.; Darling, M.; Kryshtalskyj, E.; Chiang, C.P.; Shin, H.I. Oral cancer: A multicenter study. Med. Oral Patol. Oral Cir. Bucal 2018, 23, 23–29. [Google Scholar] [CrossRef] [PubMed]
  5. Lavanya, L.; Chandra, J. Oral Cancer Analysis Using Machine Learning Techniques. Int. J. Eng. Res. Technol. 2019, 12, 596–601. [Google Scholar]
  6. Kearney, V.; Chan, J.W.; Valdes, G.; Solberg, T.D.; Yom, S.S. The application of artificial intelligence in the IMRT planning process for head and neck cancer. Oral Oncol. 2018, 87, 111–116. [Google Scholar] [CrossRef]
  7. Hamet, P.; Tremblay, J. Artificial intelligence in medicine. Metabolism 2017, 69, 36–40. [Google Scholar] [CrossRef]
  8. Head and Neck sq Squamous Cell Carcinoma. Available online: https://en.wikipedia.org/wiki/Head_and_neck_sqsquamous-cell_carcinoma (accessed on 12 January 2022).
  9. Kaladhar, D.; Chandana, B.; Kumar, P. Predicting Cancer Survivability Using Classification Algorithms. Int. J. Res. Rev. Comput. Sci. 2011, 2, 340–343. [Google Scholar]
  10. Kalappanavar, A.; Sneha, S.; Annigeri, R.G. Artificial intelligence: A dentist’s perspective. J. Med. Radiol. Pathol. Surg. 2018, 5, 2–4. [Google Scholar] [CrossRef]
  11. Krishna, A.B.; Tanveer, A.; Bhagirath, P.V.; Gannepalli, A. Role of artificial intelligence in diagnostic oral pathology-A modern approach. J. Oral Maxillofac. Pathol. JOMFP 2020, 24, 152–156. [Google Scholar] [CrossRef]
  12. Kareem, S.A.; Pozos-Parra, P.; Wilson, N. An application of belief merging for the diagnosis of oral cancer. Appl. Soft Comput. J. 2017, 61, 1105–1112. [Google Scholar] [CrossRef]
  13. Kann, B.H.; Aneja, S.; Loganadane, G.V.; Kelly, J.R.; Smith, S.M.; Decker, R.H.; Yu, J.B. Pretreatment Identification of Head and Neck Cancer Nodal Metastasis and Extranodal Extension Using Deep Learning Neural Networks. Sci. Rep. 2018, 8, 14306. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  14. Mohd, F.; Noor, N.M.; Abu Bakar, Z.; Rajion, Z.A. Analysis of Oral Cancer Prediction using Features Selection with Machine Learning. In Proceedings of the ICIT The 7th International Conference on Information Technology, Amman, Jordan, 12–15 May 2015. [Google Scholar]
  15. Ahmad, L.G.; Eshlaghy, A.G.; Poorebrahimi, A.; Ebrahimi, M.; Razavi, A.R. Using Three Machine Learning Techniques for Predicting Breast Cancer Recurrence. Health Med. Inform. 2013, 4, 2157–7420. [Google Scholar]
  16. Vidhya, M.S.; Nisha, D.S.S.; Sathik, D.M.M. Denoising the CT Images for Oropharyngeal Cancer using Filtering Techniques. Int. J. Eng. Res. Technol. 2020, 8, 1–4. [Google Scholar]
  17. Chakraborty, D.; Natarajan, C.; Mukherjee, A. Advances in oral cancer detection. Adv. Clin. Chem. 2019, 91, 181–200. [Google Scholar]
  18. Anitha, N.; Jamberi, K. Diagnosis, and Prognosis of Oral Cancer using classification algorithm with Data Mining Techniques. Int. J. Data Min. Tech. Appl. 2017, 6, 59–61. [Google Scholar]
  19. Available online: https://www.ahns.info/resources/education/patient_education/oralcavity (accessed on 12 January 2022).
  20. Aubreville, M.; Knipfer, C.; Oetter, N.; Jaremenko, C.; Rodner, E.; Denzler, J.; Bohr, C.; Neumann, H.; Stelzle, F.; Maier, A. Automatic Classification of Cancerous Tissue in Laserendomicroscopy Images of the Oral Cavity using Deep Learning. Sci. Rep. 2017, 7, 11979. [Google Scholar] [CrossRef] [Green Version]
  21. Alabi, R.O.; Youssef, O.; Pirinen, M.; Elmusrati, M.; Mäkitie, A.A.; Leivo, I.; Almangush, A. Machine learning in oral squamous cell carcinoma: Current status, clinical concerns and prospects for future—A systematic review. Artif. Intell. Med. 2021, 115, 102060. [Google Scholar] [CrossRef]
  22. Sun, M.-L.; Liu, Y.; Liu, G.-M.; Cui, D.; Heidari, A.A.; Jia, W.-Y.; Ji, X.; Chen, H.-L.; Luo, Y.-G. Application of Machine Learning to Stomatology: A Comprehensive Review. IEEE Access 2020, 8, 184360–184374. [Google Scholar] [CrossRef]
  23. Alkhadar, H.; Macluskey, M.; White, S.; Ellis, I.; Gardner, A. Comparison of machine learning algorithms for the prediction of five-year survival in oral squamous cell carcinoma. J. Oral Pathol. Med. 2021, 50, 378–384. [Google Scholar] [CrossRef]
  24. Saba, T. Recent advancement in cancer detection using machine learning: Systematic survey of decades, comparisons and challenges. J. Infect. Public Health 2020, 13, 1274–1289. [Google Scholar] [CrossRef]
  25. Kouznetsova, V.L.; Li, J.; Romm, E.; Tsigelny, I.F. Finding distinctions between oral cancer and periodontitis using saliva metabolites and machine learning. Oral Dis. 2021, 27, 484–493. [Google Scholar] [CrossRef] [PubMed]
  26. Alhazmi, A.; Alhazmi, Y.; Makrami, A.; Salawi, N.; Masmali, K.; Patil, S. Application of artificial intelligence and machine learning for prediction of oral cancer risk. J. Oral Pathol. Med. 2021, 50, 444–450. [Google Scholar] [CrossRef] [PubMed]
  27. Chu, C.S.; Lee, N.P.; Adeoye, J.; Thomson, P.; Choi, S.W. Machine learning and treatment outcome prediction for oral cancer. J. Oral Pathol. Med. 2020, 49, 977–985. [Google Scholar] [CrossRef] [PubMed]
  28. Welikala, R.A.; Remagnino, P.; Lim, J.H.; Chan, C.S.; Rajendran, S.; Kallarakkal, T.G.; Zain, R.B.; Jayasinghe, R.D.; Rimal, J.; Kerr, A.R.; et al. Automated Detection and Classification of Oral Lesions Using Deep Learning for Early Detection of Oral Cancer. IEEE Access 2020, 8, 132677–132693. [Google Scholar] [CrossRef]
  29. Mehmood, S.; Ghazal, T.M.; Khan, M.A.; Zubair, M.; Naseem, M.T.; Faiz, T.; Ahmad, M. Malignancy Detection in Lung and Colon Histopathology Images Using Transfer Learning with Class Selective Image Processing. IEEE Access 2022, 10, 25657–25668. [Google Scholar] [CrossRef]
  30. Ahmed, U.; Issa, G.F.; Khan, M.A.; Aftab, S.; Said, R.A.T.; Ghazal, T.M.; Ahmad, M. Prediction of Diabetes Empowered with Fused Machine Learning. IEEE Access 2022, 10, 8529–8538. [Google Scholar] [CrossRef]
  31. Siddiqui, S.Y.; Haider, A.; Ghazal, T.M.; Khan, M.A.; Naseer, I.; Abbas, S.; Rahman, M.; Khan, J.A.; Ahmad, M.; Hasan, M.K.; et al. IoMT Cloud-Based Intelligent Prediction of Breast Cancer Stages Empowered with Deep Learning. IEEE Access 2021, 9, 146478–146491. [Google Scholar] [CrossRef]
  32. Sikandar, M.; Sohail, R.; Saeed, Y.; Zeb, A.; Zareei, M.; Khan, M.A.; Khan, A.; Aldosary, A.; Mohamed, E.M. Analysis for Disease Gene Association Using Machine Learning. IEEE Access 2020, 8, 160616–160626. [Google Scholar] [CrossRef]
  33. Kuntz, S.; Krieghoff-Henning, E.; Kather, J.N.; Jutzi, T.; Höhn, J.; Kiehl, L.; Hekler, A.; Alwers, E. Gastrointestinal cancer classification and prognostication from histology using deep learning: Systematic review. Eur. J. Cancer 2021, 155, 200–215. [Google Scholar] [CrossRef]
  34. Heuvelmans, M.A.; Ooijen, P.M.A.; Sarim, A.; Silva, C.F.; Han, D.; Heussel, C.P.; Hickes, W.; Kauczor, H.-U.; Novotny, P.; Peschl, H.; et al. Lung cancer prediction by Deep Learning to identify benign lung nodules. Lung Cancer 2021, 154, 1–4. [Google Scholar] [CrossRef]
  35. Ibrahim, D.M.; Elshennawy, N.M.; Sarhan, A.M. Deep-chest: Multi-classification deep learning model for diagnosing COVID-19, pneumonia, and lung cancer chest diseases. Comput. Biol. Med. 2021, 132, 104348. [Google Scholar] [CrossRef] [PubMed]
  36. Liu, X.; Zhang, D.; Liu, Z.; Li, Z.; Xie, P.; Sun, K.; Wei, W.; Dai, W.; Tang, Z. Deep learning radiomics-based prediction of distant metastasis in patients with locally advanced rectal cancer after neoadjuvant chemoradiotherapy: A multicentre study. EBio Med. 2021, 69, 103442. [Google Scholar] [CrossRef] [PubMed]
  37. Wafaa Shams, K.; Zaw, Z. Oral Cancer Prediction Using Gene Expression Profiling and Machine Learning. Int. J. Appl. Eng. Res. 2017, 12, 0973–4562. [Google Scholar]
  38. Kalantari, N.; Ramamoorthi, R. Deep high dynamic range imaging of dynamic scenes. ACM Trans. Graph. 2017, 36, 144:1–144:12. [Google Scholar] [CrossRef]
  39. Deepak Kumar, J.; Surendra Bilouhan, D.; Kumar, R.C. An approach for hyperspectral image classification by optimizing SVM using self-organizing map. J. Comput. Sci. 2018, 25, 252–259. [Google Scholar]
  40. Shavlokhova, V.; Sandhu, S.; Flechtenmacher, C.; Koveshazi, I.; Neumeier, F.; Padrón-Laso, V.; Jonke, Ž.; Saravi, B.; Vollmer, M.; Vollmer, A.; et al. Deep Learning on Oral Squamous Cell Carcinoma Ex Vivo Fluorescent Confocal Microscopy Data: A Feasibility Study. J. Clin. Med. 2021, 10, 5326. [Google Scholar] [CrossRef]
  41. Available online: https://www.kaggle.com/ashenafifasilkebede/dataset?select=val (accessed on 6 January 2022).
  42. Chen, C.; Zhou, K.; Zha, M.; Qu, X.; Guo, X.; Chen, H.; Xiao, R. An effective deep neural network for lung lesions segmentation from COVID-19 CT images. IEEE Trans. Ind. Inform. 2021, 17, 6528–6538. [Google Scholar] [CrossRef]
  43. Abdel-Basset, M.; Chang, V.; Mohamed, R. HSMA_WOA: A hybrid novel Slime mould algorithm with whale optimization algorithm for tackling the image segmentation problem of chest X-ray images. Appl. Soft Comput. 2020, 95, 106642. [Google Scholar] [CrossRef]
  44. Rahman, A.-U.; Abbas, S.; Gollapalli, M.; Ahmed, R.; Aftab, S.; Ahmad, M.; Khan, M.A.; Mosavi, A. Rainfall Prediction System Using Machine Learning Fusion for Smart Cities. Sensors 2022, 22, 3504. [Google Scholar] [CrossRef]
  45. Saleem, M.; Abbas, S.; Ghazal, T.M.; Khan, M.A.; Sahawneh, N.; Ahmad, M. Smart cities: Fusion-based intelligent traffic congestion control system for vehicular networks using machine learning techniques. Egypt. Inform. J. 2022; in press. [Google Scholar] [CrossRef]
  46. Nadeem, M.W.; Goh, H.G.; Khan, M.A.; Hussain, M.; Mushtaq, M.F.; Ponnusamy, V.A. Fusion-based machine learning architecture for heart disease prediction. Comput. Mater. Contin. 2021, 67, 2481–2496. [Google Scholar] [CrossRef]
  47. Siddiqui, S.Y.; Athar, A.; Khan, M.A.; Abbas, S.; Saeed, Y.; Khan, M.F.; Hussain, M. Modelling, simulation and optimization of diagnosis cardiovascular disease using computational intelligence approaches. J. Med. Imaging Health Inform. 2020, 10, 1005–1022. [Google Scholar] [CrossRef]
Figure 1. Proposed model of oral cancer prediction empowered with transfer learning.
Figure 1. Proposed model of oral cancer prediction empowered with transfer learning.
Sensors 22 03833 g001
Figure 2. Sick and healthy oral squamous cell carcinoma biopsy dataset.
Figure 2. Sick and healthy oral squamous cell carcinoma biopsy dataset.
Sensors 22 03833 g002
Figure 3. Pre-processed (227 × 227) oral squamous cell carcinoma biopsy images.
Figure 3. Pre-processed (227 × 227) oral squamous cell carcinoma biopsy images.
Sensors 22 03833 g003
Figure 4. Customized AlexNet for oral cancer prediction using transfer learning.
Figure 4. Customized AlexNet for oral cancer prediction using transfer learning.
Sensors 22 03833 g004
Figure 5. The proposed model predicted results of oral cancer during validation.
Figure 5. The proposed model predicted results of oral cancer during validation.
Sensors 22 03833 g005
Figure 6. Proposed model of oral cancer prediction accuracy and loss with respect to iteration during training.
Figure 6. Proposed model of oral cancer prediction accuracy and loss with respect to iteration during training.
Sensors 22 03833 g006
Figure 7. Proposed model testing confusion matrix.
Figure 7. Proposed model testing confusion matrix.
Sensors 22 03833 g007
Table 1. Compare and weaknesses of previous studies.
Table 1. Compare and weaknesses of previous studies.
PublicationMethodDatasetAccuracyLimitation
A.Alhazmi [26]ANNPublic78.95%Requires data preprocessing
C.S. Chu [27]SVM, KNNPublic70.59%Requires data preprocessing
R.A.Welikala [28]ResNet101Public78.30%Requires data preprocessing and learning criteria decision method
V. Shavlokhova [40]CNNPrivate77.89%Requires better image data preprocessing techniques and learning criteria method
M. Aberville [20]Deep LearningPublic80.01%Requires data image preprocessing techniquesClass instances
H. Alkhadar [23]KNN, Logistic Regression, Decision Tree, Random ForestPublic76%Requires handcrafted features
Table 2. Pseudocode of the proposed model for oral cancer prediction.
Table 2. Pseudocode of the proposed model for oral cancer prediction.
1Start
2Input Oral Cancer Data from Data Cloud
3Pre-process Oral Cancer data
4Load Data
5Load Customized Model
6Prediction of Oral Cancer using Transfer Learning (AlexNet)
7Training Phase
8Image Testing Phase
9Compute the Performance and Accuracy of the proposed model by using the Performance Matrix
10Finish
Table 3. OSCC biopsy dataset instances.
Table 3. OSCC biopsy dataset instances.
ClassesNo. of Images
Sick (OSCC)2511
Healthy2435
Table 4. Proposed model training simulation parameters.
Table 4. Proposed model training simulation parameters.
No. of EpochsLearning RateNo. of LayersImage DimensionPooling MethodMini-Batch Loss
100.00125227 × 227 × 3MAX2.5674
202.3498
301.3600
401.4948
506.1029
600.2491
700.3736
Table 5. Proposed model training simulation accuracies.
Table 5. Proposed model training simulation accuracies.
No. of EpochsLearning RateAccuracy (%)Loss Rate (%)IterationsTime Elapsed (hh:mm:ss)
100.00176.1223.8838 per
epoch
00:03:15
2080.3519.6500:03:45
3086.1513.8500:04:34
4090.629.3800:04:55
5085.9414.0600:06:11
6094.445.5600:07:17
7097.662.3400:08:34
Table 6. Proposed model performance parameter results using transfer learning.
Table 6. Proposed model performance parameter results using transfer learning.
Instances (1483)Testing (%)
CA90.02
CMR9.08
Sensitivity92.74
Specificity87.38
F1-Score90.15
PPV87.69
NPV92.55
FPR12.62
FNR7.26
LPR7.35
LNR0.08
FMI90.18
Table 7. Comparative analysis with previous research.
Table 7. Comparative analysis with previous research.
WorkPreprocessing LayerModelsClassification AccuracyMiss-Classification Rate
A.Alhazmi [26]NoANN78.95%21.05%
C.S. Chu [27]NoSVM, KNN70.59%29.41%
R.A.Welikala [28]NoResNet10178.30%21.70%
Proposed ModelYesTransfer Learning (AlexNet)90.06%9.94%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Rahman, A.-u.; Alqahtani, A.; Aldhafferi, N.; Nasir, M.U.; Khan, M.F.; Khan, M.A.; Mosavi, A. Histopathologic Oral Cancer Prediction Using Oral Squamous Cell Carcinoma Biopsy Empowered with Transfer Learning. Sensors 2022, 22, 3833. https://doi.org/10.3390/s22103833

AMA Style

Rahman A-u, Alqahtani A, Aldhafferi N, Nasir MU, Khan MF, Khan MA, Mosavi A. Histopathologic Oral Cancer Prediction Using Oral Squamous Cell Carcinoma Biopsy Empowered with Transfer Learning. Sensors. 2022; 22(10):3833. https://doi.org/10.3390/s22103833

Chicago/Turabian Style

Rahman, Atta-ur, Abdullah Alqahtani, Nahier Aldhafferi, Muhammad Umar Nasir, Muhammad Farhan Khan, Muhammad Adnan Khan, and Amir Mosavi. 2022. "Histopathologic Oral Cancer Prediction Using Oral Squamous Cell Carcinoma Biopsy Empowered with Transfer Learning" Sensors 22, no. 10: 3833. https://doi.org/10.3390/s22103833

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop