Next Article in Journal
Application of a Quantitative Risk-Based Decision Tool for Local Railway Level Crossing Management
Previous Article in Journal
Adaptive MAC Scheme for Interference Management in Ad Hoc IoT Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Privacy-Preserving Transfer Learning Framework for Kidney Disease Detection

Department of Computer Engineering, Kahramanmaras Sutcu Imam University, Kahramanmaras 46050, Türkiye
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(19), 8629; https://doi.org/10.3390/app14198629
Submission received: 31 August 2024 / Revised: 13 September 2024 / Accepted: 18 September 2024 / Published: 25 September 2024
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
This paper introduces a new privacy-preserving transfer learning framework for the classification of kidney diseases. In the proposed framework, transfer learning is employed for feature extraction, and differential privacy is used to obtain noisy gradients. A variety of CNN architectures, including Xception, ResNet50, InceptionResNetV2, MobileNet, DenseNet201, InceptionV3, and VGG19 are utilized to evaluate the proposed framework. Analysis of a large dataset of 12,400 labeled kidney CT images shows that transfer learning architectures based on the proposed framework achieve excellent accuracy ratios in privacy-preserving classification. These results demonstrate the effectiveness of the proposed framework in enabling transfer learning models to classify kidney diseases while ensuring privacy. The MobileNet architecture stands out for its exceptional performance, with an impressive accuracy of 99.83% in privacy-preserving classification. Considering the findings of this study, it is evident that the proposed framework is appropriate for the early and private diagnosis of kidney diseases and promotes the achievement of promising results in this field.

1. Introduction

Chronic kidney disease (CKD) is a widespread health issue that impacts over 10% of the global population and directly affects an incredible amount of 800 million people. The prevalence of CKD is notably higher in older people, women, racial minorities, and those with chronic conditions like as diabetes mellitus and hypertension. CKD is a prominent contributor to global mortality and is among the limited number of non-communicable diseases that have contributed to a rise in deaths during the past two years [1]. In the year 2016, the disease ranked as the 16th most common cause of mortality. Projections indicate that by 2040, it will rank as one of the most lethal diseases, increasing to the fifth position [2]. The high number of affected individuals and the significant negative impact of CKD call for urgent solutions to develop early diagnosis systems [1].
Machine imaging techniques, including magnetic resonance imaging (MRI), computed tomography (CT), positron emission tomography (PET), mammography, ultrasound, and X-ray, are extensively employed in the medical imaging domain [3]. Recent developments in medical imaging methods have been crucial for the early diagnosis and treatment of diseases. Nevertheless, this procedure is vulnerable to human error as the task of visual interpretation is typically carried out by highly skilled professionals and these professionals are not always available. Therefore, automated systems are always needed to help doctors’ diagnosis decisions.
Although traditional machine learning approaches may not be adequate for properly and effectively analyzing the complexity of medical images, the combination of deep learning methods and the advancement of fast processors enables precise and efficient analysis of medical images. This approach enables the identification and interpretation of patterns in medical imaging. Historically, the medical sector has not completely harnessed artificial intelligence owing to hardware constraints. However, recent advancements in deep learning algorithms have seen significant progress in the healthcare domain. These advancements have resulted in a monumental revolution in the domain of medical image processing. Specifically, considering the constraints of the convolutional neural networks (CNNs) approach, deep learning models show significant promise in the healthcare sector. Utilizing this approach enables healthcare practitioners to examine complex patterns in medical images with higher precision, therefore offering more efficient and expedited solutions in the realms of diagnosis [4].
CNN is a highly efficient deep learning model in the field of image processing, primarily employed for image classification and recognition. A CNN is composed of several layers that divide the inputs into smaller regions, employing convolution filters. Customization of this model enables the identification and determination of significant characteristics in medical images. In the context of CT kidney imaging, a CNN model can provide significant assistance in the diagnosis of kidney diseases [5].
Radiological imaging is of utmost importance in the identification and assessment of kidney problems. These images provide a detailed depiction of the internal structure and operational characteristics of the kidneys. Nevertheless, the interpretation of these images is a laborious and specialized procedure. CNN models have the capability to identify anomalies in CT scans, therefore facilitating the early diagnosis of diseases. The implementation of this technology can facilitate the early diagnosis of diseases and enhance the management of treatment. The advancement and use of these technologies can serve as a crucial measure in addressing significant health issues, including CDK [6,7].
On the other hand, studies emphasize the issue that disease data, such as CT scans, are highly sensitive. Rigid privacy measures are essential in this situation to protect patients’ sensitive information from attacks and breaches. Insufficient measures may result in people being subjected to discrimination, stigmatization, or financial loss if their health information is disclosed. Furthermore, implementing privacy measures can support the establishment of trust between patients and healthcare professionals, thereby promoting early diagnosis without the concern of their confidential health information being disclosed. It is essential to use privacy-preserving methods for extracting important information from such data [8,9].
Current kidney disease detection studies face several key challenges, such as the following:
  • Lack of high detection accuracy: since kidney disease is highly likely to result in death, high detection accuracy is very important for early diagnosis. The studies in the literature still do not present higher accuracy;
  • Privacy concerns and inference attack issue: kidney disease detection studies employ sensitive data, such as CT scans, which may disclose the privacy of the patients.
In order to overcome these challenges, the proposed framework introduces these solutions:
  • Creation of optimal classifier: to obtain high detection accuracy, the proposed framework employs an optimal classifier, including appropriate dense layers, which gives better results;
  • Implementation of differential privacy for optimizer: the proposed framework utilizes differential privacy for noisy gradient computing. Hence, we applied differential privacy in an Adam optimizer in order to provide privacy.
The main contributions of this paper are listed below:
  • We introduce a novel framework that includes an optimal neural network classifier including appropriate dense layers. The proposed framework provides the best results in the literature.
  • Motivated by the privacy issue in kidney disease detection, we present the use of a differentially private Adam optimizer in transfer learning to detect kidney diseases by respecting privacy for the first time.
This study introduces a novel framework that aims to facilitate the detection of chronic kidney disorders, including tumors, cysts, stones, and normal tissue, while respecting privacy. This framework facilitates the deployment of differential privacy with CNN-based architectures for the analysis of CT images of the kidney. The objective of the study is to employ CNN-based architectures to identify anomalies in the kidney by acquiring knowledge of particular patterns and characteristics embedded in the images, while ensuring the privacy of the data is preserved.
The rest of this paper is organized as follows. Section 2 includes a summary of prior studies conducted in the field. Section 3 outlines the materials and methods employed in the study. Section 4 introduces the proposed framework that addresses the handled problem. Section 5 presents the experimental studies conducted in this paper. Finally, Section 6 concludes the research by summarizing the essential findings and their implications.

2. Literature Review

The main purpose of this section is to review the studies on detecting diseases using a CNN-based architecture with CT images. Additionally, it provides a table for future research by comparing accuracy ratios and datasets acquired from various catalogs utilized in the current literature. Below is a summary of research in the literature that investigates the use of these architectures with CT scans for different disease detection.
In [10], the authors assessed the efficiency of deep learning techniques in the task of liver segmentation. Comparative analyses were conducted on two U-Net architectures of varying sizes and a basic CNN model. The U-Net model triggered by the ReLU exhibited a highly successful performance with an F1 score of 97.58% when trained on a 256 × 256 input size. In comparison, the U-Net model trained on a 512 × 512 input size showed a 2.61% reduction in the F1 score. The F1 score values of the U-Net models trained with sigmoid and linear activation functions showed a notable reduction in comparison to the 256 × 256 U-Net model.
In [11], Kayhan introduced a 3D U-Net model that incorporates an early fusion technique for the purpose of automatically segmenting organs on CT scans. The proposed model employs a two-stage early fusion technique that incorporates CT scans with distinct color spaces. High accuracy rates were attained in the tests for large-size organs; however, the degree of success was limited for small-size organs. Training and testing of the proposed model yielded an average segmentation accuracy of 88.8% and 79.6% for all organs, respectively. The automated segmentation procedure produced successful and promising outcomes. However, further inclusion of training and test data could enhance the model’s performance and enable advances in segmenting tiny-size organs. In comparison to CNN and FCN architectures, the proposed approach demonstrated superior performance.
In [12], Katar examined the advancement of a deep learning system for the purpose of diagnosing COVID -19 using chest CT images. The present system comprises two distinct models that possess both classification and semantic segmentation functionalities. This study employed the EfficientNetB2 model as the classification model and achieved a 99.75% accuracy in classifying chest CT scans as either positive or negative for COVID-19. The U-Net model was employed for semantic segmentation, and a ResNet50-based model achieved the highest performance, producing a Jaccard index of 93.76% and a Dice score of 96.61%.
Al-Areqi et al. [13] explored the application of CT scans in the diagnosis of COVID-19 and assessed the efficacy of machine learning and transfer learning models. A COVID-CT dataset, which included 349 positive and 397 negative cases, was employed. Extraction of shape, statistical, and texture information from CT images was followed by classification using different classifiers. Further classification was conducted using transfer learning models. The findings indicate that the XGBoost classifier achieved the highest accuracy of 98.65%, whereas VGG19 and the ResNet50 model achieved an accuracy of 98.04% and 99.62%, respectively.
The utility of image processing technology in the diagnosis of lung cancer was examined in [14]. A dataset comprising CT images of lung, including healthy, malignant, and benign cases, was utilized. This study aimed to develop a lung cancer diagnosis system by leveraging pretrained convolutional neural networks and a novel hybrid model. Empirical studies have revealed that the developed model greatly enhances the precision of lung cancer diagnosis. Although the accuracy rate by deep learning architectures only reached 92%, the proposed approach demonstrated a notable result with an accuracy of 98.3%.
The performance of CNNs for predicting the risk of COVID-19 is assessed in [15]. The study utilized CT scans acquired from clinically diagnosed COVID-19 patients. The collection comprises 349 CT pictures that are positive and 397 CT images that are negative, extracted from 216 people. Analysis and evaluation were conducted on CNN-based architectures VGG16, ResNet18, ResNet50, DenseNet121, DenseNet169, EfficientNetB0, and EfficientNetB1. The use of pretrained weights resulted in a performance enhancement of approximately 10–15% for all networks, with EfficientNetB1 and DenseNet169 achieving an accuracy level above 85%.
In their study [16], Narmada et al. proposed a CNN model for accurate diagnosis of kidney diseases. The researchers utilized a total of 12,510 CT images obtained from distinct hospitals. The images were classified into four distinct categories all pertaining to kidney health. A CNN-based classification model was developed. Sigmoid and ReLU activation functions were employed. Analysis of the model’s performance was conducted using visual representations, such as a confusion matrix and performance table. The statistical analysis revealed that 99.3% of the whole samples in the test set were accurately classified.
The study by Mehr [17] introduces deep learning techniques for diagnosing lung cancer using lung CT images. The study utilized lung CT images obtained from the Data Science Bowl and Kaggle databases. The classification task utilized the AlexNet and GoogleNet designs, resulting in accuracy rates of 95.919% and 96.360%, respectively.
Wang et al. [18] introduced a novel approach to facilitate the identification of liver disease. Here, they provide an enhanced Deeplabv3+ model specifically designed for segmenting the liver in CT scans. The main enhancement of Deeplabv3+ is the substitution of its backbone with ResNet50, which aims to prevent the occurrence of vanishing gradients. Moreover, a channel attention module is incorporated into every ResNet50 residual block, enabling the network to selectively concentrate on significant information. The incorporation of transfer learning in the improved network serves to expedite the merging process and mitigate the potential for overfitting when dealing with a restricted dataset. By attaining an average union over intersection of 95.40% and a Dice coefficient of 94.80%, the proposed technique surpasses U-net and U-net++, demonstrating exceptional accuracy in liver segmentation.
The study of Hossain et al. [6] demonstrates the efficiency of neural networks in the early detection of chronic kidney disease. Classification of kidney CT images was performed using three distinct neural network classification approaches. Initially, the images underwent watershed segmentation procedures. Next, many neural networks, including EAnet, ResNet50, and a proposed CNN model, underwent training using a dataset that is accessible to the public on Kaggle. The test findings indicate that EANet, ResNet50, and the proposed CNN model attained accuracy rates of 83.65%, 87.92%, and 98.66%, respectively on the test set of classification models. The proposed CNN model demonstrates superior sensitivity and specificity, resulting in the highest overall accuracy.
A study conducted by Xu et al. [19] examined deep learning techniques for the detection of lesions in CT scans of COVID-19. This approach, designed to alleviate the workload of physicians in the diagnostic procedure, is founded on the Faster R-CNN architectural framework. In order to achieve more precise outcomes, the data underwent initial clustering using Kmeans++. Subsequently, the Faster R-CNN model was evaluated using two distinct backbone networks, namely VGG16 and ResNet50. Experimental evaluations were conducted on both the raw dataset and the enhanced dataset. The Faster R-CNN model with the VGG16 backbone demonstrated superior performance on the enriched dataset. Using all test data, this model attained a recall of 68.12% and a precision of 65.58%. The percentage of misclassification was observed as 31.88%.
Consequently, the aforementioned studies have shown that deep learning models yield highly effective outcomes in several classification tasks on medical images. CNN is the most favored model type in these experiments. CNNs exert a significant impact in the field of image processing and analysis, making them very well suited for addressing medical image classification problems.
Table 1 presents a comparison of the CNN architecture with several evaluations conducted on CT images. It shows the utilization of CNN models by several researchers in CT classification tasks and provides a summary of classification evaluations conducted on different CT images using deep learning models. Each study focuses on classification problems related to distinct organs or disorders. It can be seen that this study proposes a privacy-aware kidney disease detection for the first time and provides better results.

3. Material and Methods

This section outlines the materials and methods employed in this study.

3.1. CNN

CNN is extensively and efficiently utilized among deep learning models. The concept of CNN was initially proposed by Kunihiko Fukushima in the 1980s. Yann LeCun and his team subsequently integrated CNN with backpropagation theory to accurately identify handwritten numbers and documents. It is highly advantageous for the processing and analysis of visual data, particularly images and videos. The primary advantage of this technology lies in its capacity to acquire hierarchical representations, therefore facilitating a remarkable transformation in the domains of image processing and computer vision through direct learning from pixel data [20]. Medical image processing has seen the extensive application of CNN in recent years. This model demonstrates exceptional performance in tasks including object identification and recognition, image classification, and segmentation [21,22]. The CNN’s capacity to extract significant characteristics and its high performance have established it as a favored option, particularly for the processing of complicated medical images, like CT scans [15].
A CNN is composed of several layers that convert inputs into smaller regions using convolution filters [3]. Multi-layer architecture is employed to identify visual characteristics. These layers are partitioned into two distinct stages: extraction of features and classification. The feature extraction phase consists of an input layer, a convolution layer, and a pooling layer, whereas the classification phase consists of a fully connected layer and an output module [23].
The input data are provided to the network in their raw form, and the resulting resolution directly impacts the network’s performance. Applying filters, the convolution layer derives feature maps from the incoming data. Efficient extraction of visual features requires the use of a specific number of filters. Filters are shifted over the image and matrix multiplications are executed. Summation of the values derived from the multiplications yields the final value. This layer serves the function of identifying patterns in areas of the image that are often invisible to the human visual system. The initial convolution layers extract features at a lower level, whereas the deepest convolution layers extract features at a higher level. The majority of the intensive processing required by the CNN is carried out in this layer [24]. By decreasing the number and size of the network’s parameters, the pooling layer decreases the computational cost and minimizes memory usage. The pooling layer, following the convolution layer, has filters that are convolved with the output of the convolution layer. Within CNNs, three distinct pooling techniques are employed: maximum, minimum, and average pooling [13,25].
The final layer, known as the fully connected layer, aggregates the data by assigning weights and computes probability values for the purpose of classification [3]. Within the fully connected layer, the neurons from the preceding layer are interconnected in their entirety during the flattening phase. In order to make the data from the hidden layers appropriate for the fully connected layer, a flattening procedure is employed. As a consequence of this procedure, the data are transformed into a vector with one dimension. Iterative coefficient operations are executed on this vector in the fully connected layer. The values derived from these operations are activated by running them through the chosen activation function, resulting in the production of the output value. Within the network, the output layer is responsible for comparing the given value with the label. The error is then computed, and the weights are adjusted using the backpropagation algorithm [24].

3.2. Xception

Xception, introduced by Francois Chollet in 2017, consists of 36 layers [26]. This architecture is an advancement of the conventional CNN model and is considered to be pioneering in the field of deep learning for the classification of images. The fundamental component of Xception is the depthwise separable convolution, which effectively separates the conventional convolution process into two distinct stages: depthwise and pointwise convolutions [27]. Depthwise separable convolutions are a computational technique that integrates depthwise convolution, which involves convolving each filter channel with each input channel, and pointwise convolution, which involves convolving the produced output channels with a 1 × 1 filter [13]. With a parameter count of 22.9 million, the model has been trained to accurately classify 1000 different categories on the ImageNet dataset. By default, it accepts RGB images as input and anticipates an input with the number of dimensions (299,299,3). Furthermore, it is capable of processing images with the dimensions (224,224,3). Figure 1 depicts the architecture of Xception, as described in [28].

3.3. ResNet50

Developed in 2015 by He Kaiming, Sun Jian, and Microsoft Research Asia, the ResNet architecture is a deep neural network that is based on CNNs. In order to tackle the issue of vanishing gradients in deep neural networks, the ResNet architecture was developed. An issue arises when gradients reach inadequate magnitude during backpropagation, therefore impeding the network’s ability to train efficiently. The architecture of ResNet50 has 50 layers, which are composed of convolutional layers, pooling layers, and fully connected layers. The model has undergone pretrained training using the ImageNet dataset, an extensive collection of 1.2 million images divided into 1000 distinct categories. In order to minimize a loss function, such as the cross-entropy between the anticipated output and the actual labels, the ResNet50 architecture can be trained using the backpropagation technique and stochastic gradient descent. This extensive training dataset has established this model as the predominant CNN model in computer vision applications. The architecture of ResNet50 is shown in Figure 2 [29].

3.4. InceptionResNetV2

InceptionResNetV2 is derived from the InceptionV3 and ResNet architectures. The computational cost of this architecture is lower than that of Inception and Resnet. This architecture has integrated Inception with connections, resulting in a total stack of 164 layers. Incorporating residual connections, convolution filters not only decrease the duration of training but also mitigate the occurrence of deep network distortions [30,31]. This architecture is illustrated in Figure 3.

3.5. VGG19

VGG19 was introduced in 2015 by the Visual Geometry Group at Oxford University [32]. Its architecture comprises 19 deep layers, including 16 convolution layers, 5 max pooling layers, and 3 fully connected layers. The architecture has undergone training using the ImageNet dataset to accurately classify 1000 distinct categories, with a total of 143.7 million parameters. The model is designed to process RGB images on input with dimensions of 224 × 224 × 3. Both the convolution and pooling filters have dimensions of 3 × 3 and 2 × 2, respectively. The initial two fully connected layers consist of 4096 neurons, while the last layer comprises 1000 neurons and utilizes a SoftMax activation function. Every layer in this architecture employs the ReLU function for activation. The architectural design of VGG19 is seen in Figure 4.

3.6. MobileNet

MobileNet was developed specifically to optimize efficiency and is especially suitable for operation on embedded or mobile devices. It was officially launched in April 2017 [33]. To minimize the number of features in this model, the base layer is constructed using a depth-separable convolutional architecture, and this architecture effectively minimizes the parameter count. The key distinction between the depth-separable convolutions in the MobileNet architecture and conventional convolutions lies in the utilization of distinct layers for 3 × 3 depth convolution and 1 × 1 point convolution rather than a single 3 × 3 convolution layer. In depth-separable convolution, the convolution operation is performed independently for each input channel. Pointwise convolution is the application of convolution operators in a 1 × 1 dimension to produce convergence across channels. Figure 5 depicts the architectural design of MobileNet.

3.7. InceptionV3

InceptionV3 was initially introduced by Szegedy et al. The architecture has a substantial series of convolution and maximum pooling layers. The topology of the network consists of 48 layers. Factorized convolutions enhance the network’s efficiency by minimizing the parameter count. The neural network employs asymmetric convolution processes. In addition to pooling activities, grid size reduction operations are also executed. These actions facilitate the resolution of bottlenecks [34]. Figure 6 indicates the layers of the InceptionV3 architecture.

3.8. DenseNet201

DenseNet201 was introduced by Huang et al. in 2017 as an extension of the ResNet architecture [35]. It consists of 201 layers, starting with a 7 × 7 convolution filter and then performing a 3 × 3 max pooling operation. The layers are succeeded by 196 convolution layers, embedded with 3 mean pooling layers, and result in an output layer with a fully connected layer consisting of 1000 neurons, namely global mean pooling. This architecture has a parameter count of 20.2 million and is designed to receive RGB images with dimensions of 224 × 224 × 3 as input. The filters are organized into four compact blocks. The layers inside the dense blocks are tightly interconnected, with each layer receiving input from all preceding layers and transmitting its output to the subsequent layers. The architecture of DenseNet201 is seen in Figure 7.

3.9. Differential Privacy

Differential privacy (DP), a widely used privacy-preserving method, was initially proposed by Dwork in 2006 [36]. It aims to present a solution for statistical attacks targeting databases. Differential privacy is a resilient method that protects private information by introducing random noises to query responses. It ensures that the distribution of noisy results closely fits to the distribution of the actual results [37]. The fundamental premise of differential privacy is that statistical information should not be significantly impacted by the existence or absence of any record in a database. Thus, it protects private information against attackers with a background knowledge [9].
( ε ,   δ ) -Differential Privacy: the mechanism M is considered to satisfy ( ε ,   δ ) -differential privacy if it meets the constraint stated in Equation (1), for every output S , among neighboring datasets D 1 and D 2 . The parameters ε and δ established the degree of privacy. ε is the upper level of impact that an individual’s record can have on public information, whereas δ denotes the probability that this upper level does not hold true:
P r [ M D 1 S e ε Pr M D 2 S + δ
For a dataset, D , the mechanism M that answers query f with Gaussian noise is constructed as in Equation (2):
M ( D ) = f D + n o r m a l ( 0 , σ 2 I )

3.10. Evaluation Methods

Evaluating the performance of the proposed framework with transfer learning architectures is an essential process to measure how well the models have acquired the ability to classify test data. The architectures generate predicted labels, which are then compared with the true labels to determine the performance outcomes. This section examines various evaluation methodologies.
a.
Confusion matrix
As a typical practice, the performance evaluation of a classification algorithm involves comparing the actual and predicted values of the classes with the confusion matrix. An analysis of the predicted labels against the actual labels yields four outcomes: true positive (TP), true negative (TN), false positive (FP), and false negative (FN). TP indicates the number of positive class instances that are accurately identified as positive. TN represents the number of instances belonging to negative class instances that are accurately predicted as negative. FP is the count of negative class instances that the algorithm classifies incorrectly as positive. FN represents the number of positive class instances that the algorithm predicts as negative. A confusion matrix for two classes is presented in Figure 8.
b.
Classification metrics
Classification metrics are also used to measure the performance of the models. Metrics such as accuracy, precision, recall, and F1 score are calculated as follows [39].
Accuracy is the ratio of correct predictions to total predictions and is given in Equation (3):
A c c u r a c y = T P + T N T P + T N + F P + F N
Precision is the ratio of true positive class predictions to the sum of true positive and false positive, shown in Equation (4):
P r e s i c i o n = T P T P + F P
Recall is defined as the ratio of correct positive class predictions to the sum of correct positive class predictions and false negative class predictions and is given by Equation (5):
R e c a l l = T P T P + F N
The F1 score is defined as the weighted average of precision and sensitivity and is given by Equation (6):
F 1   S c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
The area under the process characteristic curve, or AUC, is another evaluation metric used in classification problems. The ROC curve plots the TP ratio (TPR) against the FP ratio (FPR). The ROC curve has the TP ratio on the Y-axis and the FP ratio on the X-axis. These ratios are presented in Equation (7) and Equation (8), respectively:
T P R = T P T P + F N
F P R = F P F P + T N

3.11. Dataset

The dataset obtained via the Picture Archiving and Communication System from several hospitals in Dhaka, Bangladesh, was employed in this study. Patients with kidney tumors, cysts, stones, and normal tissue were included. Following the established procedure, whole abdomens and urinary tracts were scanned, and both coronal and axial slices from both contrast and non-contrast scans were used. Next, the Dicom images were carefully chosen individually based on the diagnosis. From these images, a Dicom image stack was created by the data providers, which included the specific area of interest for each radiologic finding. Then the information and metadata of each patient were retrieved from the Dicom images and converted to a lossless jpg image format. Following the conversion process, a radiologist and a medical technician reviewed every image finding to verify the truthfulness of the data [40].
The dataset obtained consists of 12,510 distinct CT images. We selected a subset of 12,400 images, allocating 3709 to cysts, 5031 to normal tissue, 1377 to stones, and 2283 to tumors [40]. Figure 9 shows CT kidney images of normal, cyst, stone, and tumor. Normal kidneys are characterized by a smooth cortex and medulla. Kidney cysts appear on CT images as round or oval-shaped lesions, usually dark in color and with sharp edges. Kidney stones appear as dense masses, usually bright white on CT images. The size and location of the stone determines the severity of symptoms. Kidney tumors may appear different in density from normal kidney tissue. The size, shape, and location of the tumor are also important.

4. The Proposed Framework

This paper introduces a novel and comprehensive framework that offers the dual capability of kidney disease detection and privacy protection. The identification of kidney disease is achieved by the use of transfer learning methods, namely Xception, ResNet50, InceptionResnetV2, MobileNet, VGG19, InceptionV3, and DenseNet201, with fine-tuning. These methods are robust algorithms extensively employed for image classification tasks. Furthermore, the differential privacy paradigm is used to ensure data privacy. Differential privacy can be implemented at several stages of any deep learning system [41]. The proposed framework applies differential privacy to the gradients. Therefore, a differentially private Adam (DP-Adam) optimizer is used in the proposed framework, which is indicated in Figure 10.
Gradient perturbation via differential privacy ensures privacy by releasing a noisy gradient at each iteration while yet allowing for post-processing of data [8]. Developing an optimization algorithm that satisfies differential privacy is based on a differentially private stochastic gradient descent in which differential privacy is applied in the gradient computation phase. In [42], it is emphasized that traditional gradient computation is performed according to the Equation (9), in which g , m , L , x i , and θ t indicate gradients, number of elements in mini-batch, loss function, element in mini-batch, and position, respectively:
g = 1 m i = 1 m θ L ( x i , θ t )
In addition, the update rules in Adam optimizer are given in detail in [42]. After the calculation of gradients, bounded gradients g ¯ are computed to ensure the gradients are bounded by C:
g ¯ = g m a x ( 1 , g 2 C )
Now, it is time to calculate noisy gradients by adding noise picked from a distribution mechanism:
g ~ = g ¯ + z
In Equation (11), z indicates a vector of random variables. Finally, the noisy gradient g ~ satisfies differential privacy.
In this framework, input image is preprocessed and then transfer learning algorithms are applied for feature extraction. Then the features are flattened, and dense layers which consist of neural networks are applied for the classification task. Dense layers consist of three layers containing 64, 32, and 4 neurons with ReLU and Softmax activators. The DP-Adam optimizer algorithm optimizes the parameters of the neural networks by respecting privacy. Then training and validation loss are compared and then this process iterates for N epochs. Finally, the outputs are obtained.
Some details about the proposed framework shown in Figure 10 are given below:
  • Image input: the input image is read for processing;
  • Image preprocessing: the read image is firstly resized to 100 × 100, normalized, and then split for 70% training and 30% testing;
  • Feature extraction: the features of the images are extracted by using a transfer learning algorithm;
  • Flattening: the extracted features are flattened to a one-dimensional array;
  • Creation of dense layers: three dense layers are used to create a neural network architecture with the number of neurons—64, 32, and 4. ReLU activator is used in the first two layers, and SoftMax activator is employed for the last layer;
  • Compiling model: the model is compiled in order to fit the model;
  • Differentially private optimizer: the optimizer is made differentially private in order to provide privacy;
  • Loss comparison: training and validation loss are compared;
  • Iteration: in order to optimize the model weights, the process is iterated;
  • Output: the classification results are obtained.
The network structure of the proposed framework comprises three dense layers after the extraction of the features with fine-tuned transfer learning architectures. Dense1, Dense2, and Dense3 layers include 64, 32, and 4 neurons, respectively. The differentially private classification task is performed by adding Gaussian noise to the raw gradients of the Adam optimizer. With the help of this process, noisy gradients are obtained for training of the model.

5. Experimental Studies

Accurate identification and classification of CT kidney images as cysts, normal, stones, and tumors is particularly crucial in the healthcare sector. Effective implementation of this procedure is crucial in guaranteeing that patients obtain an early diagnosis and accurate and fast medical care. Thus, the proposed framework facilitates the use of transfer learning methods which rely on CNN to accurately classify medical images. The employed CT images in experimental studies are classified using both the traditional approach and the privacy-preserved approach with the proposed framework.
The experiments were carried out on a computer equipped with 16 GB of RAM, a Xenon 3.4 GHz processor, and running the Windows 11 operating system. Python 3.11, TensorFlow 2.16, as well as other libraries, including numpy, pandas, opencv, and matplotlib were used.

5.1. Preprocessing of Dataset

The dataset comprised 12,400 CT kidney images. Among these images, 3709 depicted cysts, 5031 depicted normal tissue, 1377 depicted stones, and 2283 depicted tumors. During the data preprocessing stage, the file paths of the images associated with each class were identified and lists of these files were generated. Next, every image was appended to a list and assigned the appropriate class label. To accomplish this, distinct labels were assigned to cyst, normal, stone, and tumor instances, respectively. Within the image processing phase, every image was read and scaled to dimensions of 100 × 100 pixels. Subsequently, the pixel values were standardized to the range of values between 0 and 1. This procedure organizes the digital representation of the images for subsequent processing. It was configured to assign 70% of the dataset for training and the remaining 30% for testing.

5.2. Experimental Results

This section evaluates and compares the performance of seven distinct CNN architectures, namely Xception, ResNet50, InceptionResNetV2, MobileNet, DenseNet201, InceptionV3, and VGG19, in accurately classifying CT kidney images into cyst, normal, stone, and tumor classes.
The architectures underwent training using same training dataset and hyperparameters, and their performance was evaluated using comparable test datasets. Strict consistency was maintained throughout the training procedure to ensure fair comparison of the model performance in both training and testing stages. DP-Adam was utilized as the optimizer for privacy-aware classification, whereas Adam was used for traditional classification tasks. The parameters assigned to the learning rate, batch size, number of epochs, number of microbatch, l2 norm clip, and number of epochs are 0.0001, 40, 100, 40, 1, and 100, respectively. The loss function utilized was categorical cross-entropy, whereas the activation functions applied were ReLU and Softmax. Application of the trained model enabled the generation of predictions on the test dataset, which were then compared with the true labels. Consequently, the classification performance of each model in distinguishing kidney cysts, normal tissues, stones, and tumors was assessed.
The performance results of Xception, ResNet50, InceptionResnetV2, MobileNet, VGG19, InceptionV3, and DenseNet201 architectures were evaluated and described, both with and without privacy modifications. Presented in Table 2 are the traditional classification results, whereas Table 3, Table 4 and Table 5 shows the test results of the proposed framework with different noise_multiplier (nm) and δ results. Generally, the value of δ is determined as 1 × 10−5 in the literature. Therefore, throughout the experiments, we set δ to 1 × 10−5. The images presented in Figure 11, Figure 12, Figure 13, Figure 14, Figure 15 and Figure 16 were created according to the best results obtained, as in Table 3. Figure 11, Figure 12 and Figure 13 provide the confusion matrix for each class, the accuracy rates of the training and test data, and the ROC curves for traditional classification. Given that Table 3 yielded superior outcomes in the studies, Figure 14, Figure 15 and Figure 16 were generated based on the values indicated in Table 3. Presented in Figure 14, Figure 15 and Figure 16 are the confusion matrices for each class, the accuracy rates of the training and test data, and the ROC curves for privacy-aware classification via the proposed framework. Among the architectures included in Table 1, VGG19 outperformed other architectures in traditional classification tasks. Furthermore, alternative algorithms also yielded robust outcomes for classification in terms of accuracy, recall, F-1, and precision.
According to the results in Table 3, Table 4 and Table 5, the proposed framework tests demonstrate that MobileNet achieved a remarkable accuracy ratio of 99.83% in the classification CT images. The best results were obtained with settings of nm = 0.1, ε = 772,577, and δ = 1 × 10−5. In addition, the other architectures similarly achieved satisfactory high precision, recall, F-1, and accuracy outcomes. The results demonstrate that the proposed framework facilitates the effective performance of transfer learning algorithms while maintaining privacy aspects. It can be clearly seen that when nm increases and ε decreases, the accuracy decreases.

6. Conclusions and Discussion

Technological advancements in medical imaging techniques have played a vital role in the early detection and management of illnesses. However, this process is vulnerable to human error since the task of visual interpretation is usually performed by professionals, who may not always be easily accessible. Hence, the use of automated technologies is always necessary to assist doctors in making diagnostic interpretations. However, in the literature, it is known that disease data are highly sensitive, and privacy measures should be taken to protect the privacy of patients against disclosure attacks. Disclosure of such kinds of data may negatively affect data owners, such as via discrimination, stigmatization, or financial harm. Hence privacy measures are always required. On the other hand, implementing privacy measures enables trust to be established between patients and healthcare providers. Thus, any useful process over these kinds of data, such as early diagnosis, treatment planning, etc., may be performed easily without considering if private data are disclosed. It is essential to use privacy-preserving measures for extracting important information from such data.
The objective of this study was to provide a novel framework that considers the accurate classification of kidney diseases while ensuring data privacy. In the proposed framework, transfer learning was employed for feature extraction, and differential privacy was used to obtain noisy gradients. An evaluation of the proposed framework’s efficiency was conducted by comparing the performance of seven distinct transfer learning architectures: Xception, ResNet50, InceptionResNetV2, MobileNet, DenseNet201, InceptionV3, and VGG19. Every architecture was assessed using accuracy, recall, F-1 score, and precision criteria. The acquired findings demonstrate the efficiency of each architecture in classification tasks.
The experimental findings indicate that with nm = 0.1, ε = 772,577, and δ = 1 × 10−5, the MobileNet model outperformed the other benchmark models in privacy-sensitive diagnosis of kidney diseases, with an accuracy of 99.83%. This model has exceptional performance characterized by excellent accuracy and well-balanced metrics. In addition, it can be seen from Table 3, Table 4 and Table 5 that the accuracy decreased when ε decreased. This is the expected situation in the domain of differential privacy. The results indicate that the proposed framework is applicable for the early detection of kidney diseases and presents promising privacy preserving outcomes in this domain.
While this paper has primarily focused on the detection of kidney disease with respect to privacy, it should be noted that the proposed framework can be generalized for other types of medical imaging and transfer learning tasks. With the help of appropriate preprocessing approaches and training processes, other types of medical imaging data, such as X-rays, MRI, and PET, can be easily used in the proposed framework. In addition, other transfer learning techniques can be easily employed and implemented in the feature extraction process.
In conclusion, the proposed framework facilitated transfer learning architectures to achieve robust performance in classifying CT kidney images while considering data privacy. The use of such a high-performance framework can enhance healthcare quality by facilitating precise, fast, and private disease diagnosis. Thus, the findings demonstrate that deep learning models can be properly employed in medical diagnostic procedures. Implementing CNN models in CT kidney images significantly enhances and simplifies the precise identification of cysts, tumors, stones, and healthy tissues. Improved quality of healthcare services is achieved by guaranteeing prompt and precise treatment for patients.

Author Contributions

Conceptualization, Y.C. and P.C.; methodology, Y.C. and P.C.; software, Y.C., P.C. and S.A; validation, Y.C. and P.C.; investigation, Y.C., P.C. and S.A.; writing—original draft preparation, Y.C., P.C. and S.A; writing—review and editing, Y.C. and P.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kovesdy, C.P. Epidemiology of chronic kidney disease: An update 2022. Kidney Int. Suppl. 2022, 12, 7–11. [Google Scholar] [CrossRef] [PubMed]
  2. Foreman, K.J.; Marquez, N.; Dolgert, A.; Fukutaki, K.; Fullman, N.; McGaughey, M.; Pletcher, M.A.; Smith, A.E.; Tang, K.; Yuan, C.-W. Forecasting life expectancy, years of life lost, and all-cause and cause-specific mortality for 250 causes of death: Reference and alternative scenarios for 2016–40 for 195 countries and territories. Lancet 2018, 392, 2052–2090. [Google Scholar] [CrossRef]
  3. Eker, A.G.; Duru, N. Medikal görüntü işlemede derin öğrenme uygulamaları. Acta Infol. 2021, 5, 459–474. [Google Scholar]
  4. Daniel, A.J.; Buchanan, C.E.; Allcock, T.; Scerri, D.; Cox, E.F.; Prestwich, B.L.; Francis, S.T. Automated renal segmentation in healthy and chronic kidney disease subjects using a convolutional neural network. Magn. Reson. Med. 2021, 86, 1125–1136. [Google Scholar] [CrossRef]
  5. Pavithra, D.; Vanithamani, R. Chronic kidney disease detection from clinical data using CNN. In Proceedings of the 2021 IEEE International Conference on Distributed Computing, VLSI, Electrical Circuits and Robotics (DISCOVER), Nitte, India, 19–20 November 2021; pp. 282–285. [Google Scholar]
  6. Hossain, M.S.; Hassan, S.N.; Al-Amin, M.; Rahaman, M.N.; Hossain, R.; Hossain, M.I. Kidney Disease Detection from CT Images using a customized CNN model and Deep Learning. In Proceedings of the 2023 International Conference on Advances in Intelligent Computing and Applications (AICAPS), Kochi, India, 1–3 February 2023; pp. 1–6. [Google Scholar]
  7. Bhaskar, N.; Manikandan, S. A deep-learning-based system for automated sensing of chronic kidney disease. IEEE Sens. Lett. 2019, 3, 1–4. [Google Scholar] [CrossRef]
  8. Ismetoglu, A.; Canbay, Y. Detection of COVID-19 disease by using privacy-aware artificial intelligence system. Secur. Priv. 2024, e434. [Google Scholar] [CrossRef]
  9. Canbay, Y.; Ismetoğlu, A.; Canbay, P. Deep Learning and Data Privacy in Diagnosis of COVID-19. J. Eng. Sci. Des. 2021, 9, 701–715. [Google Scholar]
  10. Avcı, F. Liver Segmentation on Tomography Images with Deep Learning Techniques. Master’s Thesis, Graduate School of Natural and Applied Sciences, Eskisehir Osmangazi University, Eskişehir, Turkey, 2019. [Google Scholar]
  11. Kayhan, B. Deep Learning Based Multi Organ Segmentation in Computed Tomography Images. Master’s Thesis, Graduate School of Natural and Applied Sciences, Konya Technical University, Konya, Turkey, 2022. [Google Scholar]
  12. Katar, O. Deep Learning Based Physician Decision Support System Design for COVID-19 Diagnosis on Computed Tomography Images. Master’s Thesis, Graduate School of Natural and Applied Sciences, Firat University, Elazığ, Turkey, 2022. [Google Scholar]
  13. Al-Areqi, F.; Konyar, M.Z. Effectiveness evaluation of different feature extraction methods for classification of COVID-19 from computed tomography images: A high accuracy classification study. Biomed. Signal Process. Control 2022, 76, 103662. [Google Scholar] [CrossRef]
  14. Karaca, Y.E. Detection of Lung Cancer with Deep Learning Using Computed Tomography Images. Master’s Thesis, Institute of Graduate Studies, Malatya Turgut Ozal University, Malatya, Turkey, 2023. [Google Scholar]
  15. Topçu, K. Prediction of COVID-19 Risk of a Person by Analyzing Computed Tomography Images Using Convolutional Neural Networks. Master’s Thesis, Middle East Technical University, Ankara, Turkey, 2024. [Google Scholar]
  16. Narmada, N.; Shekhar, V.; Singh, T. Classification of Kidney Ailments using CNN in CT Images. In Proceedings of the 2022 13th International Conference on Computing Communication and Networking Technologies (ICCCNT), Kharagpur, India, 3–5 October 2022; pp. 1–5. [Google Scholar]
  17. Mehr, H.D. Classification of Lung Ct Images Using Deep Convolutional Neural Network. Master’s Thesis, Graduate School of Natural and Applied Sciences, Gazi University, Ankara, Turkey, 2017. [Google Scholar]
  18. Wang, J.; Shen, T.; Gu, R.; Jiang, X. CT image liver segmentation based on improved Deeplabv3+. In Proceedings of the 2023 2nd International Conference on Robotics, Artificial Intelligence and Intelligent Control (RAIIC), Mianyang, China, 11–13 August 2023; pp. 376–382. [Google Scholar]
  19. Xu, L.; Mao, X.; Sun, M.; Liu, W.; Wang, Y.; Tang, Y. Lung lesions detection from CT images based on the modified Faster R-CNN. In Proceedings of the 2020 International Conference on Computer, Information and Telecommunication Systems (CITS), Hangzhou, China, 5–7 October 2020; pp. 1–5. [Google Scholar]
  20. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  21. Karayegen, G.; Aksahin, M.F. Brain tumor prediction on MR images with semantic segmentation by using deep learning network and 3D imaging of tumor region. Biomed. Signal Process. Control 2021, 66, 102458. [Google Scholar] [CrossRef]
  22. Utku, A.; Ayaz, Z.; Çiftçi, D.; Akcayol, M.A. Deep Learning Based Classification for Hoverflies (Diptera: Syrphidae). J. Entomol. Res. Soc. 2023, 25, 529–544. [Google Scholar] [CrossRef]
  23. Samuel, S.P.; Malarvizhi, K.; Karthik, S. Weed detection in agricultural fields via automatic graph cut segmentation with Mobile Net classification model. Signal Image Video Process. 2024, 18, 1549–1560. [Google Scholar] [CrossRef]
  24. Kahraman, G.; Civelek, Z. Diagnosing Covid-19 Disease from Computed Tomography Images with Deep Learning and Machine Learning. Int. J. Eng. Res. Dev. 2023, 15, 49–63. [Google Scholar] [CrossRef]
  25. Yu, D.; Wang, H.; Chen, P.; Wei, Z. Mixed pooling for convolutional neural networks. In Proceedings of the Rough Sets and Knowledge Technology: 9th International Conference, RSKT 2014, Shanghai, China, 24–26 October 2014; Proceedings 9. pp. 364–375. [Google Scholar]
  26. Chollet, F. Xception: Deep learning with depthwise separable convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1251–1258. [Google Scholar]
  27. Hu, H.; Li, S.; Huang, J.; Liu, B.; Che, C. Casting product image data for quality inspection with xception and data augmentation. J. Theory Pract. Eng. Sci. 2023, 3, 42–46. [Google Scholar] [CrossRef]
  28. Thepade, S.D.; Dindorkar, M.R. Fusing deep convolutional neural network features with Thepade’s SBTC for land usage identification. Eng. Sci. Technol. Int. J. 2022, 27, 101014. [Google Scholar] [CrossRef]
  29. Shinde, S.; Kulkarni, U.; Mane, D.; Sapkal, A. Deep learning-based medical image analysis using transfer learning. In Health Informatics: A Computational Perspective in Healthcare; Springer: Singapore, 2021; pp. 19–42. [Google Scholar]
  30. Elaraby, A.; Saad, A.; Karamti, H.; Alruwaili, M. An Optimized Deep Learning Approach for Robust Image Quality Classification. Trait. Signal 2023, 40, 1573–1579. [Google Scholar] [CrossRef]
  31. Kaur, G.; Singh, S. An ensemble based approach for violence detection in videos using deep transfer learning. Multimed. Tools Appl. 2024, 1–25. [Google Scholar] [CrossRef]
  32. Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  33. Wang, W.; Li, Y.; Zou, T.; Wang, X.; You, J.; Luo, Y. A novel image classification approach via dense-MobileNet models. Mob. Inf. Syst. 2020, 2020, 7602384. [Google Scholar] [CrossRef]
  34. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer vision. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
  35. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  36. Dwork, C. Differential Privacy. In International Colloquium on Automata, Languages and Programming; Springer: Berlin/Heidelberg, Germany, 2006; pp. 1–12. [Google Scholar]
  37. Kifer, D.; Machanavajjhala, A. No free lunch in data privacy. In Proceedings of the 2011 ACM SIGMOD International Conference on Management of Data, Athens, Greece, 12–16 June 2011; pp. 193–204. [Google Scholar]
  38. Ramadhan, A.A.; Baykara, M. A novel approach to detect COVID-19: Enhanced deep learning models with convolutional neural networks. Appl. Sci. 2022, 12, 9325. [Google Scholar] [CrossRef]
  39. Kaya, Y.; Yiner, Z.; Kaya, M.; Kuncan, F. A new approach to COVID-19 detection from X-ray images using angle transformation with GoogleNet and LSTM. Meas. Sci. Technol. 2022, 33, 124011. [Google Scholar] [CrossRef]
  40. Islam, M.N.; Hasan, M.; Hossain, M.K.; Alam, M.G.R.; Uddin, M.Z.; Soylu, A. Vision transformer and explainable transfer learning models for auto detection of kidney cyst, stone and tumor from CT-radiography. Sci. Rep. 2022, 12, 11440. [Google Scholar] [CrossRef] [PubMed]
  41. Mireshghallah, F.; Taram, M.; Vepakomma, P.; Singh, A.; Raskar, R.; Esmaeilzadeh, H. Privacy in deep learning: A survey. arXiv 2020, arXiv:2004.12254. [Google Scholar]
  42. Gylberth, R.; Adnan, R.; Yazid, S.; Basaruddin, T. Differentially private optimization algorithms for deep neural networks. In Proceedings of the 2017 International Conference on Advanced Computer Science and Information Systems (ICACSIS), Jakarta, Indonesia, 28–29 October 2017; pp. 387–394. [Google Scholar]
Figure 1. Architecture of Xception.
Figure 1. Architecture of Xception.
Applsci 14 08629 g001
Figure 2. Architecture of ResNet50.
Figure 2. Architecture of ResNet50.
Applsci 14 08629 g002
Figure 3. Architecture of InceptionResnetV2.
Figure 3. Architecture of InceptionResnetV2.
Applsci 14 08629 g003
Figure 4. Architecture of VGG19.
Figure 4. Architecture of VGG19.
Applsci 14 08629 g004
Figure 5. Architecture of MobileNet.
Figure 5. Architecture of MobileNet.
Applsci 14 08629 g005
Figure 6. Architecture of InceptionV3.
Figure 6. Architecture of InceptionV3.
Applsci 14 08629 g006
Figure 7. Architecture of DenseNet201.
Figure 7. Architecture of DenseNet201.
Applsci 14 08629 g007
Figure 8. Confusion matrix [38].
Figure 8. Confusion matrix [38].
Applsci 14 08629 g008
Figure 9. CT examples for cyst, stone, normal, and tumor.
Figure 9. CT examples for cyst, stone, normal, and tumor.
Applsci 14 08629 g009
Figure 10. The proposed framework.
Figure 10. The proposed framework.
Applsci 14 08629 g010
Figure 11. Confusion matrix of traditional classification.
Figure 11. Confusion matrix of traditional classification.
Applsci 14 08629 g011
Figure 12. Accuracy graph of traditional classification.
Figure 12. Accuracy graph of traditional classification.
Applsci 14 08629 g012
Figure 13. ROC curves of traditional classification.
Figure 13. ROC curves of traditional classification.
Applsci 14 08629 g013
Figure 14. Confusion matrix of privacy-preserved classification.
Figure 14. Confusion matrix of privacy-preserved classification.
Applsci 14 08629 g014
Figure 15. Accuracy graph of privacy-preserved classification.
Figure 15. Accuracy graph of privacy-preserved classification.
Applsci 14 08629 g015
Figure 16. ROC curves of privacy-preserved classification.
Figure 16. ROC curves of privacy-preserved classification.
Applsci 14 08629 g016
Table 1. Comparison of previous studies.
Table 1. Comparison of previous studies.
StudyMain
Architecture
Data TypeFocusConsider PrivacyDatasetNumber of RecordsAccuracy
(%)
[10]CNNCTLiverNoCombined Healthy Abdominal Organ Segmentation257997.58
[11]CNNCTMultiple OrgansNoMICCAI 2015 Multi-Atlas Abdomen Labeling Challenge300088.8
[12]CNNCTChestNoEFSCH-19400099.75
[13]CNNCTChestNoCT Scans for COVID-19 Classification19,68599.62
[14]CNNCTLungNoIQ-OTH/NCCD119092.69
[15]CNNCTChestNoCOVID-CT Dataset74681
[16]CNNCTKidneyNoPACS Systems in Dhaka, Bangladesh12,51099.36
[17]CNNCTLungNoData Science Bowl and Kaggle Lung CT Scan Images285,05896.36
[18]CNNCTLiverNo3Dircadb Dataset207495.4
[6]CNNCTKidneyNoPACS Systems in Dhaka, Bangladesh12,51098.66
[19]CNNCTLungNoOriginal26368.12
This StudyCNNCTKidneyYesPACS Systems in Dhaka, Bangladesh12,40099.83
Table 2. Traditional classification results (without privacy).
Table 2. Traditional classification results (without privacy).
ArchitectureAccuracyRecallF1Precision
Xception0.99650.99650.99650.9965
ResNet500.99780.99780.99780.9978
InceptionResNetV20.99540.99540.99540.9954
VGG191111
MobileNet0.99970.99970.99970.9997
DenseNet2010.99970.99970.99970.9997
InceptionV30.99050.99050.99050.9905
Table 3. Privacy-aware classification results with the proposed framework (nm = 0.1, ε = 772,577, and δ = 1 × 10−5).
Table 3. Privacy-aware classification results with the proposed framework (nm = 0.1, ε = 772,577, and δ = 1 × 10−5).
ArchitectureAccuracyRecallF1Precision
Xception0.97360.97360.97340.9735
ResNet500.75910.75910.71480.7347
InceptionResNetV20.95370.95370.95340.9533
VGG190.94000.94000.93810.9388
MobileNet0.99830.99830.99830.9983
DenseNet2010.98750.98750.98750.9875
InceptionV30.93920.93920.93830.9391
Table 4. Privacy-aware classification results with the proposed framework (nm = 0.2, ε = 4625, and δ = 1 × 10−5).
Table 4. Privacy-aware classification results with the proposed framework (nm = 0.2, ε = 4625, and δ = 1 × 10−5).
ArchitectureAccuracyRecallF1Precision
Xception0.94940.94940.94890.9489
ResNet500.70510.70510.64310.6925
InceptionResNetV20.92230.92230.92090.9210
VGG190.91040.91020.91030.9102
MobileNet0.99110.99110.99110.9911
DenseNet2010.98480.98480.98480.9848
InceptionV30.89970.89970.89680.8998
Table 5. Privacy-aware classification results with the proposed framework (nm = 0.3, ε = 290, and δ = 1 × 10−5).
Table 5. Privacy-aware classification results with the proposed framework (nm = 0.3, ε = 290, and δ = 1 × 10−5).
ArchitectureAccuracyRecallF1Precision
Xception0.92710.92710.92490.9265
ResNet500.66720.66720.58290.6399
InceptionResNetV20.90330.90350.90340.9033
VGG190.89470.89470.89470.8944
MobileNet0.98680.98680.98670.9868
DenseNet2010.97810.97810.97820.9781
InceptionV30.86720.86720.86760.8676
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Canbay, Y.; Adsiz, S.; Canbay, P. Privacy-Preserving Transfer Learning Framework for Kidney Disease Detection. Appl. Sci. 2024, 14, 8629. https://doi.org/10.3390/app14198629

AMA Style

Canbay Y, Adsiz S, Canbay P. Privacy-Preserving Transfer Learning Framework for Kidney Disease Detection. Applied Sciences. 2024; 14(19):8629. https://doi.org/10.3390/app14198629

Chicago/Turabian Style

Canbay, Yavuz, Seyda Adsiz, and Pelin Canbay. 2024. "Privacy-Preserving Transfer Learning Framework for Kidney Disease Detection" Applied Sciences 14, no. 19: 8629. https://doi.org/10.3390/app14198629

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop