Next Article in Journal
Polar Code Parameter Recognition Algorithm Based on Dual Space
Next Article in Special Issue
Defect Synthesis Using Latent Mapping Adversarial Network for Automated Visual Inspection
Previous Article in Journal
An Overview of Technologies for Improving Storage Efficiency in Blockchain-Based IIoT Applications
Previous Article in Special Issue
AI-Based Resource Allocation Techniques in Wireless Sensor Internet of Things Networks in Energy Efficiency with Data Optimization
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Perceptual Encryption-Based Image Communication System for Deep Learning-Based Tuberculosis Diagnosis Using Healthcare Cloud Services

Department of Computer Engineering, Chosun University, Gwangju 61452, Korea
*
Author to whom correspondence should be addressed.
Electronics 2022, 11(16), 2514; https://doi.org/10.3390/electronics11162514
Submission received: 14 July 2022 / Revised: 5 August 2022 / Accepted: 9 August 2022 / Published: 11 August 2022

Abstract

:
Block-based perceptual encryption (PE) algorithms are becoming popular for multimedia data protection because of their low computational demands and format-compliancy with the JPEG standard. In conventional methods, a colored image as an input is a prerequisite to enable smaller block size for better security. However, in domains such as medical image processing, unavailability of color images makes PE methods inadequate for their secure transmission and storage. Therefore, this study proposes a PE method that is applicable for both color and grayscale images. In the proposed method, efficiency is achieved by considering smaller block size in encryption steps that have negligible effect on the compressibility of an image. The analyses have shown that the proposed system offers better security with only 12% more bitrate requirement as opposed to 113% in conventional methods. As an application of the proposed method, we have considered a smart hospital that avails healthcare cloud services to outsource their deep learning (DL) computations and storage needs. The EfficientNetV2-based model is implemented for automatic tuberculosis (TB) diagnosis in chest X-ray images. In addition, we have proposed noise-based data augmentation method to address data deficiency in medical image analysis. As a result, the model accuracy was improved by 10%.

1. Introduction

Cloud services provide a cost-effective solution to meet the Information and Communication Technology (ICT) needs of an organization. The organization can use ICT resources, services and software of a Cloud Services Provider (CSP) via the internet without a necessity of internal infrastructure or hardware on-site installations. With the recent success of Machine Learning (ML) in the field of computer vision, automatic computer aided diagnosis (CAD) systems have emerged in healthcare organizations to assist doctors and practitioners. Particularly, Deep Learning (DL), a subfield of ML, has achieved state-of-the-art performance for image classification [1]. However, DL models are compute-intensive tasks, and their training requires cutting-edge technology and high computational resources. In this regard, healthcare organizations can avail cloud-computing services to access the latest technology in order to speed up the training process and allow DL models to scale efficiently with a lower capital cost [2,3]. In addition, training DL models requires a large volume of sample data, which in some cases such as the medical domain, is expensive and difficult to acquire. To overcome this issue, healthcare organizations can benefit from a community cloud, where services are shared by organizations with common interests. In this case, cloud storage services can be used as a shared central data repository for joint projects and collaboration among the organizations. However, like all communication systems, when data are outsourced for cloud services, there is always a risk of information leakage and a large volume of data requires high bandwidth [4,5,6,7].
Compression and encryption are two processes that satisfy the dual requirements of data transmission over bandwidth constraint and public channels. Image compression gives a compact representation to an image such that it requires less number of bits. It can be achieved either in lossless or lossy mode. In lossless compression, an image can be recovered with almost the same quality as that of the original image, whereas in lossy mode the image quality degrades. Compared to lossless mode, lossy compression offers better savings; however, resulting quality degradation in lossy mode may not be acceptable in certain domains. For example, medical images contain information crucial for correct diagnosis of diseases; therefore, their compression should be carried out in such a way that the diagnostic information remains intact in them while their sizes are reduced [8,9,10]. One of the popular approaches to achieve this goal is to compress the region-of-interest (ROI) necessary for diagnosis in lossless mode and non-ROI in lossy mode [8,9,10]. Such methods can achieve a significant reduction in the image size while preserving its important details. However, they require segmentation of an image beforehand, which is computationally expensive and is a target task to be performed using cloud-computing resources. Therefore, ROI-based methods are not suitable for efficient image data transmission [2].
Encryption makes image data unintelligible, which can only be recovered by its inverse decryption process. The number theory and chaos theory-based encryption algorithms are proven efficient for securing image data [2]. These conventional encryption algorithms perform stream encryption and/or scrambling of pixel values; however, they are only suitable for encrypting raw images. For example, the JPEG compressed image consists of format markers and any changes in them by an external operation will leave the image uninterpretable. Similarly, re-encoding a cipher image as a JPEG image results in file size increment. Different from other form of data, encryption of image data can be carried out only by disrupting their intrinsic properties. For example, changing pixel correlation and/or redundancy in an image can result in an unintelligible image with a necessary level of security. Based on this observation, a new class of encryption algorithms has been emerged called Perceptual Encryption (PE) algorithms to meet the aforementioned requirements of encrypting compressed images. The main idea is to reverse the conventional order of performing compression prior to encryption. PE performs block-based operations that hides only perceptual information of an image, thereby preserves image intrinsic properties necessary to carry out computations in the encryption domain. For example, refs. [11,12] proposed PE methods for enabling privacy-preserving DL applications. In addition, PE cipher images are JPEG compressible, which makes them suitable for numerous applications, such as cloud photo storage and social networking services [13,14] and image retrieval in the encryption domain [15]. Nonetheless, PE methods are resilient against various attacks, including brute-force and cipher-text-only attacks [16].
Based on an input image representation, PE methods can be grouped as Color-PE and Grayscale-PE methods. The Color-PE represents an input color image as a three-component image and uses same encryption keys for each component [17], whereas their extended versions encrypt each color component independently [12,18]. The latter methods have larger keyspace as they have increased number of blocks. However, this increment is limited by the smallest allowable block size in the JPEG algorithm, for instance, block size no smaller than 16 × 16 should be used for color image compression. This recommended size is necessary to avoid block artifacts resulted from the JPEG chroma-subsampling step [2]. Smaller block size such as 8 × 8, can be utilized in the JPEG algorithm without any adverse effect, for compression of grayscale images. Therefore, to exploit the smaller block size for an expanded keyspace, Grayscale-PE represents color input as a pseudo-grayscale image by combining the color components along the horizontal or vertical direction [13,14]. Overall, in conventional methods, color image as an input is a prerequisite for better security. However, in domains such as medical image processing, the unavailability of color images makes the conventional PE methods inadequate for their secure transmission and storage. Therefore, the current study proposes a PE method that is applicable for both color and grayscale images. In the proposed method, efficiency is achieved by considering smaller block size in encryption steps that have smaller effect on compressibility of an image, and, importantly, the processing does not compromise quality of the recovered images. As an application of the proposed method, we have considered a smart hospital that avail healthcare cloud services to outsource their DL computations and data storage needs. The preliminary results of this work were presented in [2].
The main contributions of the present study are summarized as: (1) proposed a PE algorithm for secure and efficient transmission and/or storage of medical images; (2) a DL-based solution is implemented for automatic Tuberculosis (TB) screening in chest X-ray (CXR) images; (3) analysis of the proposed DL model against distortions resulted from compression process; (4) proposed noise-based augmentation method to improve generalization of DL model on smaller dataset; (5) the analysis comprised of encryption, compression and DL-based classification were carried out on three datasets.

2. Related Work

2.1. Deep Learning-Based Tuberculosis Screening

Grivkov et al. [19] implemented InceptionNetV3 [20] for diagnosis of TB in Shenzhen (SH) and Montgomery (MG) datasets [21] and achieved 86.8% accuracy. Das et al. [22] exploited transfer learning to improve InceptionV3 accuracy to 91.7% on the same datasets. Priya et al. [23] implemented transfer learning on VGG19 [24], ResNet50 [25], DenseNet121 [26] and InceptionV3 models. In their analysis, pre-trained VGG19 has achieved 89% and 95% best accuracies on MG and SH datasets, respectively. Cao et al. [27] implemented DenseNet121, VGG and ResNet152 [25] models and achieved best accuracy of 90.38% classification accuracy with DenseNet121. Raman et al. [28] adopted somewhat different approach than the aforementioned methods. They have used three pre-trained models (ResNet101 [25], VGG19, and DenseNet201 [26]) to extract features from CXR images and use eXtreme Gradient Boosting (XG-Boost) (1.6.1, Tianqi Chen and Carlos Ernesto Guestrin, Seattle, WA, USA) [29] model to classify TB and non-TB in them. In their experiments, DenseNet201 with XG-Boost architecture achieved the highest accuracy of 99.92% as compared to its counterparts. Munadi et al. [30] proposed to enhance CXR quality before feeding them to pre-trained ResNet and EfficientNet [31] models. They have used three different image-enhancing techniques (unsharped masking, high-frequency emphasis filtering, and contrast limited adaptive histogram equalization). In their analysis, EfficientNet with unsharped masking image enhancement achieved 89.92% accuracy on SH dataset. Msnoda et al. [32] implemented ResNet, GoogLeNet [33], and AlexNet [34] with an extra Spatial Pyramid Pooling (SPP) [35] layer. Among the implemented architectures, GoogLeNet achieved the highest classification accuracy of 97%, which was then improved to 98% by using the SPP layer.
The methods discussed so far rely on the architecture of an individual model for classification efficiency. There are methods that combine two or even more models to form an ensemble network to achieve better performance. For example, Rajaraman et al. [36] implemented VGG16, InceptionResNetV2 [37], InceptionV3, XceptionNet [38] and DenseNet121, and then ranked them based on their accuracy. In their experiments, the top-3 models were InceptionV3 (accuracy = 94%), DenseNet121 (accuracy = 92.8%) and InceptionResNetV2 (accuracy = 92.5%). They have evaluated multiple ensemble methods to combine the top-3 models such as majority voting, simple averaging, weighted averaging stacking and blending to make an ensemble network. Their analysis showed that stacking ensemble demonstrated better performance and achieved 94.1% accuracy. Dasanayaka et al. [39] have implemented an ensemble of only two models (VGG16 and InceptionV3), and achieved 97.10% accuracy, which is higher than the ensemble of the three models proposed in [36]. Oloko-Oba et al. [40] have implemented an ensemble of VGG16, ResNet50 and InceptionV3 and achieved best accuracy of 96.14%. In their other study [41], they have explored ensemble of EfficientNets [31] for the diagnosis of TB. In their analysis of individual models, EfficientNet-B4 achieved best accuracy of 94.35% on SH dataset, which was then improved to 97.44% through ensemble learning. The ensemble was built by averaging the performance of three best individual EfficientNets (B2, B3, and B4). Saif et al. [42] proposed to combine the traditional hand-engineered feature with an ensemble of DenseNet169, ResNet50 and InceptionV3 models. Their ensemble model has achieved best accuracy of 99.7% on SH dataset. Overall, ensemble methods have shown superior performance for TB screening in CXR images than the individual models.

2.2. Perceptual Encryption Methods

The PE algorithm is block-based and performs four steps: blocks permutation, rotation and inversion, negative and positive transformation, and color channel shuffling. Based on these steps, several methods have been proposed in literature. They can be classified as Color-PE and Grayscale-PE methods based on their input image representation. In Color-PE methods, an input color image is represented as a three-component image, whereas Grayscale-PE methods represent an input as pseudo-grayscale image by concatenating its color components along the vertical or horizontal direction. In Grayscale-PE methods, the channel-shuffling step is omitted. This section provides a summary of PE related work.
Kurihara et al. [17] proposed a block-based Color-PE method that performs the encryption steps on each color component by using the same key. Since, the input is a color image, larger block size is used to avoid block artifacts in a decoded image resulted by chroma subsampling of the JPEG algorithm. However, the use of the same key for each color component and larger block size result in a smaller number of blocks, which make the scheme vulnerable to jigsaw puzzle attack. To increase the number of blocks for better security, Imaizumi et al. [43] proposed to perform the first three steps of encryption independently in each color component. As a result, the scheme has a larger key space than that of [17]; however, processing each component individually results in the JPEG compatibility issues. For example, the method is only applicable with the JPEG lossless algorithm only when using RGB colorspace. Ahmad et al. [12,18] proposed a PE method to deal with the compatibility issue of [43]. In their proposed schemes, rotation, inversion and pixel values transformations are performed on each color component independently. The use of a same key for permutation step in each component allows the JPEG algorithm with YCbCr colorspace for better compression savings. The extended PE methods in [12,18,43] have better security than the PE method proposed in [17] as the keyspace is expanded and color distribution is altered significantly. However, the main limitation of extended PE methods is that they cannot exploit the JPEG chroma subsampling.
An alternative approach has been adopted by Chuman et al. [13] that allows the use of a smaller block size. The main idea is to represent an input color image as a pseudo grayscale image by concatenating the color components along the horizontal or vertical direction; therefore, belongs to Grayscale-PE methods. Such representation allows use of a smaller block size without any adverse effect on the decoded image quality and compression savings. In addition to smaller block size, lack of color information improves robustness against jigsaw puzzle solver attack. When chroma subsampling is desirable, Sirichotedumrong et al. [14] proposed to convert input image from RGB to YCbCr colorspace, perform down sampling on the color components and then concatenate them with the luminance component. For example, to combine them horizontally, the color components must be concatenated vertically and vice versa.
Compared to Color-PE methods and their extensions, Grayscale-PE methods provide better security as their use of smaller block size increases the number of blocks and pseudo grayscale representation efficiently disrupts the color information. However, for block-based PE schemes, there is an efficiency tradeoff between compression and encryption because of the block size choice. For example, a block size of no smaller than 16 × 16 and 8 × 8 should be used in Color-PE and Grayscale PE methods, respectively.

3. Methods

Block-based compressible perceptual encryption (PE) methods proposed in [12,13,14,17,18,43] mainly consist of the following three steps as shown in Figure 1:
Step 1. Input image representation.
Step 2. Block-based image transformations such as geometric and color transformations. A geometric transformation changes position as well as orientation of a block while color transformation modifies pixel values in a block.
Step 3. PE results in an encrypted image that preserves intrinsic properties of an image necessary for compression. Therefore, the last step of compressible PE algorithms is to carry out compression by using image standards such as the JPEG.
Several PE methods have been proposed which can be grouped based on the input image representation as color-PE methods and grayscale-PE methods. The following sub-sections describe conventional and proposed PE methods.

3.1. Color-PE Methods

A compressible PE method is proposed in [12,17,18,43], which gives a user control over their images data when sharing them via internet. In conventional color-PE schemes [17], an input image is represented as a true color image with three-color components, which then undergoes the following steps to obtain an encrypted image I e .
Step 1. Divide an image I with W × H pixels into non-overlapping blocks of size B w × B h , and shuffle them by using a secret key K 1 generated randomly.
Step 2. Change orientation of each block in the shuffled image by performing rotation and inversion on them. The transformation is controlled by randomly generated key K 2 .
Step 3. Change pixel values in each block by negative-positive transformation decided by a uniformly distributed binary key K 3 . The transformation is carried out as
p ´ j , k = p j , k K 3 i = = 0 p j , k 2 L 1 K 3 i = = 1 ,
where p ´ j , k is the modified value of a pixel p j , k in original image, 0 j w , 0 k h , 0 i N   with N being total number of blocks and L is number of bits required to represent highest intensity level in the original image.
Step 4. Shuffle three-color components in each block by using key K 4 where each entry represents a unique permutation of the color components.
Since, color-PE methods represent an input image as color; therefore, for each component red (R), green (G) and blue (B), encryption consists of four keys as K i   ϵ   K i R , K i G , K i B where i 1 ,   2 ,   3 ,   4 . The method described above uses the same key in each step as K i R = K i G = K i B for i 1 ,   2 ,   3 . The use of different keys for each component can improve color-PE methods security as demonstrated in [43]. The method uses separate key as K i R K i G K i B for i 1 ,   2 ,   3 . However, when blocks are shuffled independently in each component; then the resultant encrypted image is not JPEG compressible. An alternative method is proposed in [12,18], where blocks are shuffled by using same key in each color channel and processes them independently in Step 2 and 3 as K i R K i G K i B for i 2 ,   3 . It can be seen that keyspace of color-PE methods can be improved by processing color channels independently; however, main limitation of the methods is due to the choice of a block size. In order to avoid block artifacts in recovered image due to chroma subsampling, a block-size of no less than 16 × 16 should be used. To overcome this limitation, grayscale-PE methods have been proposed in [13,14].

3.2. Grayscale-PE Methods

As opposed to color-PE methods, grayscale-PE methods [13,14] represent an input color image as a pseudo grayscale image by concatenating its color components in either vertical or horizontal direction. The methods consist of the following steps:
Step 1. Convert an image I R G B into I Y C b C r as
Y = 0.299 × R + 0.587 × G + 0.114 × B C b = 0.1687 × R 0.3313 × G + 0.5 × B + 128 C r = 0.5 × R 0.4187 × G 0.0813 × B + 128
Step 2. Generate a pseudo grayscale image I G S by concatenating I Y C b C r components along an axis. When I Y C b C r has W × H pixels in each component then, I G S will have 3 × W × H pixels. Chroma subsampling can be carried out at this stage as proposed in [14]. For example, when 4:2:0 chroma subsampling is used, I C b and I C r components dimensions are reduced to W / 2 × H / 2 , and then I G S will have 3 / 2 W × H pixels.
Step 3. Perform block-based transformations of color-PE methods as discussed in Section 3.1 without the color channel shuffling step.
As opposed to color-PE methods, grayscale-PE methods allow the use of smaller block size 8 × 8. Thereby, expands keyspace and improves encryption efficiency. The improved security of conventional methods is based on an assumption that their input is a color image. However, when a grayscale image is considered as an input then conventional methods are inadequate. Therefore, this work proposed an efficient PE method that is applicable to both color and grayscale images as described below.

3.3. Proposed PE Method

The JPEG algorithm exploits image statistical properties such as redundancy and correlation among adjacent pixels to achieve compression savings. Therefore, conventional methods perform encryption steps in such a way (for example, the use of 8 × 8 block size) that intrinsic properties necessary for compression remain intact in generated cipher images. However, each step of block-based encryption algorithm affects image properties differently. For example, block shuffling and negative-positive transformations change correlation coefficient, whereas changing block orientation only changes the correlation direction. Therefore, performing the latter step on a smaller block size can improve encryption efficiency without compromising compression efficiency. The proposed efficient PE method for grayscale images is shown in Figure 2 and is described below:
Step 1. Divide an input grayscale image I with W × H pixels into non-overlapping blocks of size B w × B h , and shuffle them by using a randomly generated key K 1 .
Step 2. Apply negative-positive transformation to each block selected by a uniformly distributed binary key K 2 as in (1).
Step 3. Divide each block into sub-blocks of size B w × B h and change their orientation by performing rotation and inversion based on randomly generated key K 3 . In this step, block size smaller than 8 × 8 can be used.

4. Results

In this section, we present the security and compression efficiency of the proposed PE method against conventional PE methods. When input image is grayscale then conventional color-PE [12,17,18,43] and grayscale-PE [13,14] methods can be implemented in the same way as described in Section 3.2. For the analysis, Shenzhen dataset [21] was used, which consists of 662 images and all images were resized to same dimension of 2048 × 2048. The JPEG algorithm was implemented in grayscale mode with luminance quantization standard table and quality factors were chosen as Q f 71 ,   72 ,   ,   100 . In addition, a deep learning model based on EfficientNetV2 [44] was implemented as image classifier.

4.1. Encryption Analysis

This subsection presents robustness of the proposed method against various statistical attacks, cipher-text only (COA) attack and transmission faults. The analysis were carried out on USC-SIPI Miscellaneous dataset [45]. In Section 4.1.4, all PE methods were extended to color images and UCID dataset [46] was used for COA robustness analysis.

4.1.1. Correlation Analysis

An encryption algorithm should eliminate correlation among adjacent pixels in an image for better security. The correlation coefficient ρ x , y between two distributions x and y each with N elements is given by,
ρ x , y = 1 N i = 1 N x i μ x σ x y i μ y σ y ,
where μ A is the mean and σ A is the standard deviation defined as
μ A = 1 N i = 1 N A i ,   and   σ A = 1 N i = 1 N A i μ A 2
The coefficient ρ 1.0 ,   1.0 , where ρ = 0 shows that there is no correlation, ρ < 0 shows negative correlation and ρ > 0 shows positive correlation between the distributions. The positive correlation means that both distributions are following the same trend while negative correlation means that they are following different trends. The example images from USC-SIPI Miscellaneous dataset used for correlation analysis are shown in Figure 3. For the correlation analysis, we have considered two scenarios: correlation between adjacent pixels randomly chosen from the whole image and correlation between adjacent pixels lying on the edges of blocks. In the first scenario, correlation among the neighboring pixels is still high as shown in Table 1. The reason is that in order to preserve compressibility of an image, PE methods do not alter correlation in a block. At first, it may seem like PE algorithms are vulnerable; however, when the second scenario is considered, it can be seen that on a block level the image has low correlation in all three directions and exhibits favorable encryption properties. Table 1 presents correlation analysis on the dataset in diagonal (D), horizontal (H) and vertical (V) directions for plain images and PE encrypted images. The last row shows mean of correlation across the whole dataset.

4.1.2. Histogram Analysis

Histogram of an image gives intensity distribution as a number of pixels at each intensity level. A plain image histogram is a skewed distribution concentrated at one location while histogram of a cipher image is a uniform distribution because the encryption process alters the intensity distribution. To quantify characteristics of a histogram H , histogram variance V H is calculated as,
V H = 1 N 1 i = 1 N H i μ H 2
where, N is the level of intensities in an image and μ is the mean calculated as in (4), and small value of V H means a uniform distribution. Table 2 shows histogram analysis, where the last row shows the mean V H across the whole dataset. For the proposed PE method, mean of V H as well as each image V H value is smaller than that of the plain images; therefore, reduces information characteristics of an image.

4.1.3. Information Entropy Analysis

The information entropy shows a degree of randomness in an image. The entropy of an image H I is given by,
H I = i = 1 N p i log 2 p i
where, p i is probability of a pixel value in an image. For a truly random image with N = 256 intensity levels, ideal value of entropy should be closer to H I = log 2 N = 8 . Table 2 shows mean of H I across the whole dataset for plain and cipher images. In all cases, H I of the proposed method is greater than that of plain images and is similar to that of conventional methods. Therefore, output of the proposed method is completely random.

4.1.4. Jigsaw Puzzle Solver Attack

PE methods perform block-based encryption and the cipher image preserves intrinsic properties of an image to make it compressible; therefore, it is necessary to evaluate their robustness against jigsaw puzzle solver attack [16]. This is a cipher-text only attack, where each block of the cipher image can be treated as a tile of jigsaw puzzle. Robustness against the attack can be quantified by using the following three measures [16,47]:
Direct comparison ( D c ): estimates the ratio of blocks that are in correct positions in a recovered image as that would have been appeared in an original image. Let I be an original image, I r recovered image, p i is the ith block and n is total number of blocks then D c I r is
D c I r = 1 n i = 1 n d c p i ,   where d c p i = 1 , I r p i = I p i 0 , I r p i I p i .
Neighbor comparison ( N c ): estimates the ratio of adjacent neighboring blocks that are correctly joined. For a recovered image I r with B boundaries among the blocks and b i is ith boundary then N c I r is given as,
N c I r = 1 B i = 1 B n c b i ,   where n c b i = 1 , if   b i   is   joined   correctly 0 , otherwise                           .
Largest component comparison ( L c ): estimates the ratio of largest joined blocks that have correct neighbor adjacencies with other blocks in that component. For the recovered image I r with n partial correctly assembled areas and l c number of blocks in the ith assembled area, L c I r is given by,
L c I r = 1 n max i l c I r , i .
The measures score D c , N c , L c 0 ,   1 where 1 means highest assembly score. A jigsaw puzzle solver mainly relies on color information present in an image; therefore, for fair comparison all PE methods were implemented for color image encryption. For this purpose, 20 images from the UCID dataset were encrypted independently and then assembled using a jigsaw puzzle solver. The proposed method performs rotation and inversion on a sub-block level; therefore, the cipher images were first assembled as a jigsaw puzzle type where only pieces orientations are unknown. Once correct orientations were recovered, an extended jigsaw puzzle solver [16] was used to recover the original image. Note that recovery of sub-blocks orientations does not guarantee the correct orientation of their entire block. Table 3 summarizes robustness of PE methods against jigsaw puzzle solver attack. The assembly scores obtained were averaged across the whole dataset. Though cipher images obtained with proposed method have color information; however, the smaller score values are attributed to smaller block size.

4.1.5. Robustness Analysis

When data are transmitted over a communication channel, they are always susceptible to noise due to transmission faults, which makes recovery of a plain image difficult [5]. In this subsection, we analyzed robustness of PE methods against noise attack. For this purpose, we have chosen Salt and Pepper (SPN) and Gaussian (GN) noises. In simulations, varying amount of noises were added to the cipher images obtained from different PE methods, and then the original images were recovered from those noisy cipher images. The amount of SPN and variance of GN both were chosen to be 0.05 ,   0.025 ,   0.0125 ,   0.01 ,   0.001 . The grayscale images were taken from USC-SIPI Miscellaneous dataset [45]. For visual analysis, Figure 4 and Figure 5 show recovered images from the noisy cipher images with GN and SPN, respectively. In both figures, (a) is an example of original image, and the recovered images are (b–d), (e–g) and (h–j) for conventional, proposed (4 × 4) and proposed (2 × 2) PE methods, respectively. The recovered images in Figure 5 have better visual appearance than that in Figure 4. In order to quantify images quality, Table 4 shows their Multiscale Structural Similarity Index Measure (MS-SSIM) [48] measure score for different methods. Overall, PE methods have strong robustness against SPN than GN.

4.1.6. Keyspace Analysis

In this section, we present the key size of PE methods. As discussed in Section 3, encryption algorithm of PE schemes for grayscale images mainly consists of three steps. Each step is a block based processing and have a randomly generated key. The key size of each step depends on its mode of processing and on number of blocks. When an image with   W × H   pixels is divided into blocks   B w × B h   size, then the number of blocks   n   are
n = W B w × H B h .
When either dimension is not divisible by the chosen block size then a necessary number of pixels should be padded to the image. For each step of PE, the key length is given as
K 1 = n ! K 2 = 8 n K 3 = 2 n ,
where A   is the cardinality function that returns number of elements in A ,   K 1   is a key used in block permutation step, K 2   is a key used for changing orientations of blocks and K 3   is a binary key for the negative-positive transformation step. Note that the rotation R o t   and inversion I n v   steps each have four possible operations, that is R o t 0 ° ,   90 ° ,   180 ° ,   270 °   and I n v H , V , H , V , where H is a horizontal flip and V is a vertical flip. However, their combination leads to a collision in the keyspace. For example, a block rotated by   90 °   followed by a vertical flip is similar to a block rotated   270 °   followed by a horizontal flip. Therefore, K 2 = 8 n   instead of K 2 = 16 n to avoid the collision. Since, input is a grayscale image; therefore, color-PE and grayscale-PE methods yield the same keyspace K c o n v   derived as
K c o n v = K 1 K 2 K 3             = n ! 8 n 2 n .
In proposed scheme, key length is independent of the input image representation and can perform Step 3 on a sub-block level; therefore, an extended keyspace   K p r o p   can be derived as
K p r o p = K 1 K 2 K 3                           = n ! 8 n sub 8 n 2 n ,
where   n sub = B w / w × B h / h   is the number of sub-blocks with   w   and h values are chosen smaller than   w   and   h values. Note that   K 2   in (13) is a product of   8 n sub   and   8 n   because recovery of sub-blocks orientations does not guarantee the correct orientation of their entire block. Therefore, proposed method has a larger keyspace and enhanced security against different attacks.

4.1.7. Discussion

In the literature, encryption efficiency of conventional PE methods has been analyzed only in terms of their key size and robustness against jigsaw puzzle solver based cipher-text-only attack. However, there are other most commonly used statistical tests to assess security of an encryption algorithm. Therefore, the current study has evaluated conventional and proposed PE methods using these tests as presented in Section 4.1. Note that results for differential attack analysis of PE methods are not included. The reason is that in differential attacks, an intruder obtains cipher images of two almost similar plain images (for example, difference of a single pixel value) and observes any statistical patterns that can be exploited to break security of the encryption algorithm. In order to be robust against such an attack, an encryption algorithm should generate completely two different cipher images. This is usually achieved by diffusion process of an encryption algorithm; however, PE algorithms lack this property and result in almost similar cipher images. For example, the images will have difference only on a block level where pixel values were modified. To overcome this issue, it has been suggested in literature to use different keys for encrypting different images. The proposed PE method follows the same practice to resist differential attacks.

4.2. Compression Analysis

Rate distortion curves (RD) of compression algorithms give a relationship between encoded image qualities with respect to bit rates. For this purpose, different image quality metrics such as Peak Signal-to-Noise Ratio (PSNR), Structural Similarity Index Measure (SSIM) [49], and MS-SSIM can be used. In the literature, only PSNR between original and PE encoded images has been computed as an image quality metric. However, SSIM and MS-SSIM are regarded as better measures to quantify visual degradation of an image. Therefore, this subsection presents analysis of PE methods with respect to SSIM (dB) and MS-SSIM (dB) along with PSNR (dB). SSIM and MS-SSIM values are 10 log 10 1 M ,   where   M is either SSIM or MS-SSIM score. This logarithmic transform is necessary when methods can achieve high quality results. For compression analysis, images in the Shenzhen dataset were first encrypted by conventional and proposed PE methods and then the resulting cipher images were compressed using the JPEG algorithm. Figure 6 shows RD curves for different PE methods on the dataset using PSNR, SSIM, and MS-SSIM measures. For the RD curves, the JPEG quality factors were set to 70–100. The original images were obtained by the JPEG image compression without any encryption. The graphs show that compression efficiency of conventional methods decreases with smaller block size. On the other hand, for proposed method, block-size has smaller effect on the JPEG compression efficiency.
Although, the RD curves give a reliable subjective analysis; however, it is difficult to quantify performance difference between two encoding schemes, as they do not have points at the exact same bit rates. Therefore, Bjøntegaard delta (BD) [50] rate measures use a polynomial fit of the curves, then sample 100 points over the fitted curve and compute areas by using the trapezoidal integration method. As a result, BD measures summarize data rate savings (BD-rate) or quality improvements (BD-QM) between two codecs. The BD-rate metric gives percent difference in area between two RD curves for equivalent quality after the logarithmic transform of bitrate. Similarly, BD-QM gives average difference in quality for equivalent bitrate. This study reports the BD measures for PSNR (BD-PSNR), SSIM (BD-SSIM) and MS-SSIM (BD-MS-SSIM) quality measures. Figure 6 shows the rate savings and quality improvements for different PE methods with respect to the JPEG compressed images using PSNR, SSIM and MS-SSIM, respectively. The compression of cipher images obtained from conventional PE methods requires 5% more bitrate for equivalent quality (in terms of MS-SSIM) of plain images compression. However, the bitrate drastically increases when smaller block size is used. For example, with block size 4 × 4, conventional method requires almost 113% more bitrate and the requirement increases to almost 220% for block size 2 × 2. The reason is that for conventional PE methods, there is a tradeoff relation between encryption and compression efficiency because of the block size choice. In contrast, proposed PE method requires about 12% more bitrate when using block size of (4 × 4) and (2 × 2). The analysis are discussed in-terms of MS-SSIM; however, the same trend can be seen when using PSNR and SSIM.

4.3. Classification Analysis

4.3.1. Dataset

In this study, a publicly available dataset of postero-anterior chest radiograph called Shenzhen (SH) China dataset [21] was used. This dataset consists of 326 chest X-ray (CXR) images of normal cases and 336 CXR images of TB cases along with radiologist reading. To balance the dataset, we have used only 326 images from each class. The dataset was split into training, validation and testing sets which account for 80%, 10 and 10%, respectively. In addition, the input images were resized with the following steps: (1) all black borders and regions on the edges of images were cropped and (2) the images were resized to 224 × 224 dimensions from the center.

4.3.2. Proposed Data Augmentation

Data augmentation significantly increases the number of samples, which in turn can improve the accuracy of a deep learning model [51]. It is popular in domains, for example, medical image processing, where the datasets available are small and it is difficult to acquire more data [52]. The conventional data augmentation techniques are based on basic image manipulations such as geometric transformations, color transformations, mixing images and deep learning approaches such as adversarial training, neural style transfer, and GAN data augmentation [53,54]. However, when the image data are corrupted due to noise, then noise-based data augmentation is performed to improve robustness and generalization of DL models and to overcome data deficiency [53]. The data augmentation can be carried out in two ways: on the fly during training—online augmentation, and transforming and storing data on the disk before training—offline augmentation [51]. The main consideration when choosing either of the augmentation techniques is their associated additional memory and computational requirements. The online augmentation can save memory at the expense of slower training time whereas offline augmentation can provide efficiency during training at the cost of higher storage [51]. In the current study, we have proposed a new noise-based data augmentation method that takes advantage of distortion resulted from the JPEG compression algorithm. The data augmentation is carried out offline to benefit from faster training time and as the dataset size is small, the incurred storage requirement is manageable. For this purpose, 528 images from the dataset were encrypted and compressed with the JPEG quality factors Q f 71 ,   75 ,   80 ,   85 ,   90 ,   95 ,   100 , and then the original images were recovered with distortions. The resulting images were combined to form a dataset that consists of 4130 samples uniformly distributed between the two labels. Figure 7 shows sample images of the proposed data augmentation.

4.3.3. Evaluation Metrics

In analysis, we considered samples with TB as positive and healthy samples as negative classes. The proposed model can either classify an observation to be positive or negative. The number of observations belonging to the positive class and classified as such are true positives (TP) and misclassified as negative class are false negatives (FN). Similarly, the number of observations belonging to the negative class and correctly classified as such are true negatives (TN) and misclassified as positive class are false positives (FP). In this study, we considered accuracy (14), sensitivity (15), and specificity (16) measures to evaluate performance of the proposed classifier. Accuracy is the ratio of correct predictions to total predictions, which measures the total number of correct predictions (TP + TN) made by the model. Accuracy is important when FN and FP have similar costs. However, for disease diagnosis, identifying the positives is crucial and the occurrence of FN is intolerable. For example, identifying a healthy person as a patient (FP) is a smaller problem than a patient identified as a healthy person (FN). For this purpose, sensitivity is the metric that measures performance of a model by how few FN are predicted. Another metric called specificity is used alongside sensitivity, which gives the ratio of true negatives to total negative in the observations. The metrics are defined as
Accuracy = TP + TN TP + TN + FP + FN .
Sensitivity = TP TP + FN .
Specificity = TN TN + FP .

4.3.4. Proposed Model

The model used in this study is based on EfficientNet family, which was selected based on their superior performance in natural images while having a low computational cost compared to popular CNNs such as VGGs and ResNets. Given their efficiency, they have been widely adopted in medical image processing domain as well [30,41,55,56,57,58,59,60,61,62]. The architecture of the proposed model is illustrated in Figure 8. It is based on EfficientNetV2-B0 [44], which has better parameter efficiency and faster training speed than the EfficientNetV1 [31]. These are achieved by combining training-aware neural architecture search with scaling during development of the model [44]. The proposed model consists of three Fused-MBConv [63] in early layers and three MBConv in later layers. The Fused-MBConv replaces the combination of depthwise Conv3×3 and expansion Conv1×1 in MBConv [31,64] with a single regular Conv3×3 in order to utilize modern accelerators fully. Squeeze-and-Excite (SE) blocks [65] were utilized in MBConv layers to perform channel-wise feature recalibration for improved representational power of the model. The classifier consists of a fully connected layer followed by dropout and fully connected layers.

4.3.5. Analysis

First, the model was trained on cleaned images for comparison with existing works. Next, to show robustness of the proposed model against different levels of distortions resulted from compression, the model was trained on images compressed with various JPEG quality factors. Finally, to improve accuracy of the model on limited amount of available data, the model is trained using proposed augmentation method.
Table 5 presents the accuracy of the proposed model for TB detection compared to existing works. Hwang et al. [66] proposed a deep CNN for TB classification, which achieved 83.7% accuracy without transfer learning. Pasa et al. [67] proposed a simple and fast CNN that achieved 84.4%. The main advantage of their model is being lightweight and the use of fewer parameters than the state-of-the-art models; however, this simplicity comes at a cost of being not robust against different levels of compression noise as shown by [2]. An et al. [68] proposed E-TBNet, a deep neural network model based on efficient channel attention mechanism for automatic detection of TB. The E-TBNet achieved 85.0% accuracy on SH dataset. Showkatian et al. [1] proposed CNN model has achieved 87.0% accuracy, which is the highest accuracy on the SH dataset among the compared works. In comparison with existing works, our model based on EfficientNetV2 has achieved 89.52% accuracy for automatic classification of CXR images as normal or TB. Note that all of the aforementioned methods used basic transformations such as rotation, scaling etc. only on a training dataset to avoid overfitting.
In Table 5, the analysis is limited to the works that have presented their model performance on SH dataset without transfer learning. Otherwise, there are methods like ensemble and pre-training on larger dataset methods that can achieve better accuracies as summarized in Section 2. For example, Hwang et al. [66] model accuracy was improved from 83.7% to 90.3% when using transfer learning. Similarly, Showkatian et al. [1] compared pre-trained Exception, InceptionV3, ResNet50, and VGG models and their analysis showed that when using transfer learning, pre-trained Xception, ResNet50 and VGG16 has achieved highest accuracy of 90.0%. For detail review of ensemble and transfer learning methods for TB classification, please refer to [69].
Table 6 shows robustness of the proposed model against various levels of noise introduced by compression of cipher images. The accuracy of the model is shown alongside the images quality in terms of MS-SSIM and Perceptual Image Quality Evaluator (PIQUE) [70]. The MS-SSIM is a full-reference image matric that compares an input image against a reference image with no distortion. A full-reference matric measures the quality of a distorted image, as a human would perceive that. Therefore, it can be seen in Table 6 that the image quality degrades as Qf becomes smaller. On the other hand, PIQUE (also abbreviated as PIQE) is a no-reference image quality matric that exploits local block features for measuring an image quality. PIQUE value is in the range of 0–100 and the score is interpreted in steps of twenties, for example, 0–20 means excellent and 81–100 means bad quality of an image. Despite visible noise in the images, local features are still intact as suggested by the PIQUE measure. It can be seen that accuracy of the proposed model on different levels of distortion remains closer, that is, the variance is less than 2 for original, Conventional (8 × 8) and proposed (2 × 2) methods and less than 3 for Conventional (4 × 4), (2 × 2) and proposed (4 × 4) methods. Therefore, proposed model is robust against compression distortions. The accuracy of a classifier gives its total number of correct predictions; however, in medical image analysis it is important to judge a model by how fewer number of FN are predicted. Table 7 presents sensitivity and specificity analysis of the proposed model. The model was able to achieve lower error rate across all distorted images.
Finally, Table 8 shows efficiency of the proposed noise-based augmentation method for TB screening. Wang et al. [71] proposed TB-Net, a self-attention deep CNN for automatic TB detection. Their dataset consists of 6939 CXR images collected from three different datasets. The TB-Net achieved 99.85% accuracy. In addition, they have also used EfficientNetB0 on the same dataset, and achieved 98.99% accuracy. On the other hand, our proposed model has achieved 99.77% classification accuracy on original images and preserved the same accuracy across different PE decoded images. Note that a user can compress images with different quality factors depending on their bandwidth requirements. Therefore, to accommodate this in our analysis, the test set was compressed with different quality factors and the model performance was evaluated on the distorted images. The test accuracy presented in Table 8 is an average value across different quality factors.

5. Conclusions and Future Work

This paper proposed a block-based perceptual encryption (PE) method that improves security of existing PE methods for both color and grayscale images. The proposed scheme allows smaller block size in rotation and inversion steps, thereby, improving robustness against different attacks, which is confirmed by various statistical tests. Experimental results showed that the proposed scheme is suitable to avail healthcare cloud services for medical image analysis. The compression was performed in lossy mode and distortion in recovered images has no effect on performance of the proposed deep learning (DL) model for tuberculosis (TB) screening in chest X-ray images. In addition, we proposed a new noise-based data augmentation method that takes advantage of distortion that resulted from the JPEG compression algorithm to improve robustness and generalization of the proposed DL model on smaller dataset.
The present study assumed a private cloud-computation server or trustworthy CSPs; therefore, the secret key information is shared with them. As future work, we are interested in caring out computation in the encryption domain without disclosing the key information and data to CSP. In addition, we have only considered X-ray images in the current study. However, multimodal medical image fusion is a common practice used to assist experts’ decision-making during diagnosis. Therefore, extending proposed compressible PE method for different modalities is an interesting research direction.

Author Contributions

Conceptualization, I.A.; methodology, I.A.; software, I.A.; validation, S.S.; formal analysis, I.A.; investigation, I.A.; resources, I.A. and S.S.; data curation, I.A.; writing—original draft preparation, I.A.; writing—review and editing, S.S.; visualization, I.A.; supervision, S.S.; project administration, S.S.; funding acquisition, S.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF-2018R1D1A1B07048338).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All the datasets used in this study are publically available. The Shenzhen dataset used for tuberculosis screening and compression analysis is published by the U.S. National Institute of Health (NIH) and is accessible at https://ceb.nlm.nih.gov/repositories/tuberculosis-chest-X-ray-image-data-sets/31. The USC-SIPI Miscellaneous dataset is accessible at: https://sipi.usc.edu/database/database.php?volume=misc (4 July 2022) and UCID dataset is accessible at: https://qualinet.github.io/databases/image/uncompressed_colour_image_database_ucid/ (accessed on 4 July 2022) were used for encryption analysis.

Acknowledgments

The experiments in this paper were performed with GPU resources provided by the Korea NIPA (National IT Industry Promotion Agency).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Showkatian, E.; Salehi, M.; Ghaffari, H.; Reiazi, R.; Sadighi, N. Deep Learning-Based Automatic Detection of Tuberculosis Disease in Chest X-ray Images. Pol. J. Radiol. 2022, 87, 118–124. [Google Scholar] [CrossRef]
  2. Ahmad, I.; Shin, S. Encryption-Then-Compression System for Cloud-Based Medical Image Services. In Proceedings of the 2022 International Conference on Information Networking (ICOIN), Jeju-si, Korea, 12–15 January 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 30–33. [Google Scholar]
  3. Awad, A.I.; Fouda, M.M.; Khashaba, M.M.; Mohamed, E.R.; Hosny, K.M. Utilization of Mobile Edge Computing on the Internet of Medical Things: A Survey. ICT Express 2022, in press. [CrossRef]
  4. Jain, J.; Jain, A. Securing E-Healthcare Images Using an Efficient Image Encryption Model. Sci. Program. 2022, 2022, 1–11. [Google Scholar] [CrossRef]
  5. Ahmad, I.; Shin, S. A Novel Hybrid Image Encryption–Compression Scheme by Combining Chaos Theory and Number Theory. Signal Process. Image Commun. 2021, 98, 116418. [Google Scholar] [CrossRef]
  6. Siriwardhana, Y.; Gür, G.; Ylianttila, M.; Liyanage, M. The Role of 5G for Digital Healthcare against COVID-19 Pandemic: Opportunities and Challenges. ICT Express 2021, 7, 244–252. [Google Scholar] [CrossRef]
  7. Macedo, E.L.C.; de Oliveira, E.A.R.; Silva, F.H.; Mello, R.R.; Franca, F.M.G.; Delicato, F.C.; de Rezende, J.F.; de Moraes, L.F.M. On the Security Aspects of Internet of Things: A Systematic Literature Review. J. Commun. Netw. 2019, 21, 444–457. [Google Scholar] [CrossRef]
  8. Ou, Y.; Sur, C.; Rhee, K.H. Region-Based Selective Encryption for Medical Imaging. In Frontiers in Algorithmics; Preparata, F.P., Fang, Q., Eds.; Lecture Notes in Computer Science; Springer: Berlin/Heidelberg, Germany, 2007; Volume 4613, pp. 62–73. ISBN 978-3-540-73813-8. [Google Scholar]
  9. Puech, W.; Rodrigues, J.M. Crypto-Compression of Medical Images by Selective Encryption of DCT. In Proceedings of the 13th Eurpoean Signal Processing Conference, Antalya, Turkey, 4–8 September 2005; pp. 1–4. [Google Scholar]
  10. Ahmad, I.; Shin, S. Region-Based Selective Compression and Selective Encryption of Medical Images. In Proceedings of the 9th International Conference on Smart Media and Applications, Jeju-si, Korea, 17 September 2020; pp. 34–38. [Google Scholar] [CrossRef]
  11. Kawamura, A.; Kinoshita, Y.; Nakachi, T.; Shiota, S.; Kiya, H. A Privacy-Preserving Machine Learning Scheme Using EtC Images. IEICE Trans. Fundam. 2020, 103, 1571–1578. [Google Scholar] [CrossRef]
  12. Ahmad, I.; Kim, E.; Hwang, S.-S.; Shin, S. Privacy-Preserving Surveillance for Smart Cities. In Proceedings of the 2022 Thirteenth International Conference on Ubiquitous and Future Networks (ICUFN), Barcelona, Spain, 5 July 2022; pp. 301–306. [Google Scholar] [CrossRef]
  13. Chuman, T.; Sirichotedumrong, W.; Kiya, H. Encryption-Then-Compression Systems Using Grayscale-Based Image Encryption for JPEG Images. IEEE Trans. Inf. Forensics Secur. 2019, 14, 1515–1525. [Google Scholar] [CrossRef]
  14. Sirichotedumrong, W.; Kiya, H. Grayscale-Based Block Scrambling Image Encryption Using YCbCr Color Space for Encryption-Then-Compression Systems. APSIPA Trans. Signal Inf. Process. 2019, 8, e7. [Google Scholar] [CrossRef]
  15. Iida, K.; Kiya, H. Privacy-Preserving Content-Based Image Retrieval Using Compressible Encrypted Images. IEEE Access 2020, 8, 200038–200050. [Google Scholar] [CrossRef]
  16. Chuman, T.; Kurihara, K.; Kiya, H. On the Security of Block Scrambling-Based EtC Systems against Extended Jigsaw Puzzle Solver Attacks. IEICE Trans. Inf. Syst. 2018, 101, 37–44. [Google Scholar] [CrossRef]
  17. Kurihara, K.; Shiota, S.; Kiya, H. An Encryption-Then-Compression System for JPEG Standard. In Proceedings of the 2015 Picture Coding Symposium (PCS), Cairns, QLD, Australia, 31 May–3 June 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 119–123. [Google Scholar]
  18. Ahmad, I.; Shin, S. Block-Based Perceptual Encryption Algorithm with Improved Color Components Scrambling. In Proceedings of the Korean Institute of Next Generation Computing, Jeju-si, Korea, 19 May 2022; pp. 155–158. [Google Scholar]
  19. Grivkov, A.V.; Smirnov, A.A. Application of Convolutional Neural Networks for Diagnostics of Tuberculosis; AIP Publishing LLC: Ekaterinburg, Russia, 2020; p. 080011. [Google Scholar]
  20. Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the Inception Architecture for Computer Vision. arXiv 2015, arXiv:1512.00567. [Google Scholar]
  21. Jaeger, S.; Candemir, S.; Antani, S.; Wáng, Y.-X.J.; Lu, P.-X.; Thoma, G. Two Public Chest X-Ray Datasets for Computer-Aided Screening of Pulmonary Diseases. Quant. Imaging Med. Surg. 2014, 4, 475–477. [Google Scholar]
  22. Das, D.; Santosh, K.C.; Pal, U. Inception-Based Deep Learning Architecture for Tuberculosis Screening Using Chest X-Rays. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milan, Italy, 10–15 January 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 3612–3619. [Google Scholar]
  23. Anu Priya, P.; Vimina, E.R. Tuberculosis Detection from CXR: An Approach Using Transfer Learning with Various CNN Architectures. In International Conference on Communication, Computing and Electronics Systems; Bindhu, V., Tavares, J.M.R.S., Boulogeorgos, A.-A.A., Vuppalapati, C., Eds.; Lecture Notes in Electrical Engineering; Springer: Singapore, 2021; Volume 733, pp. 407–418. ISBN 978-981-334-908-7. [Google Scholar]
  24. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015, arXiv:1409.1556. [Google Scholar]
  25. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  26. Huang, G.; Liu, Z.; van der Maaten, L.; Weinberger, K.Q. Densely Connected Convolutional Networks. arXiv 2018, arXiv:1608.06993. [Google Scholar]
  27. Cao, K.; Zhang, J.; Huang, M.; Deng, T. X-Ray Classification of Tuberculosis Based on Convolutional Networks. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence and Industrial Design (AIID), Guangzhou, China, 28–30 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 125–129. [Google Scholar]
  28. Rahman, M.; Cao, Y.; Sun, X.; Li, B.; Hao, Y. Deep Pre-Trained Networks as a Feature Extractor with XGBoost to Detect Tuberculosis from Chest X-ray. Comput. Electr. Eng. 2021, 93, 107252. [Google Scholar] [CrossRef]
  29. Chen, T.; Guestrin, C. XGBoost: A Scalable Tree Boosting System. In Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, San Francisco, CA, USA, 13–17 August 2016; pp. 785–794. [Google Scholar]
  30. Munadi, K.; Muchtar, K.; Maulina, N.; Pradhan, B. Image Enhancement for Tuberculosis Detection Using Deep Learning. IEEE Access 2020, 8, 217897–217907. [Google Scholar] [CrossRef]
  31. Tan, M.; Le, Q.V. EfficientNet: Rethinking Model Scaling for Convolutional Neural Networks. arXiv 2020, arXiv:1905.11946. [Google Scholar]
  32. Msonda, P.; Uymaz, S.A.; Karaağaç, S.S. Spatial Pyramid Pooling in Deep Convolutional Networks for Automatic Tuberculosis Diagnosis. Trait. Du Signal 2020, 37, 1075–1084. [Google Scholar] [CrossRef]
  33. Szegedy, C.; Wei, L.; Yangqing, J.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going Deeper with Convolutions. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar] [CrossRef]
  34. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet Classification with Deep Convolutional Neural Networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 8691, 346–361. [Google Scholar]
  36. Rajaraman, S.; Antani, S.K. Modality-Specific Deep Learning Model Ensembles Toward Improving TB Detection in Chest Radiographs. IEEE Access 2020, 8, 27318–27326. [Google Scholar] [CrossRef]
  37. Szegedy, C.; Ioffe, S.; Vanhoucke, V.; Alemi, A. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning. arXiv 2016, arXiv:1602.07261. [Google Scholar] [CrossRef]
  38. Chollet, F. Xception: Deep Learning with Depthwise Separable Convolutions. arXiv 2017, arXiv:1610.02357. [Google Scholar]
  39. Dasanayaka, C.; Dissanayake, M.B. Deep Learning Methods for Screening Pulmonary Tuberculosis Using Chest X-Rays. Comput. Methods Biomech. Biomed. Eng. Imaging Vis. 2021, 9, 39–49. [Google Scholar] [CrossRef]
  40. Oloko-Oba, M.; Viriri, S. Ensemble of Convolution Neural Networks for Automatic Tuberculosis Classification. In Computational Collective Intelligence; Nguyen, N.T., Iliadis, L., Maglogiannis, I., Trawiński, B., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12876, pp. 549–559. ISBN 978-3-030-88080-4. [Google Scholar]
  41. Oloko-Oba, M.; Viriri, S. Ensemble of EfficientNets for the Diagnosis of Tuberculosis. Comput. Intell. Neurosci. 2021, 2021, 1–12. [Google Scholar] [CrossRef]
  42. Saif, A.F.M.; Imtiaz, T.; Shahnaz, C.; Zhu, W.-P.; Ahmad, M.O. Exploiting Cascaded Ensemble of Features for the Detection of Tuberculosis Using Chest Radiographs. IEEE Access 2021, 9, 112388–112399. [Google Scholar] [CrossRef]
  43. Imaizumi, S.; Kiya, H. A Block-Permutation-Based Encryption Scheme with Independent Processing of RGB Components. IEICE Trans. Inf. Syst. 2018, 101, 3150–3157. [Google Scholar] [CrossRef]
  44. Tan, M.; Le, Q.V. EfficientNetV2: Smaller Models and Faster Training. arXiv 2021, arXiv:2104.00298. [Google Scholar]
  45. SIPI Image Database-Misc. Available online: https://sipi.usc.edu/database/database.php?volume=misc (accessed on 4 July 2022).
  46. Schaefer, G.; Stich, M. UCID: An Uncompressed Color Image Database; Yeung, M.M., Lienhart, R.W., Li, C.-S., Eds.; SPIE: San Jose, CA, USA, 2003; pp. 472–480. [Google Scholar]
  47. Gallagher, A.C. Jigsaw Puzzles with Pieces of Unknown Orientation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 382–389. [Google Scholar]
  48. Wang, Z.; Simoncelli, E.P.; Bovik, A.C. Multiscale Structural Similarity for Image Quality Assessment. In Proceedings of the The Thrity-Seventh Asilomar Conference on Signals, Systems & Computers, Pacific Grove, CA, USA, 9–12 November 2003; IEEE: Piscataway, NJ, USA, 2003; pp. 1398–1402. [Google Scholar]
  49. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  50. Gisle Bjontegaard. Calculation of Average PSNR Differences between RD-Curves. Doc. VCEG-M33 ITU-T Q6/16, Austin, USA, 2–4 April 2001.
  51. Shorten, C.; Khoshgoftaar, T.M. A Survey on Image Data Augmentation for Deep Learning. J. Big Data 2019, 6, 60. [Google Scholar] [CrossRef]
  52. Kim, D.; Joo, J.; Kim, S.C. Fake Data Generation for Medical Image Augmentation Using GANs. In Proceedings of the 2022 International Conference on Artificial Intelligence in Information and Communication (ICAIIC), Jeju-si, Korea, 21–24 February 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 197–199. [Google Scholar]
  53. Momeny, M.; Neshat, A.A.; Hussain, M.A.; Kia, S.; Marhamati, M.; Jahanbakhshi, A.; Hamarneh, G. Learning-to-Augment Strategy Using Noisy and Denoised Data: Improving Generalizability of Deep CNN for the Detection of COVID-19 in X-Ray Images. Comput. Biol. Med. 2021, 136, 104704. [Google Scholar] [CrossRef] [PubMed]
  54. Kim, E.; Kim, J.; Lee, H.; Kim, S. Adaptive Data Augmentation to Achieve Noise Robustness and Overcome Data Deficiency for Deep Learning. Appl. Sci. 2021, 11, 5586. [Google Scholar] [CrossRef]
  55. Luz, E.; Silva, P.L.; Silva, R.; Silva, L.; Moreira, G.; Menotti, D. Towards an Effective and Efficient Deep Learning Model for COVID-19 Patterns Detection in X-Ray Images. Res. Biomed. Eng. 2022, 38, 149–162. [Google Scholar] [CrossRef]
  56. Chetoui, M.; Akhloufi, M.A. Deep Efficient Neural Networks for Explainable COVID-19 Detection on CXR Images. In Advances and Trends in Artificial Intelligence; Fujita, H., Selamat, A., Lin, J.C.-W., Ali, M., Eds.; Artificial Intelligence Practices; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12798, pp. 329–340. ISBN 978-3-030-79456-9. [Google Scholar]
  57. Khan, I.U.; Aslam, N.; Anwar, T.; Aljameel, S.S.; Ullah, M.; Khan, R.; Rehman, A.; Akhtar, N. Remote Diagnosis and Triaging Model for Skin Cancer Using EfficientNet and Extreme Gradient Boosting. Complexity 2021, 2021, 1–13. [Google Scholar] [CrossRef]
  58. Marques, G.; Agarwal, D.; de la Torre Díez, I. Automated Medical Diagnosis of COVID-19 through EfficientNet Convolutional Neural Network. Appl. Soft Comput. 2020, 96, 106691. [Google Scholar] [CrossRef]
  59. Chetoui, M.; Akhloufi, M.A.; Yousefi, B.; Bouattane, E.M. Explainable COVID-19 Detection on Chest X-Rays Using an End-to-End Deep Convolutional Neural Network Architecture. Big Data Cogn. Comput. 2021, 5, 73. [Google Scholar] [CrossRef]
  60. Kim, S.; Rim, B.; Choi, S.; Lee, A.; Min, S.; Hong, M. Deep Learning in Multi-Class Lung Diseases’ Classification on Chest X-Ray Images. Diagnostics 2022, 12, 915. [Google Scholar] [CrossRef]
  61. Zhu, H.; Salcudean, S.; Rohling, R. Gaze-Guided Class Activation Mapping: Leveraging Human Attention for Network Attention in Chest X-rays Classification. arXiv 2022, arXiv:2202.07107. [Google Scholar]
  62. Liu, J.; Sun, W.; Zhao, X.; Zhao, J.; Jiang, Z. Deep Feature Fusion Classification Network (DFFCNet): Towards Accurate Diagnosis of COVID-19 Using Chest X-Rays Images. Biomed. Signal Process. Control 2022, 76, 103677. [Google Scholar] [CrossRef]
  63. EfficientNet-EdgeTPU: Creating Accelerator-Optimized Neural Networks with AutoML. Google AI Blog. Available online: https://ai.googleblog.com/2019/08/efficientnet-edgetpu-creating.html (accessed on 4 July 2022).
  64. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.-C. MobileNetV2: Inverted Residuals and Linear Bottlenecks. arXiv 2019, arXiv:1801.04381. [Google Scholar]
  65. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Piscataway, NJ, USA, 2018. [Google Scholar]
  66. Hwang, S.; Kim, H.-E.; Jeong, J.; Kim, H.-J. A Novel Approach for Tuberculosis Screening Based on Deep Convolutional Neural Networks; Tourassi, G.D., Armato, S.G., Eds.; SPIE: San Diego, CA, USA, 2016; p. 97852W. [Google Scholar]
  67. Pasa, F.; Golkov, V.; Pfeiffer, F.; Cremers, D.; Pfeiffer, D. Efficient Deep Network Architectures for Fast Chest X-Ray Tuberculosis Screening and Visualization. Sci. Rep. 2019, 9, 6268. [Google Scholar] [CrossRef]
  68. An, L.; Peng, K.; Yang, X.; Huang, P.; Luo, Y.; Feng, P.; Wei, B. E-TBNet: Light Deep Neural Network for Automatic Detection of Tuberculosis with X-Ray DR Imaging. Sensors 2022, 22, 821. [Google Scholar] [CrossRef]
  69. Oloko-Oba, M.; Viriri, S. A Systematic Review of Deep Learning Techniques for Tuberculosis Detection from Chest Radiograph. Front. Med. 2022, 9, 830515. [Google Scholar] [CrossRef]
  70. Venkatanath, N.; Praneeth, D.; Maruthi Chandrasekhar, B.h.; Channappayya, S.S.; Medasani, S.S. Blind Image Quality Evaluation Using Perception Based Features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Mumbai, India, 27 February–1 March 2015; IEEE: Piscataway, NJ, USA, 2015; pp. 1–6. [Google Scholar]
  71. Wong, A.; Lee, J.R.H.; Rahmat-Khah, H.; Sabri, A.; Alaref, A. TB-Net: A Tailored, Self-Attention Deep Convolutional Neural Network Design for Detection of Tuberculosis Cases from Chest X-Ray Images. Front. Artif. Intell. 2021, 5, 827299. [Google Scholar] [CrossRef]
Figure 1. Illustration of block-based perceptual encryption algorithm. (a) The main steps of compressible perceptual encryption algorithm. (b) The geometric and color transformations of perceptual encryption algorithms.
Figure 1. Illustration of block-based perceptual encryption algorithm. (a) The main steps of compressible perceptual encryption algorithm. (b) The geometric and color transformations of perceptual encryption algorithms.
Electronics 11 02514 g001
Figure 2. Comparison of conventional and proposed perceptual encryption algorithms. The encryption steps are illustrated in (bd) for conventional methods and (eh) for proposed method. (a) original blocks in the image. (b,e) are blocks permutation steps. (c) is blocks rotation and inversion step. (d,f) are negative and positive transformation steps where shaded blocks are the transformed ones. (g) is division of blocks into sub-blocks shown by dashed vertical and horizontal lines. (h) is sub-blocks rotation and inversion step.
Figure 2. Comparison of conventional and proposed perceptual encryption algorithms. The encryption steps are illustrated in (bd) for conventional methods and (eh) for proposed method. (a) original blocks in the image. (b,e) are blocks permutation steps. (c) is blocks rotation and inversion step. (d,f) are negative and positive transformation steps where shaded blocks are the transformed ones. (g) is division of blocks into sub-blocks shown by dashed vertical and horizontal lines. (h) is sub-blocks rotation and inversion step.
Electronics 11 02514 g002
Figure 3. Example images from USC-SIPI Miscellaneous dataset with different resolutions. (a) is image 5.1.09 of dimension (256 × 256), (e) is image 5.2.10 of dimension (512 × 512) and (i) is image 5.3.01 of dimension (1024 × 1024). The cipher images of (a) are in (bd), (e) are in (fh) and (i) are in (jl) for conventional, proposed (4 × 4) and proposed (2 × 2) PE methods, respectively. The bottom right corner is zoomed in every image and is shown beside them.
Figure 3. Example images from USC-SIPI Miscellaneous dataset with different resolutions. (a) is image 5.1.09 of dimension (256 × 256), (e) is image 5.2.10 of dimension (512 × 512) and (i) is image 5.3.01 of dimension (1024 × 1024). The cipher images of (a) are in (bd), (e) are in (fh) and (i) are in (jl) for conventional, proposed (4 × 4) and proposed (2 × 2) PE methods, respectively. The bottom right corner is zoomed in every image and is shown beside them.
Electronics 11 02514 g003
Figure 4. Robustness analysis of perceptual encryption (PE) methods against Gaussian noise (GN). (a) is original image without noise. Recovered images from noisy cipher images by conventional PE method (bd), proposed PE method (4 × 4) (eg) and proposed PE method (2 × 2) (hj). The GN variance is 0.05 ,   0.0125 ,   0.001 in Column 2–4, respectively. The telephone set is zoomed in every image and is shown beside them.
Figure 4. Robustness analysis of perceptual encryption (PE) methods against Gaussian noise (GN). (a) is original image without noise. Recovered images from noisy cipher images by conventional PE method (bd), proposed PE method (4 × 4) (eg) and proposed PE method (2 × 2) (hj). The GN variance is 0.05 ,   0.0125 ,   0.001 in Column 2–4, respectively. The telephone set is zoomed in every image and is shown beside them.
Electronics 11 02514 g004
Figure 5. Robustness analysis of perceptual encryption (PE) methods against Salt and Pepper noise (SPN). (a) is original image without noise. Recovered images from noisy cipher images by conventional PE method (bd), proposed PE method (4 × 4) (eg) and proposed PE method (2 × 2) (hj). The SPN intensity is 0.05 ,   0.0125 ,   0.001 in Column 2–4, respectively. The telephone set is zoomed in every image and is shown beside them.
Figure 5. Robustness analysis of perceptual encryption (PE) methods against Salt and Pepper noise (SPN). (a) is original image without noise. Recovered images from noisy cipher images by conventional PE method (bd), proposed PE method (4 × 4) (eg) and proposed PE method (2 × 2) (hj). The SPN intensity is 0.05 ,   0.0125 ,   0.001 in Column 2–4, respectively. The telephone set is zoomed in every image and is shown beside them.
Electronics 11 02514 g005
Figure 6. Compression performance analysis of perceptual encryption methods. (ac) shows rate distortion curves for different image quality measures and (df) show rate savings (Bjøntegaard Delta) relative to JPEG under different image quality measures. Proposed method allows smaller block with acceptable degradation in compression savings.
Figure 6. Compression performance analysis of perceptual encryption methods. (ac) shows rate distortion curves for different image quality measures and (df) show rate savings (Bjøntegaard Delta) relative to JPEG under different image quality measures. Proposed method allows smaller block with acceptable degradation in compression savings.
Electronics 11 02514 g006
Figure 7. Example images of the proposed noise-based augmentation method. (a) is original image. (bd) are conventional method decoded noisy images. (eg) are proposed method decoded noisy images. The JPEG quality factor is set to (80, 90, 100) from left to right. The top left corners are zoomed in every image and are shown beside them.
Figure 7. Example images of the proposed noise-based augmentation method. (a) is original image. (bd) are conventional method decoded noisy images. (eg) are proposed method decoded noisy images. The JPEG quality factor is set to (80, 90, 100) from left to right. The top left corners are zoomed in every image and are shown beside them.
Electronics 11 02514 g007
Figure 8. Illustration of the proposed deep learning model for automatic tuberculosis screening in chest X-ray images. The architecture is based on EfficientNetV2 [44] model.
Figure 8. Illustration of the proposed deep learning model for automatic tuberculosis screening in chest X-ray images. The architecture is based on EfficientNetV2 [44] model.
Electronics 11 02514 g008
Table 1. Correlation coefficient (whole image|block level) analysis of perceptual encryption algorithms in diagonal (D), horizontal (H) and vertical (V) directions.
Table 1. Correlation coefficient (whole image|block level) analysis of perceptual encryption algorithms in diagonal (D), horizontal (H) and vertical (V) directions.
MethodsPlain ImagesConventional PE ImagesProposed PE Images
4 × 42 × 2
DHVDHVDHVDHV
256 × 256
5.1.090.90|0.850.90|0.830.94|0.880.68|0.040.81|−0.040.80|−0.030.67|−0.010.79|−0.020.80|−0.010.67|−0.050.78|0.020.80|0.05
5.1.100.82|0.640.91|0.780.86|0.660.62|0.040.78|0.030.78|−0.020.56|−0.080.74|0.050.73|−0.010.57|−0.020.73|0.020.70|0.00
5.1.110.89|0.830.96|0.880.94|0.910.75|0.030.86|−0.010.87|−0.040.75|−0.060.86|−0.020.87|0.070.75|−0.020.86|−0.020.86|0.02
5.1.120.94|0.890.96|0.920.97|0.930.74|0.030.86|−0.020.86|−0.040.73|0.010.85|−0.040.86|0.00.75|0.010.85|0.010.86|0.02
5.1.130.76|0.490.87|0.710.87|0.720.70|0.020.83|−0.050.84|0.040.67|−0.080.81|0.000.82|0.040.67|−0.040.81|0.000.80|−0.04
5.1.140.85|0.680.95|0.850.90|0.780.70|0.020.83|0.000.83|0.020.66|0.000.81|0.010.80|0.020.66|0.050.80|−0.020.80|0.05
512 × 512
5.2.080.86|0.740.94|0.870.89|0.800.65|−0.010.80|0.010.80|−0.030.61|−0.010.78|0.040.76|0.010.61|0.000.77|−0.010.74|−0.01
5.2.090.80|0.600.90|0.740.86|0.680.71|0.000.84|0.000.84|0.010.69|0.000.82|−0.010.82|0.000.69|0.010.82|−0.010.81|0.01
5.2.100.90|0.820.94|0.890.93|0.860.70|−0.040.82|−0.020.82|−0.010.67|0.010.81|0.000.80|0.000.68|−0.010.80|0.010.80|0.04
7.1.010.91|0.820.96|0.910.92|0.840.72|0.030.84|−0.010.85|0.010.70|0.000.83|−0.020.83|−0.020.70|0.010.83|0.010.82|−0.02
7.1.050.89|0.810.94|0.890.91|0.830.71|0.010.83|−0.020.83|0.010.69|0.010.82|0.020.81|0.000.69|−0.010.82|0.000.80|0.02
7.1.100.93|0.860.96|0.930.95|0.880.72|−0.020.85|0.010.84|0.010.70|0.010.83|0.000.82|0.000.70|0.000.83|−0.030.81|−0.02
1024 × 1024
5.3.010.97|0.930.98|0.950.98|0.960.75|−0.020.86|0.010.86|−0.010.74|0.010.85|−0.020.86|−0.010.74|0.000.85|0.010.85|−0.01
5.3.020.86|0.730.91|0.780.90|0.830.72|0.000.84|−0.010.84|0.000.71|0.010.83|0.000.84|0.010.71|0.000.83|0.010.83|0.00
7.2.010.95|0.930.96|0.940.95|0.940.76|0.000.87|−0.020.87|0.000.76|0.000.87|−0.010.87|0.000.76|0.000.87|0.000.87|0.00
Mean0.88|0.770.94|0.860.92|0.830.71|0.010.83|−0.010.84|−0.010.69|−0.010.82|0.000.82|0.010.69|0.000.82|0.000.81|0.01
Table 2. Histogram variance and information entropy analysis of perceptual encryption algorithms. The values are given as information entropy|(histogram variance ×105).
Table 2. Histogram variance and information entropy analysis of perceptual encryption algorithms. The values are given as information entropy|(histogram variance ×105).
Original ImagesConventional PEProposed PE
4 × 42 × 2
256 × 256
5.1.096.71|1.366.75|1.36.75|1.36.75|1.3
5.1.107.31|0.517.54|0.377.54|0.377.54|0.37
5.1.116.45|2.227.30|0.817.29|0.827.29|0.83
5.1.126.71|2.837.22|1.237.22|1.237.22|1.23
5.1.131.55|120.32.46|59.832.46|59.842.46|59.83
5.1.147.34|0.517.58|0.337.58|0.327.58|0.33
512 × 512
5.2.087.20|12.007.26|11.487.26|11.487.26|11.47
5.2.096.99|17.747.58|5.637.58|5.627.58|5.63
5.2.105.71|47.615.78|42.095.78|42.15.78|42.08
7.1.016.03|46.326.56|35.976.56|35.966.56|35.95
7.1.056.56|22.387.24|11.687.23|11.697.24|11.68
7.1.105.91|47.976.60|28.096.60|28.16.60|28.11
1024 × 1024
5.3.017.52|113.947.83|45.137.83|45.137.83|45.15
5.3.026.83|317.27.41|110.187.41|110.157.41|110.15
7.2.015.64|1156.516.58|494.96.58|494.896.58|494.87
Mean6.30|127.296.78|56.66.78|56.66.78|56.6
Table 3. Robustness analysis of block-based perceptual encryption methods against jigsaw puzzle solver attack in terms of direct comparison (Dc), largest component comparison (Lc) and neighbor comparison (Nc).
Table 3. Robustness analysis of block-based perceptual encryption methods against jigsaw puzzle solver attack in terms of direct comparison (Dc), largest component comparison (Lc) and neighbor comparison (Nc).
Methods (Block Size)DcLcNc
Color-PE (16 × 16)0.010.120.11
Grayscale-PE (8 × 8)0.0010.0020.001
Proposed-PE (4 × 4)0.010.020.05
Proposed-PE (2 × 2)0.010.020.06
Table 4. Robustness analysis of perceptual encryption algorithms against salt and pepper noise and Gaussian noise. The image quality is in terms of MS-SSIM (dB).
Table 4. Robustness analysis of perceptual encryption algorithms against salt and pepper noise and Gaussian noise. The image quality is in terms of MS-SSIM (dB).
Noise TypeConventionalProposed
4 × 42 × 2
Salt and Pepper Noise
0.050.780.780.78
0.0250.870.870.87
0.01250.920.930.92
0.010.940.940.93
0.0010.990.990.99
Gaussian Noise
0.050.570.570.57
0.0250.680.680.68
0.01250.770.770.77
0.010.80.80.8
0.0010.960.960.96
Table 5. Performance analysis of the proposed deep learning model in terms of classification accuracy (%) on Shenzhen dataset with comparison of existing methods.
Table 5. Performance analysis of the proposed deep learning model in terms of classification accuracy (%) on Shenzhen dataset with comparison of existing methods.
[66][67][68][1]Proposed
83.784.485.087.089.52
Table 6. Robustness analysis of the proposed model against different types of distortions resulted by the JPEG compression of plain and cipher images. Image quality is PIQUE|MS-SSIM, Acc is classification accuracy (%), σ2 is variance and σ is standard deviation.
Table 6. Robustness analysis of the proposed model against different types of distortions resulted by the JPEG compression of plain and cipher images. Image quality is PIQUE|MS-SSIM, Acc is classification accuracy (%), σ2 is variance and σ is standard deviation.
QfOriginalConventional (8 × 8)Conventional (4 × 4)Conventional (2 × 2)Proposed (4 × 4)Proposed (2 × 2)
Image
Quality
AccImage
Quality
AccImage
Quality
AccImage
Quality
AccImage
Quality
AccImage
Quality
Acc
43.16|−89.5243.16|−89.5243.16|−89.5243.16|−89.5243.16|−89.5243.16|−89.52
7142.65|38.4987.942.60|38.4984.6842.19|32.5984.6840.43|31.6684.6843.37|38.2885.4843.51|38.3484.68
7542.77|39.2187.142.77|39.2186.2942.37|32.8184.6841.18|32.0084.6843.38|39.0586.2943.41|39.0884.68
8042.87|40.2884.6842.92|40.3385.4842.75|33.1786.2941.81|32.4786.2943.45|40.1188.7143.54|40.1384.68
8543.12|41.0986.2942.92|41.1187.942.83|33.4785.4842.43|32.9284.6843.44|40.9087.943.54|40.9087.1
9043.03|42.9084.6843.13|42.986.2943.06|33.7888.7142.73|33.3988.7143.31|42.6284.6843.41|42.6185.48
9543.02|44.0487.943.07|44.0484.6843.11|34.0384.6843.06|33.8384.6843.15|43.8684.6843.14|43.8387.1
10043.13|45.9986.2943.13|45.8887.143.12|34.2187.143.20|34.1584.6843.11|45.8884.6843.10|45.6386.29
σ20.03|7.371.8200.04|7.191.450.14|0.372.351.06|0.872.380.02|7.442.750.03|6.971.24
σ0.18|2.711.3500.20|2.681.210.37|0.611.531.03|0.931.540.14|2.731.660.19|2.641.11
Table 7. Performance analysis of the proposed model in-terms of specificity (SPE) and sensitivity (SEN).
Table 7. Performance analysis of the proposed model in-terms of specificity (SPE) and sensitivity (SEN).
Qf.OriginalConventional (8 × 8)Conventional (4 × 4)Conventional (2 × 2)Proposed (4 × 4)Proposed (2 × 2)
SPE|SENAccSPE|SENAccSPE|SENAccSPE|SENAccSPE|SENAccSPE|SENAcc
0.87|0.9289.520.87|0.9289.520.87|0.9289.520.87|0.9289.520.87|0.9289.520.87|0.9289.52
710.87|0.8987.90.81|0.8984.680.87|0.8284.680.79|0.9084.680.87|0.8485.480.76|0.9484.68
750.84|0.9087.10.81|0.9286.290.81|0.8984.680.77|0.9284.680.81|0.9286.290.76|0.9484.68
800.79|0.9084.680.79|0.9285.480.82|0.9086.290.81|0.9286.290.85|0.9288.710.82|0.8784.68
850.79|0.9486.290.87|0.8987.90.85|0.8585.480.84|0.8584.680.85|0.9087.90.81|0.9487.1
900.82|0.8784.680.77|0.9586.290.87|0.9088.710.90|0.8788.710.79|0.9084.680.84|0.8785.48
950.85|0.9087.90.79|0.9084.680.85|0.8484.680.79|0.9084.680.84|0.8584.680.82|0.9287.1
1000.76|0.9086.290.90|0.8487.10.76|0.8587.10.81|0.8984.680.85|0.8484.680.79|0.9486.29
σ2 × 1000.17|0.041.820.22|0.111.450.15|0.132.350.2|0.072.380.08|0.132.750.14|0.091.24
σ0.04|0.021.350.05|0.031.210.04|0.041.530.04|0.031.540.03|0.041.660.04|0.031.11
Table 8. Performance analysis of the proposed noise-based augmentation method.
Table 8. Performance analysis of the proposed noise-based augmentation method.
[71]Proposed Data Augmentation Method
OriginalConventionalProposed
8 × 84 × 42 × 24 × 42 × 2
99.8699.7799.7799.5499.3199.7199.77
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ahmad, I.; Shin, S. A Perceptual Encryption-Based Image Communication System for Deep Learning-Based Tuberculosis Diagnosis Using Healthcare Cloud Services. Electronics 2022, 11, 2514. https://doi.org/10.3390/electronics11162514

AMA Style

Ahmad I, Shin S. A Perceptual Encryption-Based Image Communication System for Deep Learning-Based Tuberculosis Diagnosis Using Healthcare Cloud Services. Electronics. 2022; 11(16):2514. https://doi.org/10.3390/electronics11162514

Chicago/Turabian Style

Ahmad, Ijaz, and Seokjoo Shin. 2022. "A Perceptual Encryption-Based Image Communication System for Deep Learning-Based Tuberculosis Diagnosis Using Healthcare Cloud Services" Electronics 11, no. 16: 2514. https://doi.org/10.3390/electronics11162514

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop