Next Article in Journal
ARID1B Immunohistochemistry Is an Important Test for the Diagnosis of Dedifferentiated and Undifferentiated Gynecologic Malignancies
Next Article in Special Issue
SkinLesNet: Classification of Skin Lesions and Detection of Melanoma Cancer Using a Novel Multi-Layer Deep Convolutional Neural Network
Previous Article in Journal
Administration of Enfortumab Vedotin after Immune-Checkpoint Inhibitor and the Prognosis in Japanese Metastatic Urothelial Carcinoma: A Large Database Study on Enfortumab Vedotin in Metastatic Urothelial Carcinoma
Previous Article in Special Issue
G4-QuadScreen: A Computational Tool for Identifying Multi-Target-Directed Anticancer Leads against G-Quadruplex DNA
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Adversarial Attacks on Medical Image Classification

Institute of Information Management, National Yang Ming Chiao Tung University, Hsin-Chu 300, Taiwan
*
Author to whom correspondence should be addressed.
Cancers 2023, 15(17), 4228; https://doi.org/10.3390/cancers15174228
Submission received: 23 April 2023 / Revised: 30 July 2023 / Accepted: 18 August 2023 / Published: 23 August 2023
(This article belongs to the Collection Artificial Intelligence and Machine Learning in Cancer Research)

Abstract

:

Simple Summary

As we increasingly rely on advanced imaging for medical diagnosis, it’s vital that our computer programs can accurately interpret these images. Even a single mistaken pixel can lead to wrong predictions, potentially causing incorrect medical decisions. This study looks into how these tiny mistakes can trick our advanced algorithms. By changing just one or a few pixels on medical images, we tested how various computer models handled these changes. The findings showed that even small disruptions made it hard for the models to correctly interpret the images. This raises concerns about how reliable our current computer-aided diagnostic tools are and underscores the need for models that can resist such small disturbances.

Abstract

Due to the growing number of medical images being produced by diverse radiological imaging techniques, radiography examinations with computer-aided diagnoses could greatly assist clinical applications. However, an imaging facility with just a one-pixel inaccuracy will lead to the inaccurate prediction of medical images. Misclassification may lead to the wrong clinical decision. This scenario is similar to the adversarial attacks on deep learning models. Therefore, one-pixel and multi-pixel level attacks on a Deep Neural Network (DNN) model trained on various medical image datasets are investigated in this study. Common multiclass and multi-label datasets are examined for one-pixel type attacks. Moreover, different experiments are conducted in order to determine how changing the number of pixels in the image may affect the classification performance and robustness of diverse DNN models. The experimental results show that it was difficult for the medical images to survive the pixel attacks, raising the issue of the accuracy of medical image classification and the importance of the model’s ability to resist these attacks for a computer-aided diagnosis.

1. Introduction

The use of ML models in the medical domain enables AI to assist doctors in hospitals as a critical diagnostic tool. ML is utilized widely in healthcare for its power to assist in the early diagnosis of diseases reasonably quickly and accurately. The ML model can provide a quick diagnosis based on medical images due to the development of deep learning in the medical field. Although ML is shown to have remarkable performance in various domains, many researchers have found ML models to be susceptible to different types of adversarial attacks. For example, Mahmood et al. [1] demonstrated adversarial attacks taking place in the physical realm with eyeglass frames designed to fool face-recognition systems, and [2] proposed that an adversarial attack can reprogram neural networks to perform novel adversarial tasks.
Real-life adversarial attacks on deep learning models pose significant concerns due to their potential danger. In a notable study, Eykholt et al. [3] demonstrated the feasibility of robust physical-world attacks by perturbing a real stop sign using black and white stickers, resulting in targeted misclassification. These attacks are particularly concerning due to their ability to withstand various physical conditions, including changes in view angles, distances, and resolutions. This highlights the potential for adversaries to use deep learning models to manipulate real-world objects, such as traffic signs, leading to misperception. Such attacks have critical implications for safety and security, as they can deceive autonomous systems, including self-driving cars, and cause dangerous consequences on roads.
Real-life adversarial attacks on deep learning models not only pose significant concerns, but they also have potentially life-threatening consequences, particularly in the medical field. A misclassified medical image due to an attack could lead to delayed or incorrect treatment, potentially endangering patients’ lives. Acknowledging the criticality of securing ML models in the medical domain and developing robust defense mechanisms to mitigate the risks associated with adversarial attacks is crucial. Ensuring the integrity and reliability of AI-powered diagnostic tools is paramount to safeguarding patients’ well-being and providing accurate medical care.
This study uses a one-pixel attack on various medical images and a multi-pixel attack on the COVID-19 chest X-ray dataset. These datasets contain multiclass and binary-class multi-label classifications. According to the experiment results, most medical images could be perturbed into adversarial images. As the original one-pixel attack paper [4] did not cover the topic of multi-labels, a method was devised to evaluate the effectiveness of a multi-label dataset attack. The overall contribution of this study is as follows:
  • Test the effectiveness of a pixel attack on various types of medical images;
  • Develop an effective way to evaluate the effectiveness of attacks on multi-label datasets.

2. Related Work

Most adversarial attacks can currently be divided into the following categories:

2.1. White Box Attack

Szegedy et al. [5] were the first to show that an adversarial attack can cause the network to misclassify an image based on applying hardly perceptible perturbation. In a white box attack, the attacker needs access to the victim model, including its structure, training datasets, or parameters. Goodfellow et al. [6] proposed a fast gradient sign method (FGSM) that can quickly generate adversarial samples without hyperparameter tuning or additional processing.

2.2. Black Box Attack

In a black box attack, the attacker does not need access to training samples and the model’s structure. The two main types of black box attacks are query-based and transfer-based. Yan et al. [7] proposed the subspace attack method, which reduces the complexity of the query by limiting the search directions of gradient estimation by promising subspaces that are spanned by the input gradients of a few reference models. Meanwhile, Dong et al. [8] proposed a translation-invariant attack method to generate adversarial examples with better transferability.

2.3. One-Pixel Attack

Adversarial images are generated by changing only one pixel in the original image. The original one-pixel attack paper [4] contained a simple but effective way to generate adversarial images by utilizing a different evolution algorithm to select the sensitive pixels and optimal perturbation in a CIFAR-10 dataset. To determine why a one-pixel attack may work, Danilo [9] used propagation maps that demonstrated the effect of the perturbation in each layer of the model. Moreover, he found that pixels near the perturbed one in the one-pixel attack tended to share the same susceptibility, showing that neither neurons nor pixels are the primary sources of vulnerability but rather the receptive fields. This is because most image recognition neural networks contain convoluted layers, which makes them more susceptible.

2.4. Adversarial Attack on Medical Images

In their study, Ma et al. [10] focused on adversarial attacks on deep learning-based medical image analysis systems. Their experiments on benchmark medical image datasets showed that adversarial attacks on medical images are easier to craft due to the unique characteristics of medical image data and deep neural network models. They also demonstrated that medical adversarial examples tend to target areas outside pathological regions, resulting in fundamentally different and deep features that are easy to separate.
In their study, Paul et al. [11] investigated the impact of adversarial attacks on the accuracy of predicting lung nodule malignancy and developed an ensemble-based defense strategy. They also explored the addition of adversarial images in the training set to enhance robustness against adversarial attacks. Experiments conducted on a subset of cases from the National Lung Screening Trial (NLST) dataset revealed that adversarial attacks, specifically the Fast Gradient Sign Method (FGSM) and one-pixel attacks, significantly affected the accuracy of CNN predictions. They found that a defense strategy involving multi-initialization ensembles and training with adversarial images can improve classification accuracy.
Ozbulak et al. [12] investigated the impact of adversarial examples on deep learning models to segment medical images with a specific focus on skin lesion and glaucoma optic disc segmentation and introduced the Adaptive Mask Segmentation Attack, a novel algorithm capable of generating adversarial examples with realistic prediction masks. This algorithm is based on utilizing perturbations that are mainly imperceptible to the human eye but lead to misclassification. This research sheds light on the potential risks associated with adversarial examples in segmenting medical images and highlights the need for robust defenses to ensure the reliability and accuracy of segmentation models used in clinical applications.

3. Approach

3.1. One-Pixel Attack

Considering that the original image could be represented by an n-dimensional array x = x 1 , x 2 , , x n , f is the model we chose to attack. The input of model f is the original image x, from which the confidence level of what category x is could be obtained, which is f(x). The adversarial image is generated by perturbing a pixel in the original image x. Here, the perturbed pixel is defined as e x = e 1 , e 2 , , e n and the limit of the perturbation length is specified as L. Supposing that the class set in the dataset is C = c 1 , c 2 , , c n , the original image belongs to class cori, and we want to change it into an adversarial class cadv. cadv and cori  C, hence this can be done using the following equation:
                              max e x * f c a d v x + e x   s u b j e c t   t o   e x 0 L
In a one-pixel attack scenario, since we only want to change one pixel, the value of L is set to 1. The most direct way to find the best solution is an exhaustive search, which involves trying every different pixel in the image. For a 224 × 224 RGB image, there will be as many as 224 × 224 × 256 × 256 × 256 = 841 , 813 , 590 , 016 possibilities. As a result, a more effective way to simulate adversarial attacks is a differential evolution.

3.2. Differential Evolution

A differential evolution (DE) [13] is a branch of an evolution strategy (ES) [14]. The algorithm is developed by mimicking the natural breeding process. The DE process is shown in Figure 1 with each stage described as follows:
(1)
Initial populations
The process starts with the generation of possible solutions to the issue to be solved. Each potential solution is called a “gene”. A set of solutions is produced in each “generation”, which is the process of a specific run of the ES. This set of solutions is called a “population”. As mentioned above, f is the model to be attacked for the base image x. In a one-pixel attack, the solution is in the form of (X, Y, R, G, B) array if the base image is colored, or the form of (X, Y, I) array if the image is a greyscale one. X denotes the value of the x-coordinate, Y denotes the value of the y-coordinate, and I denotes the value of the grey level. The solution for the two-pixel attacks and three-pixel attacks is (X1, Y1, I1, X2, Y2, I2) and (X1, Y1, I1, X2, Y2, I2, X3, Y3, I3), respectively (greyscale image). The population size is set to 100, meaning that there will be 100 adversarial arrays in each generation of the DE. The initial population will be developed randomly, after which a set of parental adversarial arrays will be set A R R j = a r r 1 j , a r r 2 j , ,   a r r 100 j . The superscript indicates the number of generations, and the subscript indicates the index.
(2)
Mutation
The following formula was used to generate new genes in the mutation process:
a r r i j = a r r r 1 j + F · a r r r 2 j a r r r 3 j
a r r i j means that this is the j generation array with index i, and the apostrophe means that this is the offspring population. r is a random number ranging between 1 and the size of the parent population. F is the mutant factor that ranges from 0 to 1 and decides the strength of the mutation. According to the above formula, the mutant gene firstly comprises a random parental gene a r r r 1 j , and secondly, the difference between the two parental genes a r r r 2 j ,   a r r r 3 j . The mutant factor decides how much the difference between the two random parental genes will affect the “base gene” a r r r 1 j . The offspring population is generated by repeating the above equation 100 times. Assuming that this is generation j in the DE process, the generated offspring population is denoted as A R R j = a r r 1 j , a r r 2 j , ,   a r r 100 j .
(3)
Crossover
Since the original one-pixel attack did not include a crossover, it was not used in this work.
(4)
Selection
Unlike many other evolution strategies that enable the next generation of top performance genes to survive, DE uses a pairwise survival strategy to select the group of genes that will survive. The selection process will be applied to each parent and offspring pair. There will now be two sets of arrays in our work: A R R j and A R R j , each of which contains 100 arrays in the form of (X, Y, I) (for the one-pixel attack on a grayscale image). Each array will generate a corresponding adversarial image modified from the original image. Hence, the algorithm will now have two groups of adversarial images X j = X 1 j , X 2 j , ,   X 100 j and X j = X 1 j , X 2 j , ,   X 100 j . These images will then be input to trained model f to generate two sets of confidence level arrays, C L j = c l 1 j , c l 2 j , ,   c l 100 j and C L j = c l 1 j , c l 2 j , ,   c l 100 j . The performance of the adversarial images can be evaluated based on the confidence level. Supposing that the class set in the dataset C = c , c 2 , , c n and the original image belongs to class cori., the confidence level arrays that are generated can be denoted as c l i j = c l i 1 j , c l i 2 j , ,   c l i 100 j ,   c l i k j 0 ,   1 ,   k = 1 n   c l i k j = 1 . Each element of each confidence level array corresponds to the confidence level of the class. c l i k j is how confident the model is that the image belongs to class ck. The groups of adversarial arrays that survive will then be selected when the algorithm pairs the confidence level arrays in an equation. It will compare each group of confidence levels, which means it will compare ( c l i j ,   c l i j ) on the kth position of the confidence level array. Supposing that ck is the target class, the target attack experiment aims to maximize the fitness score. This means that the confidence level with the higher value should be reserved. For instance, the parental gene a r r i j will perform better if c l i j < c l i j . Notably, the algorithm will preserve the parental gene when the performance of both genes is similar. This group of preserved genes is then passed to the next step.
(5)
Termination
An early-stop mechanism is established to determine if the performance is good enough. Based on the above selection process, the algorithm has 100 adversarial arrays corresponding to 100 adversarial images, each belonging to a specific class in class set C. In a non-target attack, the process will be terminated if one image has a class that is different from the original image. On the other hand, the process will be terminated in a targeted attack if one image has the same class as the target class; otherwise, the preserved group of genes will become the new parental initial population, and the DE process will be re-run. The process will also be terminated when the maximum iteration is reached.

3.3. One-Pixel Attack Fitness Score Setting on a Multiclass Dataset and Multi-Label Dataset

Only multiclass datasets were used in the original one-pixel attack paper, but in medical images one image could contain multiple diseases, making a multi-label classification problem. The classifier not only needs to determine whether the image is diseased or not, but it also needs to identify all the diseases in the image. Recalling that a one-pixel attack uses DE to generate the adversarial images, and DE requires a fitness score to assess the performance of the generated images, the target or original class confidence level was used as the fitness score in the original one-pixel attack study. Suppose that the class set in the dataset is C = c 1 , c 2 , , c n , and the original image belongs to the class c o r i . c o r i C , hence when this image is processed by the classifier it will generate a confidence interval vector c l = c l 1 , c l 2 , , c l n ,   c l i 0 , 1 ,   i = 1 n c l i = 1 , and c l o r i will denote the confidence level of the original class. c l o r i c l . If the experiment involves conducting a non-target attack, the goal will be to minimize c l o r i . If the experiment involves conducting a targeted attack and the adversarial images need to become class c a d v ( c a d v   C ), the goal will be to maximize the confidence level of class c l a d v . c l a d v C . The same technique will be used for the multiclass datasets in this study [8], but for the multi-label datasets the algorithm may not look at just one specific class. It will look at multiple class confidence levels at once. Supposing that the class set in the dataset is C = c 1 , c 2 , , c n , the image label can be constructed with an array l = l 1 , l 2 , l n . l i   i N ,   i 0 , n . Each l i corresponds to each class c i in the class set C and the value will be 0 or 1. If the value is 0, the image does not contain this class, but if the value is 1 it does. The image can be considered to be a multi-dimensional array x = x 1 , x 2 , , x n   , and the classifier can be represented by f . When the image x is input to the classifier f , f x will produce a set of confidence levels c l = c l 1 ,   c l 2 , , c l n ,   c l i 0 , 1 ,   i N . Threshold γ can be set with a range from 0 to 1, where, if c l i > γ , the algorithm will consider it to contain a class c i disease. Therefore, if all c i < γ , the image is of a normal patient with no disease. Suppose that an image x o r i that we want to attack is found with the original class set c o r i . c o r i is a subset of C and has a label form l o r i . l o r i is a subset of l . By inputting the image x o r i into the classifier it becomes f x o r i , and the classifier has successfully predicted the image and produced confidence level c l o r i . Because all the elements in the c o r i needed to be considered at once, cosine similarity is used in this study to construct the fitness score. The generated adversarial image x a d v will be input into the classifier and become f x a d v , hence producing an adversarial confidence level c l a d v . If this is a non-target attack, the formula will be as follows:
      s i m i l a r i t y l o r i , c l a d v = l o r i · c l a d v max l o r i c l a d v , δ ,   δ = 10 5    
δ is a very small number that prevents the denominator from becoming 0. The goal is to minimize the above formula.
If this is the target attack and the target class set is c a d v , c a d v is a subset of C and has a label form l a d v , l a d v is a subset of l , the formula will be as follows:
        s i m i l a r i t y l a d v , c l a d v = l a d v · c l a d v max l a d v c l a d v , δ ,   δ = 10 5  
The goal is to maximize the above formula. The algorithm transformed the label and confidence level range from [0, 1] to [−1, 1] as a reminder to calculate the cosine similarity. This is because if the label is all zeroes, the cosine similarity will always be zero, causing the fitness function to fail.

4. Experimental Setting

All experiments are non-target attack experiments except for the COVID-19 dataset. For non-target attacks, each class in each multiclass dataset will undergo 100 individual experiments. We attack all test images in each class for a targeted attack on the COVID-19 dataset. For the multi-label dataset, which is the Chest dataset, only five classes were involved in the experiment with the five larges amounts of data in the whole dataset. The reason for this will be explained in the Chest dataset section.

4.1. Differential Evolution

This DE process has no crossover like the original one-pixel attack paper, and the mutant factor is set to 0.5. The population size is set to 100. The maximum number of iterations to run the DE process is limited to 100. It may stop if it meets the early-stop condition, indicating that the total number of successful images has surpassed 1% of the total number of adversarial images generated in that particular iteration.

4.2. Model

We used a stochastic gradient descent (SGD) optimizer with a learning rate = 0.001 and momentum = 0.9. The main model was ResNet50 [15]. Each model trained 100 epochs with batch size = 64. To prevent overfitting and optimize the training efficiency, we implemented early stopping with a condition that terminates training when the model achieves the 95% accuracy threshold.

4.3. Hardware and Software

The CPU is 11th Gen Intel(R) Core (TM) i7-11700KF @ 3.60GHz 3.60 GHz and the RAM is 16 GB. The GPU is NVIDIA GeForce RTX 3060 and the RAM is 12 GB.
The software is the Python package “torchvision”. At the same time, the model is pretrained and adjusted in the first layer according to the dataset size and number of channels. The last layer is also adjusted based on the class number of the dataset. The software specifications are OS: Windows 10 × 64 Education. Programming language: Python 3.9.7. Programming modules: numpy 1.20.3, torchvision 0.11.3, pytorch 1.10.2, tqdm 4.62.3, scikit-learn 0.24.2, pillow 8.4.0, tiffile 2021.7.2, matplotlib 3.4.3, pandas 1.3.4, pyyaml 6.0, sqlite 3.36.0, plotly 5.8.2.

4.4. Dataset

The Derma, Pneumonia, and OCT datasets will be split into training and testing sets with a ratio of 8:2. The training and testing split for the COVID-19 dataset will be the same as the original setting of [16]. For the Chest dataset some of the labels in the Chest multi-label dataset contain only one piece of data, which cannot be split into training and testing sets. Instead, the provider of this dataset produced a list of training and testing images that will be used for the experiment on the Chest dataset. Multi-label datasets have a different training option from multiclass datasets.
The multiclass training set will be used to train and obtain the test accuracy. The multi-label training set will exclude all “normal” cases because a “normal” label is all zeroes, which will easily cause the model to train in the wrong direction.
The “normal” case also has the most significant data points in the dataset that will contribute to 94% accuracy of the model by predicting all cases to be “normal”. Therefore, the experiment excludes this class in order to train the model appropriately. An overview of the dataset information is provided in Table 1. The images in different datasets come in different sizes and will be resized as 224 × 224. The contents of each tested database are as follows:
(1)
Derma
This is a multiclass dermatoscopic colored image dataset of common pigmented skin lesions. It consists of 10,015 images and seven categories. The dataset [17] has a size of 3 × 600 × 450. An overview of the dataset information is provided in Table 2.
(2)
Pneumonia
The Pneumonia dataset, sourced from the study [18], consists of a comprehensive collection of 5856 greyscale chest X-ray images, which provide a valuable insight into a diagnosis of pneumonia. Comprised of two classes, “Normal” (1583 images) and “Pneumonia” (4273 images), this dataset provides a solid foundation for training and evaluating machine learning models in the detection and differentiation of pneumonia cases. The images will be center-cropped with a 20% reduction in width and height, effectively removing most artificial labels to ensure an accurate analysis. The precision of the “Pneumonia” class is 88%, with a recall of 100%. The precision of the “Normal” class is 100%, with a recall of 65%.
(3)
OCT
The Optical Coherence Tomography (OCT) dataset, which is referenced in [18], is a large and diverse collection of medical imaging data specifically designed for the study and analysis of retinal diseases. It contains 109,309 high-resolution grayscale images. The dataset is organized into four distinctive classes, each of which represents a specific retinal disease or condition: “Normal” (51,390 images), “Choroidal neovascularization” (37,455 images), “Diabetic macular edema” (11,598 images), and “Drusen” (8866 images). The precision of the “Normal” class is 97%, with a recall of 100%. The precision of the “Choroidal neovascularization” class is 97% with a recall of 97%. The precision of the “Diabetic macular edema” class is 95% with a recall of 93%, and the precision of the “Drusen” class is 92% with a recall of 79%.
(4)
COVID-19 Chest X-ray
This public COVID-19 chest X-ray dataset [16] consists of images and other associated metadata, including patients’ age, gender, etc. The latest V9A version of the training and testing dataset was used. This contains 30,130 training images and 400 testing images across three classes: normal, pneumonia, and COVID-19 cases. The same training and testing split as the original setting of [16] was used in the experiment, and the image distribution is shown in Table 3.
(5)
Chest
The Chest dataset is a binary-class multi-label frontal-view X-ray chest greyscale image dataset with a total of 112,120 images and 14 classes of disease. The dataset is taken from the NIH-Chest X-ray 14 dataset [19]. The original size of the images is 1 × 1024 × 1024. The labels contained in the dataset are shown in Table 4. For example, an image with the label [0, 1, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0] shows that this is an image with the labels Cardiomegaly and Effusion. If we want to perform each possible combination, there will be 2 14 = 16 , 384 possible combinations considering there are 14 classes in the dataset. This will mean conducting 1,638,400 experiments only on the Chest dataset. However, the dataset only contains 247 different combinations of labels, most of which have a total number of images under 100, and even under 10. To prevent the model from not classifying certain types of images due to a lack of training images, we only chose the five combinations of labels with the most data in the dataset: Normal with 60,361 images, Infiltration with 9547 images, Atelectasis with 4215 images, Effusion with 3955 images, and Nodule with 2705 images. An overview of the dataset information is provided in Table 5.

5. Experimental Results and Data Analysis

The experiments will be divided into multiclass and multi-label types. Some of the terms used will be explained before any further discussion. The attack will be considered a success if “the resulting label is different from the original label”. This applies to both multiclass and multi-label datasets. For the multi-label dataset, if the resulting label contains the original label but with different classes added or deducted, it will still be a success because the resulting label will be “different” from the original one. The “success rate” is used as the index of the attack performance in this study, and the formula is as follows:
              Success   rate = Number   of   successful   adversarial   images Total   number   of   experiments    
In other words, this is the ratio of the successful experiments compared to the total number of experiments. The number of classes transformed from the original label to another class called the “transform class” will also be counted. Several tables are defined as follows:
  • Example of successfully attacked images
Because 100 experiments will be conducted on each class in each dataset, and there will be a large number of successful images if the success rate is high, only one successful image will be presented for each class. The successfully modified image will be shown with a red circle to indicate the modified pixel. There will be two description lines under each image: a bold black line to indicate the original class and a bold red line to show the class that has been transformed. The number in parentheses shows the confidence level of that class. Notably, there will be some confidence levels that indicate “~100%”. This represents a confidence level above 99.94% because the number will be rounded to one decimal point.
  • Success rate table
This shows the success rate for each class in the dataset in percentages. The first row is the class name. The second row is the corresponding attack success rate. There will sometimes be an asterisk next to the number in parentheses, indicating that different images were used to construct the attack. Ideally, the algorithm should use 100 different images for the experiment. However, many datasets had too few images to train the model, making the model’s overall accuracy too low to identify some labels or insufficient images were provided to attack that specific class.
  • Class transformation table
This is the number of images being transformed for each dataset class. The vertical axis is the original class of the image. The horizontal axis is the resulting class after the transformation. Notably, the axis order is inverted for multi-label datasets, namely the Chest dataset, due to the large number of classes being transformed.
  • The ratio of class transformation table
One of the purposes of this study is to explore the resistance of images in each class in the dataset. The class transformation table presented above shows the ratio of class transformation of each class. The ratio can be calculated as follows: Supposing that the original class is denoted as c o r i and the adversarial class is denoted as c a d v ,   there is a total of K classes in the dataset, which make o r i , a d v 0 , K ,   o r i , a d v ,   K N ,   K > 1 . The number of transformations from class c o r i to c a d v is denoted as n o r i , a d v . When o r i = a d v , the attack is considered to be a failure, so that n o r i , a d v = 0 . The transformation ratio of the specific adversarial class should be found, which is denoted as T a d v ., which can be calculated as follows:
T a d v = o r i = 0 K r o r i , a d v K F
where r o r i , a d v is the ratio of transformation from class c o r i to c a d v and is calculated as follows:
r o r i , a d v = n o r i , a d v a d v = 0 K n o r i , a d v
Based on the above formula, T a d v is the average effect for each class in c o r i .
Notably, the formula for T a d v ’s denominator has F in it. This denotes the total number of classes that failed the attack. The reason the formula needs to deduct it is that, if one class failed the test, it would not contribute any value on r o r i , a d v but it would still add 1 to the denominator (consisting of K ). Although it will not affect the order of the class transformation ratio, it will lower the total transformation ratio to 100%, which would be incorrect since all transformations should be contained in K classes. Therefore, the effect of the attack ratio can be eliminated using the above formula to obtain a relatively unbiased result of class transformation power.
  • Table of class conversion for disease type
This table shows the number and ratio of data converted from “Disease” to “Normal” or “Normal” to “Disease”. The “Class conversion” column indicates whether the type is “Disease to Normal” or “Normal to Disease”. The “Count” column indicates the total images that belong to the conversion type. The “Percentage” column indicates the ratio of the “Count” column.
  • Other necessary graphs or tables
Some datasets needed extra figures and tables to further elaborate the results. More will be mentioned and explained as needed.

5.1. Multiclass Dataset

(1)
Derma
Different Derma class successful attack images are shown in Figure 2 and Figure 3. The class type that has the disease “Melanoma” has been transformed to the normal class “Melanocytic nevi” in Figure 2. This may be because these two types are similar to each other. In Figure 3, several diseased images are categorized into normal images after the one-pixel attack. Table 6 shows that the class Dermatofibroma is more susceptible to attack than other classes. Table 7 lists the class transformation of the Derma database. It is apparent from Table 8 that the top three classes that are more easily transformed are “Melanocytic nevi”, “Basal cell carcinoma”, and “Benign keratosis”. After further analysis of the data, Table 9 shows that it is easier to transform normal images into diseased images than to transform diseased images into normal images.
(2)
Pneumonia (Multi-pixel)
The Pneumonia dataset encompasses two different image classes, namely “Normal” and “Pneumonia”. A multi-pixel attack strategy was used on both classes in this research with the objective of transforming the images into their respective opposite categories.
As illustrated in Table 10, the success rate of converting normal images into pneumonia images is positively correlated with the number of perturbed pixels. However, the algorithm’s performance in transforming pneumonia images into normal images does not yield successful results, even with an increase in the perturbed pixels. The successful attack images are shown in Figure 4.
(3)
OCT (Multi-pixel and target attack)
The OCT dataset comprises four distinctive classes: “Normal”, “Choroidal neovascularization” (CNV), “Diabetic macular edema” (DME), and “Drusen”. Three targeted attacks were conducted, each aiming to transform the images from one of the three disease classes into normal images. The outcomes of these attacks are presented in Table 11. Consistent with the results observed in the pneumonia dataset, the attack that converted normal images into diseased images demonstrated a high success rate by exhibiting a positive correlation with the number of perturbed pixels.
In the targeted attacks, the algorithm struggled to generate successful attacks when attempting to convert CNV and DME images into the normal class using only one or two perturbed pixels. It was not until the number of perturbed pixels was increased to three that the attack began to generate successful transformations.
When converting Drusen images into the normal class, the success rate remained unchanged despite increasing the number of perturbed pixels. This result indicates that the algorithm may find it difficult to adapt to specific classes of images or that the Drusen class exhibits certain characteristics that render the algorithm less effective. The successful attack images are shown in Figure 5 and Figure 6.
(4)
COVID-19 (Multi-pixel and target attack)
The average confidence level was used as an additional evaluation metric, alongside the success rate, for the multi-pixel attack in order to better understand the effect of the variation in the number of perturbed pixels on the confidence of the classification.
A v g C L = i = 1 N s u c C L i N s u c
N s u c   is the number of adversarial samples, and the victim model is classified into the target class. As shown in the above equation, AvgCL was the average confidence level of the target class in the adversarial images.
  • Diseased images to Normal images
This dataset contained two diseases: “Pneumonia” and “Covid”. The goal was to perturb a pixel in the original disease image and turn it into a normal image. Since the attacked image had to be classified as “Normal”, this was a targeted attack. The test dataset belonging to “Pneumonia” and “Covid” was attacked, and the results were evaluated based on the Success Rate and Average Confidence Level. A further experiment was conducted on two- and three-pixel attacks to understand the influence of the number of perturbed pixels on the performance of the victim model, and the results are shown in Table 12. The examples of attack images are shown in Figure 7.
Logically, with the increase in perturbed pixels, the value of AvgCL should also increase because it changes more parts of the image, making the attack easier.
However, the AvgCL was found to have dropped significantly in the three-pixel attack on the “Covid” class. It is believed that this may have been caused by the newly generated three-pixel attack images, which are more difficult to attack so the confidence levels in the target class were also lower. This may have affected the overall AvgCL.
  • Normal images to Diseased images
An experiment was also conducted using non-target attacks that turn normal images into diseased images to compare it with target attacks that turn diseased images into normal ones. The results can be seen in Table 12. Obviously, it was found to be easier to turn normal images into diseased images.
The class distribution was further examined after the attack in Table 13, and the results show that the proportion of two different disease classes generated is similar, but the AvgCL of “Covid” is always higher. This may imply that the victim model tends to “believe” that an image belongs to “Covid”. The result may explain why it is harder to attack “Covid” images.
  • Other models
All the above experiments were conducted using ResNet50 [15]. To understand how different models may perform under adversarial attacks, the attack method was also tested using DenseNet121 [20], and the results are shown in Table 14. The success rate was found to have declined slightly in both the “Covid” and “Normal” classes, meaning that the model is more robust under pixel-level attacks. However, the success rate of the class “Pneumonia” had increased, although with a lower AvgCL. These results indicate that a larger model like DenseNet121 may be even more susceptible than ResNet50 in some circumstances. Also, the results of pixel-level attacks on the COVID-19dataset can be generalized over different CNN networks.

5.2. Multi-Label Dataset

The Chest dataset is a multi-label dataset. Some successful attack examples are shown in Figure 8. The class “Normal” has no confidence level because it means that all labels’ confidence level is below 50%. Figure 9 illustrates some images transformed from a diseased class to a normal class.
It is evident that the “Normal” type had the highest success rate, which is shown in Table 15. The details of the conversion are shown in Table 16. Figure 10 shows the number of diseases in the transformed class of the Chest dataset. The top three were “Infiltration”, “Atelectasis”, and “Effusion”. Table 9 shows that it is easier to transform diseased images into normal ones.
Unlike multiclass datasets, in which the confidence level of the original class needs to be reduced, and the confidence level of other classes increased simultaneously, the confidence level of the original label of multi-label datasets does not need to be reduced too much. The confidence level of other classes only needs to be increased to generate a different label, but when researching the number of disease tables for these two datasets, the final label with one disease is found to be the most common for both datasets. This means that the algorithm will first try to reduce the original class and simultaneously increase another particular disease.
It can be seen from the table of the number of diseases in the Chest dataset that the second most common disease is no disease, which is the normal patient image. This explains why the algorithm will reduce the original label’s confidence level first before trying to increase the other classes’ confidence levels. If reducing the original label has resulted in a different label, which will always be the “Normal” label, the algorithm will stop because it has already reached its goal. Due to this phenomenon, the Chest dataset “Disease to normal” ratio increased dramatically, causing severe concern about the medical system.
In terms of the number of each disease, the types of diseases in the number of diseases table are much the same as the top five labels used in the experiments, meaning that the final table contained fewer rare diseases. This is likely to be because the model is slightly overfitted, so it will tend to predict the image with these labels.

6. Discussion

Appendices, if needed, appear before the acknowledgments.

6.1. Why Use the One-Pixel Attack?

The major reason for choosing the one-pixel attack is that it is simple and effective. Unlike many other attack methods, it does not rely on gradient descent. This has two main benefits: (1) The problem that needs to be solved does not need to be differentiated, making this attack method more flexible. (2) It requires less computing power and takes less time and cost to implement. Another reason for choosing this attack method is that this kind of “adversarial” image may not be purposely generated. Many other methods that change a lot of pixels in the image to generate adversarial images need time and knowledge, and it is highly likely that the adversarial images will need to be generated purposely. It is possible for adversarial images to be generated naturally using a one-pixel attack. These images may be generated by malfunctioning medical instruments, humans or computers processing images wrongly, or engineers making careless mistakes. They may not be adversarial images that are sufficiently powerful to fool all models, but the goal of a one-pixel attack is not to generate a powerful adversarial image but just a workable one using a reasonable amount of time and computing power.

6.2. Comparison of the Colored and Greyscale Datasets

The colored dataset used in the experiment consisted of three RGB channels, whereas the greyscale dataset only contained one channel. More channels can be seen in colored images than in greyscale images, which is likely to affect the time taken for the experiment and the attack success rate. Table 17 shows that the color channel factor had a greater effect on the attack success rate and DE experiment time.
Potential strategies to counter pixel attacks:
  • Adversarial Training: Employing adversarial training techniques can enhance the robustness of models by including adversarial examples in the training process. Training the model on both clean and perturbed images makes it more resilient to pixel attacks;
  • Input Preprocessing: Input preprocessing techniques such as denoising, image resizing, or blurring can help to remove or reduce the impact of perturbations in the input images, making the model more resilient to pixel attacks; Model Ensemble: An ensemble of multiple models can improve the classification system’s overall robustness. Aggregating the predictions of multiple models can minimize the impact of individual pixel attacks.
Limitations of this study:
Although many experiments were conducted in this study to illustrate that there may be several potential threats when using AI to recognize medical images, it is imperative to understand the severity of the impact of this phenomenon on the medical system. Will successful adversarial images cause the medical system to collapse? A real doctor dedicated to chest diseases was interviewed, and he provided some important advice:
  • The medical dataset used for labeling may not be 100% correct. The people who provided the dataset may not have correctly labeled the image, and those who used it may have been unaware of or unable to recognize the mistakes. For example, Figure 11 may not be labeled correctly. Although the image in this dataset is labeled “Infiltration”, which means that the original image had a disease named “Infiltration”, the doctor thought that the image was actually a “Normal” image. He guessed that the lower part of the image might cause the people who labeled the image to think it had been infiltrated, but it may actually be normal muscle cells. The problem of incorrect labeling questions the use of an open-source dataset;
  • A one-pixel attack may be capable of generating many adversarial images, but a real doctor will check multiple images with different angles and opacity on one patient. Unless multiple images have been attacked, it may have little impact on the medical system. Also, when doctors make a diagnosis, they do not determine if the patient has a disease by just looking at medical images. They also collect other information, including the patient’s medical records, to assist with their diagnosis. If the doctor thinks the patient may potentially have a disease, but he cannot diagnose it immediately, he may invite the patient to have follow-up treatments. All these measurements in today’s medical system indicate that the treatment may not be disrupted purely by one adversarial image;
  • Successful adversarial images were generated in this study with all possible outcomes. These outcomes can be classified by disease type as “Normal to Normal”, “Normal to Disease”, “Disease to Disease”, and “Disease to Normal”. The only type that is of interest to doctors is “Disease to Normal” because “Normal to Normal” will not affect the patient’s life, and “Normal to Disease” and “Disease to Disease” will be more likely to be subject to additional treatment and have a high chance of being re-classified into their correct category. Only “Disease to Normal” may affect the diagnosis.

6.3. Future Directions

Although many experiments were conducted in this study using various medical image datasets, many aspects could not be covered due to time and space constraints. The following are some of the aspects that might be considered for further research in the future:
  • ResNet50 was used in this study as the basic model being attacked, and some experiments were conducted to illustrate the likely consequences of using different models. However, there are still other neural networks and models that can be used for image identification. For example, capsule-type models [21], Transformers [22], and so on. Another related direction would be to use a real model in the medical domain, such as IBM CODAIT’s MAX breast detector, for comparison;
  • Image classification is just one of the tools used in the medical domain. Image segmentation is also a popular type. FGSM was used in a prior study [23] to generate adversarial images for image segmentation tasks on medical images, but a one-pixel attack was not included in the paper. Few researchers have specifically covered the one-pixel attack for medical image segmentation scenarios so far;
  • The experiments in this study demonstrated possible weaknesses in several medical image datasets under the one-pixel type of attack. Researchers can use this study to develop an effective way to defend against this attack, specifically for medical image datasets.

7. Conclusions

The experiments conducted in this study have shown that nearly all types of medical image datasets, whether greyscale or colored, X-ray, or many other types, can be modified into adversarial images with a one-pixel attack. In addition, multi-pixel attacks were made on the COVID-19 dataset to investigate the impact of the number of perturbed pixels on the success rate and average confidence level. Furthermore, a robust evaluation method was devised for the multi-label Chest dataset to assess the effectiveness of attacks specifically designed for multi-label datasets. The results demonstrated the great risk of using AI or machine learning models to identify medical images and give medical assistance or advice. These experiments showed that adversarial images could result from having no disease to having diseases or vice versa, misjudging the severity of the disease or even classifying the disease type incorrectly. If AI malfunctioned, it would have devastating consequences. Giving the wrong medical advice could delay much-needed treatment, and taking the wrong medicine would cause severe side effects. Medical resources would be wasted, and patients’ lives would be lost in worst-case scenarios. Furthermore, one-pixel attack-like adversarial images could be generated effortlessly, and related knowledge may be assumed mistakenly as a result. In summary, this study provides a major warning of the possible weaknesses of implementing AI in the medical system.

Author Contributions

Conceptualization, M.-J.T.; methodology, M.-J.T.; software, P.-Y.L. and M.-E.L.; validation, P.-Y.L. and M.-E.L.; formal analysis, M.-J.T.; investigation, P.-Y.L. and M.-E.L.; data curation, M.-J.T.; writing—M.-J.T. and M.-E.L.; supervision, M.-J.T. All authors have read and agreed to the published version of the manuscript.

Funding

The National Science Council partially supported this work in Taiwan, Republic of China, under MOST 109-2410-H-009-022-MY3 and NSTC 112-2410-H-A49-024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Patient consent was waived because no human beings were involved in the research.

Data Availability Statement

All data are publicly available.

Acknowledgments

This work was partially supported by the National Science Council in Taiwan, Republic of China, under MOST 109-2410-H-009 -022 -MY3 and NSTC 112-2410-H-A49-024. In addition, the authors thank the National Center for High-performance Computing (NCHC) of National Applied Research Laboratories (NARLabs) in Taiwan for providing computational and storage resources.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sharif, M.; Bhagavatula, S.; Bauer, L.; Reiter, M. A general framework for adversarial examples with objectives. ACM Trans. Priv. Secur. 2019, 22, 1–30. [Google Scholar] [CrossRef]
  2. Elsayed, G.F.; Goodfellow, I.; Sohl-Dickstein, J. Adversarial reprogramming of neural networks. In Proceedings of the International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  3. Eykholt, K.; Evtimov, I.; Fernandes, E.; Li, B.; Rahmati, A.; Xiao, C.; Song, D. Robust physical-world attacks on deep learning visual classification. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  4. Su, J.; Vargas, D.V.; Sakurai, K. One Pixel Attack for Fooling Deep Neural Networks. IEEE Trans. Evol. Comput. 2019, 23, 828–841. [Google Scholar] [CrossRef]
  5. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.J.; Fergus, R. Intriguing properties of neural networks. In Proceedings of the International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  6. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  7. Yan, Z.; Guo, Y.; Zhang, C. Subspace attack: Exploiting promising subspaces for query-efficient black-box attacks. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; pp. 3820–3829. [Google Scholar]
  8. Dong, Y.; Pang, T.; Su, H.; Zhu, J. Evading defenses to transferable adversarial examples by translation-invariant attacks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 4312–4321. [Google Scholar]
  9. Vargas, D.V.; Su, J. Understanding the one-pixel attack: Propagation maps and locality analysis. In CEUR Workshop Proceedings; CEUR-WS: Aachen, Germany, 2020; Volume 2640. [Google Scholar]
  10. Ma, X.; Niu, Y.; Gu, L.; Wang, Y.; Zhao, Y.; Bailey, J.; Lu, F. Understanding adversarial attacks on deep learning. Pattern Recognit. 2021, 110, 107332. [Google Scholar] [CrossRef]
  11. Paul, R.; Schabath, M.; Gillies, R.; Hall, L.; Goldgof, D. Mitigating adversarial attacks on medical image understanding systems. In Proceedings of the 2020 IEEE 17th International Symposium on Biomedical Imaging (ISBI), Iowa City, IA, USA, 3–7 April 2020; pp. 1517–1521. [Google Scholar]
  12. Ozbulak, U.; Van Messem, A.; De Neve, W. Impact of adversarial examples on deep learning models for biomedical image segmentation. In Medical Image Computing and Computer Assisted Intervention, Proceedings of the MICCAI 2019: 22nd International Conference, Shenzhen, China, 13–17 October 2019; Part II; Springer International Publishing: Berlin/Heidelberg, Germany, 2019; Volume 22. [Google Scholar]
  13. Storn, R.; Price, K. Differential Evolution—A Simple and Efficient Heuristic for global Optimization over Continuous Spaces. J. Glob. Optim. 1997, 11, 341–359. [Google Scholar] [CrossRef]
  14. Beyer, H.-G.; Schwefel, H.-P. Evolution strategies—A comprehensive introduction. Nat. Comput. 2002, 1, 3–52. [Google Scholar] [CrossRef]
  15. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  16. Wang, C.L.; Lin, Z.; Wong, A. Covid-net: A tailored deep convolutional neural network design for detection of covid-19 cases from chest x-ray images. Sci. Rep. 2020, 10, 19549. [Google Scholar] [CrossRef] [PubMed]
  17. Tschandl, P.; Rosendahl, C.; Kittler, H. The HAM10000 dataset, a large collection of multi-source dermatoscopic images of common pigmented skin lesions. Sci. Data 2018, 5, 1–9. [Google Scholar] [CrossRef] [PubMed]
  18. Kermany, D.S.; Goldbaum, M.; Cai, W.; Valentim, C.C.; Liang, H.; Baxter, S.L.; Zhang, K. Identifying Medical Diagnoses and Treatable Diseases by Image-Based Deep Learning. Cell 2018, 172, 1122–1131.e9. [Google Scholar] [CrossRef] [PubMed]
  19. Wang, X.; Peng, Y.; Lu, L.; Lu, Z.; Bagheri, M.; Summers, R.M. ChestX-Ray8: Hospital-Scale Chest X-Ray Database and Benchmarks on Weakly-Supervised Classification and Localization of Common Thorax Diseases. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2097–2106. [Google Scholar]
  20. Huang, G.; Liu, Z.; Maaten, L.V.D.; Weinberger, K.Q. Densely Connected Convolutional Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  21. Sabour, S.; Frosst, N.; Hinton, G.E. Dynamic Routing Between Capsules. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  22. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. In Proceedings of the Advances in Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017; Volume 30. [Google Scholar]
  23. Pervin, M.T.; Huq, L.; He, A.; Huo, Z.; Li, L. Adversarial Attack Driven Data Augmentation for Accurate and Robust Medical Image Segmentation. arXiv 2021, arXiv:2105.12106. [Google Scholar]
Figure 1. Overview of differential evolution procedure.
Figure 1. Overview of differential evolution procedure.
Cancers 15 04228 g001
Figure 2. Derma successful attack case examples.
Figure 2. Derma successful attack case examples.
Cancers 15 04228 g002
Figure 3. Examples of Successful Derma attacks.
Figure 3. Examples of Successful Derma attacks.
Cancers 15 04228 g003
Figure 4. Examples of the attack results on the Pneumonia dataset with confidence level of ResNet50: (a) One-pixel attack turning Pneumonia to Normal; (b) One-pixel attack turning Normal to Pneumonia; (c) Two-pixel attack turning Normal to Pneumonia; (d) Three-pixel attack turning Normal to Pneumonia.
Figure 4. Examples of the attack results on the Pneumonia dataset with confidence level of ResNet50: (a) One-pixel attack turning Pneumonia to Normal; (b) One-pixel attack turning Normal to Pneumonia; (c) Two-pixel attack turning Normal to Pneumonia; (d) Three-pixel attack turning Normal to Pneumonia.
Cancers 15 04228 g004
Figure 5. Examples of the attack results on the OCT dataset with confidence level of ResNet50: (a) Three-pixel attack turning CNV to Normal; (b) Three-pixel attack turning DME to Normal; (c) One-pixel attack turning Drusen to Normal; (d) Two-pixel attack turning Drusen to Normal.
Figure 5. Examples of the attack results on the OCT dataset with confidence level of ResNet50: (a) Three-pixel attack turning CNV to Normal; (b) Three-pixel attack turning DME to Normal; (c) One-pixel attack turning Drusen to Normal; (d) Two-pixel attack turning Drusen to Normal.
Cancers 15 04228 g005
Figure 6. Examples of the attack results on the OCT dataset with confidence level of ResNet50: (a) Three-pixel attack turning Drusen to Normal; (b) One-pixel untargeted attack on Normal images (c) Two-pixel untargeted attack on Normal images; (d) Three-pixel untargeted attack on Normal images.
Figure 6. Examples of the attack results on the OCT dataset with confidence level of ResNet50: (a) Three-pixel attack turning Drusen to Normal; (b) One-pixel untargeted attack on Normal images (c) Two-pixel untargeted attack on Normal images; (d) Three-pixel untargeted attack on Normal images.
Cancers 15 04228 g006
Figure 7. Examples of the attack results on the COVID-19 dataset with confidence level of ResNet50: (a) The original COVID-19 image; (b) One-pixel attack; (c) Two-pixel attack; (d) Three-pixel attack.
Figure 7. Examples of the attack results on the COVID-19 dataset with confidence level of ResNet50: (a) The original COVID-19 image; (b) One-pixel attack; (c) Two-pixel attack; (d) Three-pixel attack.
Cancers 15 04228 g007
Figure 8. Successful chest attack examples.
Figure 8. Successful chest attack examples.
Cancers 15 04228 g008
Figure 9. Chest attack cases: Disease to Normal type.
Figure 9. Chest attack cases: Disease to Normal type.
Cancers 15 04228 g009
Figure 10. Bar Chart of Number of Each Chest Disease.
Figure 10. Bar Chart of Number of Each Chest Disease.
Cancers 15 04228 g010
Figure 11. Wrongly labeled image.
Figure 11. Wrongly labeled image.
Cancers 15 04228 g011
Table 1. Overview of the dataset information.
Table 1. Overview of the dataset information.
NameNumber
of Data
Number
of Class
Original Size of ImageType of DatasetColor InformationType of Image
Derma10,0157600 × 450Multi-classColorDermatoscopic
Pneumonia58562(384~2916) × (127~2713)Multi-classGreyscaleX-ray
OCT109,3094(1, 3) × (384~1536) × (277~512)Multi-classGreyscaleOptical Coherence Tomography
COVID-1930,53031024 × 1024Multi-classGreyscaleX-ray
Chest112,120141024 × 1024Binary-class
Multi-label
GreyscaleX-ray
Table 2. Overview of the Derma dataset information.
Table 2. Overview of the Derma dataset information.
ClassCountPercentageDisease TypePrecisionRecall
Actinic keratoses and intraepithelial carcinoma3273.27%Disease57%62%
Basal cell carcinoma5145.13%Disease65%74%
Benign keratosis-like lesions109910.97%Normal70%61%
Dermatofibroma1151.15%Normal67%35%
Melanoma111311.11%Disease68%56%
Melanocytic nevi670566.95%Normal90%95%
Vascular lesions1421.42%Disease91%75%
Table 3. Overview of the COVID-19 dataset information.
Table 3. Overview of the COVID-19 dataset information.
ClassCountPercentageDisease TypePrecisionRecall
traintesttraintest
Normal808510026.83%25%Normal90%95%
Covid16,49020054.73%50%Disease99%91%
Pneumonia555510018.44%25%Disease91%94%
Table 4. Labels of Chest dataset.
Table 4. Labels of Chest dataset.
LabelDiseaseLabelDisease
0Atelectasis7Pneumothorax
1Cardiomegaly8Consolidation
2Effusion9Edema
3Infiltration10Emphysema
4Mass11Fibrosis
5Nodule12Pleural Thickening
6Pneumonia13Hernia
Table 5. Overview of the Chest dataset information.
Table 5. Overview of the Chest dataset information.
Disease countCountPercentageDisease Type
Normal11,92853.17%Normal
1625927.90%Disease
2291412.99%Disease
39894.41%Disease
42751.23%Disease
5540.24%Disease
6120.05%Disease
720.01%Disease
Table 6. Derma success rate.
Table 6. Derma success rate.
ClassBenign KeratosisMelanocytic NeviDermatofibromaMelanomaVascular Skin LesionsBasal Cell CarcinomaActinic Keratoses
Success rate41%4%80%16%10%32%41%
Table 7. Derma class transformation.
Table 7. Derma class transformation.
TransformBenign KeratosisMelanocytic NeviDermatofibromaMelanomaVascular Skin LesionsBasal Cell CarcinomaActinic Keratoses
Original
Benign Keratosis 1818095
Melanocytic Nevi3 00010
Dermatofibroma218 00573
Melanoma4110 010
Vascular Skin Lesions0900 10
Basal Cell Carcinoma712430 6
Actinic Keratoses10043024
Table 8. Derma ratio of class transformation.
Table 8. Derma ratio of class transformation.
ClassBenign KeratosisMelanocytic NeviDermatofibromaMelanomaVascular Skin LesionsBasal Cell CarcinomaActinic Keratoses
Transform ratio21.25%37.52%3.53%5.17%0.00%27.57%4.96%
Table 9. One-Pixel Class conversion type.
Table 9. One-Pixel Class conversion type.
DatasetClass ConversionCountPercentage
DermaDisease to Normal6142.36%
Normal to Disease8357.64%
ChestDisease to Normal15662.15%
Normal to Disease9537.85%
Table 10. Pneumonia: Attack Results.
Table 10. Pneumonia: Attack Results.
ResNet50One-PixelTwo-PixelThree-Pixel
Normal64%66%70%
Pneumonia0%0%0%
Table 11. OCT: Attack Results.
Table 11. OCT: Attack Results.
ResNet50One-PixelTwo-PixelThree-Pixel
Normal (untargeted)62%66%69%
CNV to Normal0%0%1%
DME to Normal0%0%1%
DRUSEN to Normal3%3%3%
Table 12. COVID-19: Results of RESNET-50.
Table 12. COVID-19: Results of RESNET-50.
ResNet50One-PixelTwo-PixelThree-Pixel
PneumoniaSR4%5%4%
AvgCL0.8360.8570.915
CovidSR1.5%2%2.5%
AvgCL0.7180.7200.653
Normal to DiseaseSR21%23%22%
AvgCL0.8350.8300.826
Table 13. COVID-19: Normal Images to Diseased Images.
Table 13. COVID-19: Normal Images to Diseased Images.
ResNet50One-PixelTwo-PixelThree-Pixel
CovidPneumoniaCovidPneumoniaCovidPneumonia
Proportion57%43%48%52%45%55%
AvgCL0.8780.7780.8870.7780.8620.796
Table 14. COVID-19: Results of DenseNet121.
Table 14. COVID-19: Results of DenseNet121.
DenseNet121One-PixelTwo-PixelThree-Pixel
PneumoniaSR4%6%7%
AvgCL0.5780.6160.630
CovidSR1%1.5%2%
AvgCL0.7110.6710.603
Normal to DiseaseSR18%17%16%
AvgCL0.7170.7050.738
Table 15. Chest success rate.
Table 15. Chest success rate.
ClassNormalInfiltrationAtelectasisEffusionNodule
Success rate95%85%94%87%91%
Table 16. Chest class transformation.
Table 16. Chest class transformation.
OriginalNormalInfiltrationAtelectasisEffusionNodule
Transform
Atelectasis133N/A35
Atelectasis, Effusion017150
Atelectasis, Effusion, Infiltration00100
Atelectasis, Infiltration291520
Atelectasis, Nodule00001
Cardiomegaly11100
Effusion816N/A4
Effusion, Infiltration140190
Effusion, Mass00011
Effusion, Mass, Pleural00020
Infiltration51N/A171317
Infiltration, Edema06000
Infiltration, Nodule250021
Mass131220
Mass, Nodule100014
Nodule2000N/A
NormalN/A54443028
Pneumothorax10100
Table 17. Color channel analysis.
Table 17. Color channel analysis.
Color ChannelAverage Attack Success RateAverage DE Time
Grayscale19%2 min 49 sec
Colored25%3 min 07 sec
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tsai, M.-J.; Lin, P.-Y.; Lee, M.-E. Adversarial Attacks on Medical Image Classification. Cancers 2023, 15, 4228. https://doi.org/10.3390/cancers15174228

AMA Style

Tsai M-J, Lin P-Y, Lee M-E. Adversarial Attacks on Medical Image Classification. Cancers. 2023; 15(17):4228. https://doi.org/10.3390/cancers15174228

Chicago/Turabian Style

Tsai, Min-Jen, Ping-Yi Lin, and Ming-En Lee. 2023. "Adversarial Attacks on Medical Image Classification" Cancers 15, no. 17: 4228. https://doi.org/10.3390/cancers15174228

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop