Evaluating Deep Learning Resilience in Retinal Fundus Classification with Generative Adversarial Networks Generated Images
Abstract
1. Introduction
- Training and testing of seven CNNs (one designed and developed by the authors), to evaluate the ability to distinguish real and GAN-generated images;
- In this way, some networks that provide optimal performances, guarantee better models for further diagnostic classification, within a better generalization and good resilience to adversarial images;
- The experimental analysis was conducted considering three input sizes of images, starting from 28 × 28 pixels of retinal images, going through 64 × 64 and finally resized to 128 × 128 pixels, from a validated and certified public dataset. This means the analysis was enlarged to different study cases, enhancing the initial variability.
2. Related Work
3. The Method
3.1. GAN: Operating Principles
3.2. The Approach
- Conv2D: this layer performs spatial convolution on 2D input data using learnable filters to extract features relevant to the task, such as image processing;
- MaxPooling2D: this layer is a fundamental component in CNNs, primarily used for down-sampling feature maps. It operates by sliding a window (often referred to as a “kernel” or “pooling window”) over the input feature map and selecting the maximum value within each window. This maximum value becomes the output for that particular region, thus summarizing the presence of certain features in that region;
- Flatten: its primary function is to convert the input data into a one-dimensional array, also known as a vector. This transformation is crucial for connecting the output of one layer to the input of another layer with a different shape, such as when transitioning from convolutional layers to fully connected layers;
- Dropout: this layer aims to prevent overfitting by randomly deactivating neurons during training. It randomly sets input units to 0 with a frequency of rate at each step during training time. It is aimed to prevent overfitting. Inputs not set to 0 are scaled up by 1/(1 − rate) such that the sum over all inputs is unchanged;
- Dense: this layer, also known as a fully connected layer, is a type of layer that connects every neuron in one layer to every neuron in the next layer. Dense layers in CNNs are used towards the end of the network architecture to transform the high-level features extracted by convolutional and pooling layers into predictions or decisions. It provide a way for the network to learn complex patterns and relationships in the data, making them a critical component of CNN architectures.
4. Experimental Analysis
- From a general point of view, when the dataset changes to 64 × 64 pixels, a model performance improves in terms of accuracy, precision, and recall;
- On the other hand, an increasing loss appears, but this behavior represents not relevant overfitting because this value depends also on the training loss; in all of cases the difference is negligible;
- The main contributions remain the improvement in the resilience of the networks, resulting in models with a greater generalization during the training, so in this case, the GAN application makes networks better classifiers for this kind of medical image.
- VGG19 model represents the worst model for this problem. The model allocates all images in the original class and completely does not distinguish real and fake retinal images.
5. Conclusions and Future Work
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
- Suganthi, K. Review of medical image synthesis using GAN techniques. ITM Web Conf. 2021, 37, 01005. [Google Scholar]
- Sun, L.; Chen, J.; Xu, Y.; Gong, M.; Yu, K.; Batmanghelich, K. Hierarchical amortized GAN for 3D high resolution medical image synthesis. IEEE J. Biomed. Health Inform. 2022, 26, 3966–3975. [Google Scholar] [CrossRef] [PubMed]
- Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef] [PubMed]
- Farahanipad, F.; Rezaei, M.; Nasr, M.S.; Kamangar, F.; Athitsos, V. A survey on GAN-based data augmentation for hand pose estimation problem. Technologies 2022, 10, 43. [Google Scholar] [CrossRef]
- Vyas, B.; Rajendran, R.M. Generative Adversarial Networks for Anomaly Detection in Medical Images. Int. J. Multidiscip. Innov. Res. Methodol. 2023, 2, 52–58. [Google Scholar]
- Xia, X.; Pan, X.; Li, N.; He, X.; Ma, L.; Zhang, X.; Ding, N. GAN-based anomaly detection: A review. Neurocomputing 2022, 493, 497–535. [Google Scholar] [CrossRef]
- Yu, W.; Lei, B.; Wang, S.; Liu, Y.; Feng, Z.; Hu, Y.; Shen, Y.; Ng, M.K. Morphological feature visualization of Alzheimer’s disease via multidirectional perception GAN. IEEE Trans. Neural Netw. Learn. Syst. 2022, 34, 4401–4415. [Google Scholar] [CrossRef] [PubMed]
- Lamba, S.; Baliyan, A.; Kukreja, V. GAN based image augmentation for increased CNN performance in Paddy leaf disease classification. In Proceedings of the 2022 2nd International Conference on Advance Computing and Innovative Technologies in Engineering (ICACITE), Piscataway, NJ, USA, 28–29 April 2022; pp. 2054–2059. [Google Scholar]
- Alotaibi, A. Deep generative adversarial networks for image-to-image translation: A review. Symmetry 2020, 12, 1705. [Google Scholar] [CrossRef]
- Yu, B.; Zhou, L.; Wang, L.; Shi, Y.; Fripp, J.; Bourgeat, P. Ea-GANs: Edge-aware generative adversarial networks for cross-modality MR image synthesis. IEEE Trans. Med. Imaging 2019, 38, 1750–1762. [Google Scholar] [CrossRef]
- Martinelli, F.; Mercaldo, F.; Di Giammarco, M.; Santone, A. Data Poisoning Attacks over Diabetic Retinopathy Images Classification. In Proceedings of the 2023 IEEE International Conference on Big Data (BigData), Sorrento, Italy, 15–18 December 2023; pp. 3698–3703. [Google Scholar]
- Rosenblatt, M.; Scheinost, D. Data poisoning attack and defenses in Connectome-Based predictive models. In Workshop on the Ethical and Philosophical Issues in Medical Imaging; Springer: Cham, Switzerland, 2022; pp. 3–13. [Google Scholar]
- You, A.; Kim, J.K.; Ryu, I.H.; Yoo, T.K. Application of generative adversarial networks (GAN) for ophthalmology image domains: A survey. Eye Vis. 2022, 9, 6. [Google Scholar] [CrossRef]
- Yu, Z.; Xiang, Q.; Meng, J.; Kou, C.; Ren, Q.; Lu, Y. Retinal image synthesis from multiple-landmarks input with generative adversarial networks. Biomed. Eng. Online 2019, 18, 62. [Google Scholar] [CrossRef] [PubMed]
- Chen, J.S.; Coyner, A.S.; Chan, R.P.; Hartnett, M.E.; Moshfeghi, D.M.; Owen, L.A.; Kalpathy-Cramer, J.; Chiang, M.F.; Campbell, J.P. Deepfakes in ophthalmology: Applications and realism of synthetic retinal images from generative adversarial networks. Ophthalmol. Sci. 2021, 1, 100079. [Google Scholar] [CrossRef] [PubMed]
- Costa, P.; Galdran, A.; Meyer, M.I.; Niemeijer, M.; Abràmoff, M.; Mendonça, A.M.; Campilho, A. End-to-end adversarial retinal image synthesis. IEEE Trans. Med. Imaging 2017, 37, 781–791. [Google Scholar] [CrossRef] [PubMed]
- Costa, P.; Galdran, A.; Meyer, M.I.; Abramoff, M.D.; Niemeijer, M.; Mendonça, A.M.; Campilho, A. Towards adversarial retinal image synthesis. arXiv 2017, arXiv:1701.08974. [Google Scholar]
- Diaz-Pinto, A.; Colomer, A.; Naranjo, V.; Morales, S.; Xu, Y.; Frangi, A.F. Retinal image synthesis and semi-supervised learning for glaucoma assessment. IEEE Trans. Med. Imaging 2019, 38, 2211–2218. [Google Scholar] [CrossRef] [PubMed]
- Chen, X.; Xu, Y.; Yan, S.; Wong, D.W.K.; Wong, T.Y.; Liu, J. Automatic feature learning for glaucoma detection based on deep learning. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Cham, Switzerland, 2015; pp. 669–677. [Google Scholar]
- Alghamdi, H.S.; Tang, H.L.; Waheeb, S.A.; Peto, T. Automatic optic disc abnormality detection in fundus images: A deep learning approach. In Ophthalmic Medical Image Analysis International Workshop; University of Iowa: Iowa, IA, USA, 2016; Volume 3. [Google Scholar]
- Divya, S.; Suresh, L.P.; John, A. Medical mr image synthesis using dcgan. In Proceedings of the 2022 First International Conference on Electrical, Electronics, Information and Communication Technologies (ICEEICT), Trichy, India, 16–18 February 2022; pp. 1–4. [Google Scholar]
- Lal, S.; Rehman, S.U.; Shah, J.H.; Meraj, T.; Rauf, H.T.; Damaševičius, R.; Mohammed, M.A.; Abdulkareem, K.H. Adversarial attack and defence through adversarial training and feature fusion for diabetic retinopathy recognition. Sensors 2021, 21, 3922. [Google Scholar] [CrossRef] [PubMed]
- Shin, H.C.; Tenenholtz, N.A.; Rogers, J.K.; Schwarz, C.G.; Senjem, M.L.; Gunter, J.L.; Andriole, K.P.; Michalski, M. Medical image synthesis for data augmentation and anonymization using generative adversarial networks. In Proceedings of the Simulation and Synthesis in Medical Imaging: Third International Workshop, SASHIMI 2018, Held in Conjunction with MICCAI 2018, Granada, Spain, 16 September 2018; Proceedings 3. Springer: Cham, Switzerland, 2018; pp. 1–11. [Google Scholar]
- Kwon, H.; Kim, Y.; Yoon, H.; Choi, D. Captcha image generation systems using generative adversarial networks. IEICE Trans. Inf. Syst. 2018, 101, 543–546. [Google Scholar] [CrossRef]
- Mangaokar, N.; Pu, J.; Bhattacharya, P.; Reddy, C.K.; Viswanath, B. Jekyll: Attacking medical image diagnostics using deep generative models. In Proceedings of the 2020 IEEE European Symposium on Security and Privacy (EuroS&P), Genoa, Italy, 7–11 September 2020; pp. 139–157. [Google Scholar]
- Nguyen, H.; Di Troia, F.; Ishigaki, G.; Stamp, M. Generative adversarial networks and image-based malware classification. J. Comput. Virol. Hacking Tech. 2023, 19, 579–595. [Google Scholar] [CrossRef]
- Renjith, G.; Laudanna, S.; Aji, S.; Visaggio, C.A.; Vinod, P. GANG-MAM: GAN based enGine for Modifying Android Malware. SoftwareX 2022, 18, 100977. [Google Scholar]
- Martinelli, F.; Mercaldo, F.; Santone, A. Evaluating the Impact of Generative Adversarial Network in Android Malware Detection. In Proceedings of the ENASE, Angers, France, 28–29 April 2024; pp. 590–597. [Google Scholar]
- Mercaldo, F.; Martinelli, F.; Santone, A. Deep Convolutional Generative Adversarial Networks in Image-Based Android Malware Detection. Computers 2024, 13, 154. [Google Scholar] [CrossRef]
- Nagaraju, R.; Stamp, M. Auxiliary-classifier GAN for malware analysis. In Artificial Intelligence for Cybersecurity; Springer: Cham, Switzerland, 2022; pp. 27–68. [Google Scholar]
- Hu, W.; Tan, Y. Generating adversarial malware examples for black-box attacks based on GAN. In Proceedings of the International Conference on Data Mining and Big Data, Beijing, China, 21–24 November 2022; pp. 409–423. [Google Scholar]
- Yuan, J.; Zhou, S.; Lin, L.; Wang, F.; Cui, J. Black-box adversarial attacks against deep learning based malware binaries detection with GAN. In ECAI 2020; IOS Press: Amsterdam, The Netherlands, 2020; pp. 2536–2542. [Google Scholar]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K.Q. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
- Wen, L.; Li, X.; Li, X.; Gao, L. A new transfer learning based on VGG-19 network for fault diagnosis. In Proceedings of the 2019 IEEE 23rd International Conference on Computer Supported Cooperative Work in Design (CSCWD), Porto, Portugal, 6–8 May 2019; pp. 205–209. [Google Scholar]
- Di Giammarco, M.; Iadarola, G.; Martinelli, F.; Mercaldo, F.; Ravelli, F.; Santone, A. Explainable Deep Learning for Alzheimer Disease Classification and Localisation. In Proceedings of the International Conference on Applied Intelligence and Informatics, Reggio Calabria, Italy, 1–3 September 2022. in press. [Google Scholar]
- Di Giammarco, M.; Iadarola, G.; Martinelli, F.; Mercaldo, F.; Santone, A. Explainable Retinopathy Diagnosis and Localisation by means of Class Activation Mapping. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–8. [Google Scholar]
- Xia, X.; Xu, C.; Nan, B. Inception-v3 for flower classification. In Proceedings of the 2017 2nd International Conference on Image, Vision and Computing (ICIVC), Chengdu, China, 2–4 June 2017; pp. 783–787. [Google Scholar]
- Tan, M.; Le, Q. Efficientnet: Rethinking model scaling for convolutional neural networks. In Proceedings of the International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019; pp. 6105–6114. [Google Scholar]
- Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. Mobilenets: Efficient convolutional neural networks for mobile vision applications. arXiv 2017, arXiv:1704.04861. [Google Scholar]
Layer | Type | Output Shape | Parameters |
---|---|---|---|
1 | InputLayer | (256, 256, 3) | 0 |
2 | Conv2D | (254, 254, 32) | 896 |
3 | MaxPooling2D | (127, 127, 32) | 0 |
4 | Conv2D | (125, 125, 64) | 18,496 |
5 | MaxPooling2D | (62, 62, 64) | 0 |
6 | Conv2D | (60, 60, 128) | 73,856 |
7 | MaxPooling2D | (30, 30, 128) | 0 |
8 | Flatten | (115, 200) | 0 |
9 | Dropout | (115, 200) | 0 |
10 | Dense | (512) | 58,982,912 |
11 | Dropout | (512) | 0 |
12 | Dense | (256) | 131,328 |
13 | Dropout | (256) | 0 |
14 | Dense | (2) | 514 |
CNN | Accuracy | Precision | Recall | F-Measure | AUC. | Loss |
---|---|---|---|---|---|---|
ResNet 50 | 0.99 | 0.99 | 099 | 0.99 | 0.99 | 0.04 |
DenseNet | 0.99 | 0.99 | 0.99 | 0.99 | 0.99 | 0.01 |
VGG19 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.69 |
Standard_CNN | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 5.65 × 10 |
Inception V3 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.69 |
MobileNet | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.12 |
EfficientNet | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 32.1 |
CNN | Accuracy | Precision | Recall | F-Measure | AUC. | Loss |
---|---|---|---|---|---|---|
ResNet 50 | 0.99 | 0.99 | 099 | 0.99 | 0.99 | 0.15 |
DenseNet | 0.99 | 0.99 | 0.99 | 0.99 | 0.99 | 0.07 |
VGG19 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.69 |
Standard_CNN | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 2.38 × 10 |
Inception V3 | 0.98 | 0.98 | 0.98 | 0.98 | 0.99 | 0.07 |
MobileNet | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 2.65 × 10 |
EfficientNet | 0.99 | 0.99 | 0.99 | 0.99 | 0.99 | 0.51 |
CNN | Accuracy | Precision | Recall | F-Measure | AUC. | Loss |
---|---|---|---|---|---|---|
ResNet 50 | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 2.326 × 10 |
DenseNet | 0.97 | 0.97 | 0.97 | 0.97 | 0.97 | 2.47 |
VGG19 | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 0.69 |
Standard_CNN | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.002 |
Inception V3 | 0.89 | 0.89 | 0.89 | 0.89 | 0.90 | 0.48 |
MobileNet | 1.0 | 1.0 | 1.0 | 1.0 | 1.0 | 0.001 |
EfficientNet | 0.5 | 0.5 | 0.5 | 0.5 | 0.5 | 19.21 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Di Giammarco, M.; Santone, A.; Cesarelli, M.; Martinelli, F.; Mercaldo, F. Evaluating Deep Learning Resilience in Retinal Fundus Classification with Generative Adversarial Networks Generated Images. Electronics 2024, 13, 2631. https://doi.org/10.3390/electronics13132631
Di Giammarco M, Santone A, Cesarelli M, Martinelli F, Mercaldo F. Evaluating Deep Learning Resilience in Retinal Fundus Classification with Generative Adversarial Networks Generated Images. Electronics. 2024; 13(13):2631. https://doi.org/10.3390/electronics13132631
Chicago/Turabian StyleDi Giammarco, Marcello, Antonella Santone, Mario Cesarelli, Fabio Martinelli, and Francesco Mercaldo. 2024. "Evaluating Deep Learning Resilience in Retinal Fundus Classification with Generative Adversarial Networks Generated Images" Electronics 13, no. 13: 2631. https://doi.org/10.3390/electronics13132631
APA StyleDi Giammarco, M., Santone, A., Cesarelli, M., Martinelli, F., & Mercaldo, F. (2024). Evaluating Deep Learning Resilience in Retinal Fundus Classification with Generative Adversarial Networks Generated Images. Electronics, 13(13), 2631. https://doi.org/10.3390/electronics13132631