An Integrated Active Deep Learning Approach for Image Classification from Unlabeled Data with Minimal Supervision
Abstract
:1. Introduction
- An innovative active learning approach: The proposed approach utilizes uncertainty sampling to select informative unlabeled data points for labeling and retraining with pseudo-labeling of confident data, efficiently expanding the training set.
- Building a robust image classifier with fewer labeled samples: By strategically picking and labeling the most valuable samples in a loop, the model achieves high accuracy with fewer manually labeled samples compared to traditional supervised deep learning approaches.
2. Literature Review
3. Methodology
3.1. Base Classifier
3.2. Proposed Approach
3.2.1. Initialization Methods
3.2.2. Choosing Samples with Uncertainty
- By iteratively selecting uncertain samples for manual annotation and pseudo-labeling highly confident samples, the training data are expanded selectively while minimizing labeling effort. In the case of “Least Confidence ()”, you should organize the unannotated data samples by sorting them in ascending order according to their values, which are calculated as follows:
- 2.
- With margin sampling [33], unlabeled samples are sorted in ascending order based on the margin between the model’s top two predicted class probabilities. For each sample , the class probability is computed for every class . The highest and second-highest probabilities are identified. The margin is calculated as their difference as shown in Equation (3):
- 3.
- Entropy measures the uncertainty in the model’s predicted class probability distribution for a given unlabeled sample. Samples with high entropy are more ambiguous and informative for the model [33]. Entropy uses the predicted probabilities across all classes to quantify the uncertainty as follows:
3.2.3. Fine-Tuning of the Deep Convolutional Neural Network Model
Algorithm 1: Proposed approach |
Input Unannotated samples X Initially annotated samples Y; Number of uncertain samples K High confidence threshold for sample selection Threshold decay rate Optimal iteration count T Fine-tuning intervals Output CNN network parameters Algorithm Begin Initialize using Y. Set iteration to 0. While iteration < T do: Add K uncertain instances to Y based on Equations (2) or (3) or (4). Obtain pseudo-labeled samples Z according to Equation (5). If iteration % == 0 then Update through fine-tuning using Equation (6). Update based on Equation (9). Increment iteration by 1 Return End |
4. Experimental Results and Discussion
4.1. Datasets Description
4.2. Experimental Setup
4.3. Results and Discussion
5. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Guo, J.; Wang, Q. Informativeness-guided active learning for deep learning–based façade defects detection. Comput.-Aided Civ. Infrastruct. Eng. 2023, 38, 123–135. [Google Scholar] [CrossRef]
- Yao, X.; Guo, Q.; Li, A. Cloud Detection in Optical Remote Sensing Images with Deep Semi-supervised and Active Learning. IEEE Geosci. Remote Sens. Lett. 2023, 20, 45–57. [Google Scholar] [CrossRef]
- Kang, C.J.; Peter WC, H.; Siang, T.P.; Jian, T.T.; Li, Z.; Wang, Y.-H. An active learning framework featured Monte Carlo dropout strategy for deep learning-based semantic segmentation of concrete cracks from images. Struct. Health Monit. 2023, 22, 14759217221150376. [Google Scholar] [CrossRef]
- Li, X.; Du, M.; Zuo, S.; Zhou, M.; Peng, Q.; Chen, Z.; Zhou, J.; He, Q. Deep convolutional neural networks using an active learning strategy for cervical cancer screening and diagnosis. Front. Bioinform. 2023, 3, 1101667. [Google Scholar] [CrossRef] [PubMed]
- Guan, X.; Li, Z.; Zhou, Y.; Shao, W.; Zhang, D. Active learning for efficient analysis of high throughput nanopore data. Bioinformatics 2023, 39, btac764. [Google Scholar] [CrossRef] [PubMed]
- Zhao, C.; Qin, B.; Feng, S.; Zhu, W.; Sun, W.; Li, W.; Jia, X. Hyperspectral image classification with multi-attention transformer and adaptive superpixel segmentation-based active learning. IEEE Trans. Image Process. 2023, 32, 3606–3621. [Google Scholar] [CrossRef] [PubMed]
- Gu, X.; Lu, W.; Ao, Y.; Li, Y.; Song, C. Seismic Stratigraphic Interpretation Based on Deep Active Learning. IIEEE Trans. Geosci. Remote. Sens. 2023, 61, 1–11. [Google Scholar] [CrossRef]
- Fu, X.; Cao, H.; Hu, H.; Lian, B.; Wang, Y.; Huang, Q.; Wu, Y. Attention-Based Active Learning Framework for Segmentation of Breast Cancer in Mammograms. Appl. Sci. 2023, 13, 852. [Google Scholar] [CrossRef]
- Yuan, D.; Chang, X.; Liu, Q.; Yang, Y.; Wang, D.; Shu, M.; He, Z.; Shi, G. Active learning for deep visual tracking. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef]
- Liu, Y.; Lou, J.; Wang, M.; Zhang, Q.; Ye, J.; Wu, T. Active learning for truss structure design. ACM Trans. Graph. (TOG) 2021, 40, 1–14. [Google Scholar]
- Wu, Y.; Kirchhoff, K.F.; Bilenko, M. Active learning for ML enhanced scientific simulation codes. Int. Conf. Mach. Learn. 2021, 77, 11279–11289. [Google Scholar]
- Joshi, M.; Sahoo, D.; Hoi, S.C.; Li, J. Online active learning: A review. arXiv 2021, arXiv:2103.12857. [Google Scholar]
- Tanno, R.; Saeedi, A.; Sankaranarayanan, S.; Alexander, D.C.; Arridge, S. Learning from noisy labels by regularization with virtual adversarial perturbations. IEEE Trans. Med. Imaging 2021, 41, 137–153. [Google Scholar]
- Ren, P.; Xiao, Y.; Chang, X.; Huang, P.Y.; Li, Z.; Gupta, B.B.; Chen, X.; Wang, X. A survey of deep active learning. ACM Comput. Surv. 2021, 54, 1–40. [Google Scholar] [CrossRef]
- Shi, J.; Wang, Q.; Zhang, Q. Active deep metric learning. Proc. AAAI Conf. Artif. Intell. 2021, 35, 4634–4641. [Google Scholar]
- Chen, X.; Song, Y.; Islam, A.; Ren, X.; Ryu, J.K. Noise-aware unsupervised domain adaptation via stochastic conditional shift embedding. Proc. AAAI Conf. Artif. Intell. 2021, 35, 3980–3988. [Google Scholar]
- Jiang, L.; Meng, D.; Zhao, Q.; Shan, S.; Hauptmann, A.G. Self-paced curriculum learning. In Proceedings of the Twenty-ninth AAAI conference on artificial intelligence, Austin, TX, USA, 25–30 January 2015. [Google Scholar]
- Wang, K.; Zhang, D.; Li, Y.; Zhang, R.; Lin, L. Cost-effective active learning for deep image classification. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 2591–2600. [Google Scholar] [CrossRef]
- Gal, Y.; Ghahramani, Z. Bayesian convolutional neural networks with Bernoulli approximate variational inference. In Proceedings of the International Conference on Learning Representations (ICLR) Workshop Track, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
- Yoo, D.; Kweon, I.S. Learning loss for active learning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–19 June 2019; pp. 93–102. [Google Scholar]
- Jamal, M.A.; Li, S.; Mong, S.; An, G.; Shuai, Q.; Vasconcelos, N. Rethinking class balanced self-training. arXiv 2020, arXiv:2010.08806. [Google Scholar]
- Munjal, B.; Chakraborty, S.; Goyal, P.K. Towards efficient active learning for video classification using temporal coherence. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 17362–17371. [Google Scholar]
- Zhang, Y.; Ding, Z.; Li, J.; Ogunbona, P.; Xu, D.; Wang, H. Importance-aware semantic segmentation for autonomous driving. Proc. AAAI Conf. Artif. Intell. 2020, 34, 13046–13053. [Google Scholar]
- Sofiiuk, K.; Barinova, O.; Konushin, A.; Aliev, T.; Vetrov, D.P. f-ALDA: F-divergences minimization for active learning. arXiv 2021, arXiv:2103.08333. [Google Scholar]
- Tanno, R.; Arulkumaran, K.; Alexander, D.C.; Criminisi, A.; Nori, A. Adaptive neural trees. Int. Conf. Mach. Learn. 2020, 9438–9447. [Google Scholar] [CrossRef]
- Chen, X.; Pu, X.; Chen, Z.; Li, L.; Zhao, K.N.; Liu, H.; Zhu, H. Application of EfficientNet-B0 and GRU-based deep learning on classifying the colposcopy diagnosis of precancerous cervical lesions. Cancer Med. 2023, 12, 8690–8699. [Google Scholar] [CrossRef] [PubMed]
- Zhu, W.; Hu, J.; Sun, G.; Cao, X.; Qian, X. A/B test: Towards rapid traffic splitting for personalized web service. In Proceedings of the Web Conference, Taipei, Taiwan, 20–24 April 2020; pp. 78–88. [Google Scholar]
- Raza, R.; Zulfiqar, F.; Owais Khan, M.; Arif, M.; Alvi, A.; Iftikhar, M.A.; Alam, T. Lung-EffNet: Lung cancer classification using EfficientNet from CT-scan images. Eng. Appl. Artif. Intell. 2023, 126, 106902. [Google Scholar] [CrossRef]
- Jiang, Y.; Huang, D.; Zhang, C. Beyond synthetic noise: Deep learning on controlled noisy labels. Int. Conf. Mach. Learn. 2020, 4804–4815. [Google Scholar] [CrossRef]
- Wu, Y.; Winston, E.; Kaushik, D.; Lipton, Z. Domain adaptation with asymmetrically relaxed distribution alignment. Int. Conf. Mach. Learn. 2020, 10283–10293. [Google Scholar] [CrossRef]
- Biswas, S.; Pal, S.; Maity, S.P.; Balasubramanian, V.N. Effects of noisy labels on deep neural network architectures. Neural Netw. 2020, 133, 19–29. [Google Scholar]
- Wang, Y.; Wang, H.; Shen, Y.; Fei, J.; Li, W.; Jin, G.; Wu, L.; Zhao, R.; Le, X. Semi-supervised semantic segmentation using unreliable pseudo-labels. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4248–4257. [Google Scholar]
- Chen, S.; Wang, R.; Lu, J. A meta-framework for multi-label active learning based on deep reinforcement learning. Neural Netw. 2023, 162, 258–270. [Google Scholar] [CrossRef]
- Cacciarelli, D.; Kulahci, M. A survey on online active learning. arXiv 2023, arXiv:2302.08893. [Google Scholar]
- Tan, F.; Zheng, G. Active learning for deep object detection by fully exploiting unlabeled data. Connect. Sci. 2023, 35, 2195596. [Google Scholar] [CrossRef]
- Shen, X.; Dai, Q.; Ullah, W. An active learning-based incremental deep-broad learning algorithm for unbalanced time series prediction. Inf. Sci. 2023, 642, 119103. [Google Scholar] [CrossRef]
Labeled Samples | Accuracy | Precision | Recall | F1 Score | ||||
---|---|---|---|---|---|---|---|---|
Proposed | Conventional | Proposed | Conventional | Proposed | Conventional | Proposed | Conventional | |
9% | 64.7% | 61.7% | 74% | 71% | 75% | 73% | 0.75 | 0.72 |
18% | 86.4% | 74.1% | 88% | 80% | 85% | 78% | 0.87 | 0.79 |
27% | 90.3% | 80.2% | 91% | 84% | 91% | 84% | 0.91 | 0.84 |
36% | 94.6% | 84.5% | 94% | 87% | 94% | 87% | 0.94 | 0.87 |
45% | 96.3% | 87.1% | 96% | 90% | 99% | 94% | 0.98 | 0.92 |
54% | 96.5% | 88.5% | 93% | 91% | 94% | 92% | 0.93 | 0.92 |
63% | 97.5% | 88.9% | 96% | 91% | 95% | 90% | 0.95 | 0.91 |
72% | 98.5% | 89.8% | 98% | 92% | 96% | 91% | 0.97 | 0.92 |
81% | 98.7% | 91.2% | 99% | 94% | 97% | 93% | 0.98 | 0.94 |
90% | 98.9% | 91.9% | 99% | 95% | 98% | 94% | 0.98 | 0.95 |
100% | 98.9% | 92.3% | 99% | 96% | 99% | 96% | 0.99 | 0.96 |
Labeled Samples | Accuracy | Precision | Recall | F1 Score | ||||
---|---|---|---|---|---|---|---|---|
Proposed | Conventional | Proposed | Conventional | Proposed | Conventional | Proposed | Conventional | |
9% | 71.7% | 57.7% | 78% | 61% | 80% | 63% | 0.79 | 0.62 |
18% | 86.4% | 61.7% | 87% | 65% | 82% | 61% | 0.84 | 0.63 |
27% | 92.3% | 62.2% | 92% | 66% | 94% | 66% | 0.93 | 0.66 |
36% | 94.6% | 65.5% | 94% | 69% | 94% | 69% | 0.94 | 0.69 |
45% | 95.3% | 67.1% | 95% | 71% | 96% | 74% | 0.96 | 0.72 |
54% | 96.5% | 70.5% | 96% | 73% | 96% | 76% | 0.96 | 0.75 |
63% | 97.5% | 71.9% | 97% | 74% | 97% | 73% | 0.97 | 0.74 |
72% | 98.5% | 72.8% | 98% | 75% | 97% | 74% | 0.98 | 0.75 |
81% | 98.7% | 73.2% | 99% | 76% | 98% | 75% | 0.98 | 0.75 |
90% | 98.9% | 73.9% | 99% | 76% | 99% | 75% | 0.99 | 0.76 |
100% | 99.3% | 74.3% | 99% | 77% | 99% | 77% | 0.99 | 0.77 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Abdelwahab, A.; Afifi, A.; Salama, M. An Integrated Active Deep Learning Approach for Image Classification from Unlabeled Data with Minimal Supervision. Electronics 2024, 13, 169. https://doi.org/10.3390/electronics13010169
Abdelwahab A, Afifi A, Salama M. An Integrated Active Deep Learning Approach for Image Classification from Unlabeled Data with Minimal Supervision. Electronics. 2024; 13(1):169. https://doi.org/10.3390/electronics13010169
Chicago/Turabian StyleAbdelwahab, Amira, Ahmed Afifi, and Mohamed Salama. 2024. "An Integrated Active Deep Learning Approach for Image Classification from Unlabeled Data with Minimal Supervision" Electronics 13, no. 1: 169. https://doi.org/10.3390/electronics13010169
APA StyleAbdelwahab, A., Afifi, A., & Salama, M. (2024). An Integrated Active Deep Learning Approach for Image Classification from Unlabeled Data with Minimal Supervision. Electronics, 13(1), 169. https://doi.org/10.3390/electronics13010169