Next Article in Journal
Decentralized Coordination of DERs for Dynamic Economic Emission Dispatch
Previous Article in Journal
Differential-Evolution-Assisted Optimization of Classical Compensation Topologies for 1 W Current-Fed IMD Wireless Charging Systems
Previous Article in Special Issue
SAFP-YOLO: Enhanced Object Detection Speed Using Spatial Attention-Based Filter Pruning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Editorial on the Special Issue: New Trends in Image Processing III

1
Department of Computer Science and Engineering, Sejong University, Seoul 05006, Republic of Korea
2
Centre for Visual Computing, University of Bradford, Bradford BD7 1DP, UK
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(22), 12430; https://doi.org/10.3390/app132212430
Submission received: 3 November 2023 / Accepted: 14 November 2023 / Published: 17 November 2023
(This article belongs to the Special Issue New Trends in Image Processing III)
The image processing field is undergoing a significant transformation owing to rapid advancements in deep learning, computer vision, and artificial intelligence. This has led to a surge in innovation in image classification [1], object detection [2,3,4], object segmentation [4,5], and image super-resolution [6], a crucial aspect of this domain that enables a more precise and versatile analysis of visual data. In recent years, image super-resolution has garnered significant interest, particularly in the domain of enhancing image quality. The primary objective of image super-resolution is to upgrade low-resolution and blurry images to high-resolution images with a primary focus on single-image super-resolution. Several trends have emerged in this field, including reconstruction-based methods [7,8,9]. These techniques have numerous applications, such as improving security in urban areas, object recognition, small object detection [10], face recognition in surveillance videos [11], medical diagnostics [12,13,14], remote sensing, astronomy [15,16,17], and microscopy image processing [18,19].
Furthermore, the domain of interest has observed a notable trend towards self-supervised learning, which enables models to extract robust features and representations from vast datasets, resulting in a reduced need for labeled data. These advancements have been applied in areas such as autonomous vehicles, medical imaging, and security surveillance. Recent breakthroughs include achieving higher precision and objectivity in disease classification and tumor microenvironment descriptions [20], outperforming CNNs in image classification with a pure transformer [21], prominent applications for remote sensing imagery, such as land use/land cover classification [22], strong zero-shot classification performance based on natural language names [23], and guidelines for and evaluation of clinically explainable AI in medical image analysis [24]. However, this approach still faces challenges, particularly in the case of limited labeled samples, which is addressed through CNN-based frameworks [25], especially in hyperspectral image classification [26]. In this research, the utilization of lightweight networks as student models and Emonet as teacher models integrates knowledge distillation, resulting in an efficient facial expression detection system with a minimal decrease in performance and notable improvements in computational efficiency, making it suitable for practical applications [27]. There are significant improvements in real-time image detection with works such as the state-of-the-art YOLO framework, which can be implemented in real-time on edge computing platforms [28], recent innovations in image classification approaches have seen the incorporation of quantum computing, offering groundbreaking solutions to address these challenges [7], and the paper proposes a blockchain-empowered privacy-preserving federated learning (PPFL) for a remote sensing image classification framework with the poisonous dishonest majority, which can defend against encrypted model poisoning attacks without compromising users’ privacy [29]. To enhance the execution speed and reduce the number of filters in low-cost embedded boards, the paper introduces Spatial Attention-based Filter Pruning (SAFP) for YOLOv4 and YOLOv7. This innovative approach leads to a significant improvement in processing speed without compromising the accuracy of real-time object detection [30].
These recent references reflect image classification’s dynamic and exciting directions in 2023. Additionally, image segmentation also plays a vital role in image processing and computer vision. The development of image segmentation methods is closely connected to several disciplines and fields, e.g., industrial inspection [31], intelligent medical technology [32], augmented reality [33], and autonomous vehicles [34]. Image segmentation is currently witnessing a remarkable surge in trends and techniques, with an expanding emphasis on leveraging deep learning models, such as graph neural networks (GNNs) [35], and attention mechanisms [36]. Furthermore, the combination of generative adversarial networks (GANs) [37] and reinforcement learning [38] holds promise for accomplishing more robust and adaptable segmentation models. As research continues to push the boundaries of image segmentation, the convergence of these cutting-edge techniques is poised to revolutionize not only the field of computer vision but also numerous industries reliant on accurate and efficient image analysis.
In the field of computer vision, the success of transformer networks in natural language processing tasks [39] has inspired researchers to apply this architecture to computer vision tasks, resulting in the development of Vision Transformers (ViT) [40]. In ViT, global attention is directed towards 16 × 16 patches of the entire image, focusing on the global salient features of the image, and resolving the long-range dependency among the image’s contents. This approach effectively balances the incorporation of the global context and maintains computational efficiency. The potential of vision transformers to address various problems in computer vision has grabbed the interest of numerous researchers, and their applications continue to be a subject of ongoing research. Vision transformers (ViTs) have introduced a novel and groundbreaking approach to object detection when integrated with frameworks like DETR (detection transformers (DETRs) [41]. ViTs capture intricate spatial relationships between objects within an image, and ViTs effectively analyze how different objects relate to each other and their context within the scene [42]. The utilization of DETR in conjunction with ViTs is of utmost importance in object detection because it enables a comprehensive understanding of the image, which is essential for identifying and localizing objects. This approach allows for simultaneous attention to all parts of the image, significantly improving the accuracy and efficiency of the detection tasks. The significant advances in object detection achieved using this approach highlight its potential. Researchers are actively exploring ways to refine these models for various computer-vision challenges, making object detection a promising and evolving field of research.
In the future, we anticipate a substantial change towards more advanced and multi-functional image processing techniques. This transformation will be propelled by rapid developments in artificial intelligence, machine learning, and computational capabilities. Key trends include the integration of deep learning models into image processing systems, resulting in increased accuracy and versatility in tasks such as image segmentation, classification, image super-resolution, and object detection. The increasing demand for real-time and efficient processing will give rise to novel solutions that leverage hardware acceleration and cloud computing. In summary, the future of image processing is likely to be a dynamic and transformative journey that will redefine how we interact with and harness visual data across various domains.

Author Contributions

Conceptualization, H.M. and I.M.; methodology, I.M.; validation, H.M.; formal analysis, H.M.; investigation, I.M.; resources, H.M.; writing—original draft preparation, I.M.; writing—review and editing, H.M.; supervision, H.M. All authors have read and agreed to the published version of the manuscript.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dimitrovski, I.; Kitanovski, I.; Kocev, D.; Simidjievski, N. Current trends in deep learning for Earth Observation: An open-source benchmark arena for image classification. ISPRS J. Photogramm. Remote Sens. 2023, 197, 18–35. [Google Scholar]
  2. Menezes, A.G.; de Moura, G.; Alves, C.; de Carvalho, A.C. Continual Object Detection: A review of definitions, strategies, and challenges. Neural Netw. 2023, 161, 476–493. [Google Scholar] [CrossRef]
  3. Kamath, V.; Renuka, A. Deep Learning Based Object Detection for Resource Constrained Devices-Systematic Review, Future Trends and Challenges Ahead. Neurocomputing 2023, 531, 34–60. [Google Scholar] [CrossRef]
  4. Talaat, F.M.; ZainEldin, H. An improved fire detection approach based on YOLO-v8 for smart cities. Neural Comput. Appl. 2023, 35, 20939–20954. [Google Scholar] [CrossRef]
  5. Lu, Y.; Jiang, Q.; Chen, R.; Hou, Y.; Zhu, X.; Ma, Y. See more and know more: Zero-shot point cloud segmentation via multi-modal visual data. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–6 October 2023; pp. 21674–21684. [Google Scholar] [CrossRef]
  6. Lepcha, D.C.; Goyal, B.; Dogra, A.; Goyal, V. Image super-resolution: A comprehensive review, recent trends, challenges and applications. Inf. Fusion 2023, 91, 230–260. [Google Scholar] [CrossRef]
  7. Alsubai, S.; Alqahtani, A.; Binbusayyis, A.; Sha, M.; Gumaei, A.; Wang, S. Quantum Computing Meets Deep Learning: A Promising Approach for Diabetic Retinopathy Classification. Mathematics 2023, 11, 2008. [Google Scholar]
  8. Tan, R.; Yuan, Y.; Huang, R.; Luo, J. Video super-resolution with spatial-temporal transformer encoder. In Proceedings of the 2022 IEEE International Conference on Multimedia and Expo (ICME), Taipei, Taiwan, 18–22 July 2022. [Google Scholar]
  9. Li, H.; Yang, Y.; Chang, M.; Chen, S.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Srdiff: Single image super-resolution with diffusion probabilistic models. Neurocomputing 2022, 479, 47–59. [Google Scholar] [CrossRef]
  10. Li, H.; Zhang, P. Spatio-temporal fusion network for video super-resolution. In Proceedings of the 2021 International Joint Conference on Neural Networks (IJCNN), Shenzhen, China, 18–22 July 2021; pp. 1–9. [Google Scholar] [CrossRef]
  11. Li, B.; Li, X.; Lu, Y.; Liu, S.; Feng, R.; Chen, Z. HST: Hierarchical Swin Transformer for Compressed Image Super-Resolution. In Computer Vision—ECCV 2022 Workshops. ECCV 2022. Lecture Notes in Computer Science; Karlinsky, L., Michaeli, T., Nishino, K., Eds.; Springer: Cham, Switzerland; Tel Aviv, Israel, 2023; Volume 13802. [Google Scholar] [CrossRef]
  12. Lepcha, D.C.; Goyal, B. Medical-modality super-resolution for increased visualisation of intracranial tissue details and structural details. In Proceedings of the 2021 9th International Conference on Reliability, Infocom Technologies and Optimization (Trends and Future Directions) (ICRITO), Noida, India, 3–4 September 2021. [Google Scholar] [CrossRef]
  13. Lepcha, D.C.; Goyal, B.; Dogra, A.; Wang, S.H. An efficient medical image super resolution based on piecewise linear regression strategy using domain transform filtering. Concurr. Comput. Pract. Exp. 2022, 34, e6644. [Google Scholar] [CrossRef]
  14. Qiu, D.; Cheng, Y.; Wang, X. Medical image super-resolution reconstruction algorithms based on deep learning: A survey. Comput. Methods Programs Biomed. 2023, 238, 107590. [Google Scholar] [CrossRef]
  15. Yang, X.; Fan, J.; Wu, C.; Zhou, D.; Li, T. NasmamSR: A fast image super-resolution network based on neural architecture search and multiple attention mechanism. Multimed. Syst. 2022, 28, 321–334. [Google Scholar] [CrossRef]
  16. Wang, X.; Sun, L.; Chehri, A.; Song, Y. A Review of GAN-Based Super-Resolution Reconstruction for Optical Remote Sensing Images. Remote Sens. 2023, 15, 5062. [Google Scholar] [CrossRef]
  17. Zhang, B.; Ma, M.; Wang, M.; Hong, D.; Yu, L.; Wang, J.; Gong, P.; Huang, X. Enhanced resolution of FY4 remote sensing visible spectrum images utilizing super-resolution and transfer learning techniques. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7391–7399. [Google Scholar] [CrossRef]
  18. Qiao, C.; Li, D.; Liu, Y.; Zhang, S.; Liu, K.; Liu, C.; Guo, Y.; Jiang, T.; Fang, C.; Li, N. Rationalized deep learning super-resolution microscopy for sustained live imaging of rapid subcellular processes. Nat. Biotechnol. 2023, 41, 367–377. [Google Scholar] [CrossRef]
  19. Fu, P.; Cao, W.; Chen, T.; Huang, X.; Le, T.; Zhu, S.; Wang, D.-W.; Lee, H.J.; Zhang, D. Super-resolution imaging of non-fluorescent molecules by photothermal relaxation localization microscopy. Nat. Photonics 2023, 17, 330–337. [Google Scholar] [CrossRef]
  20. Filiot, A.; Ghermi, R.; Olivier, A.; Jacob, P.; Fidon, L.; Mac Kain, A.; Saillard, C.; Schiratti, J.-B. Scaling Self-Supervised Learning for Histopathology with Masked Image Modeling. medRxiv 2023. [Google Scholar] [CrossRef]
  21. Jamil, S.; Jalil Piran, M.; Kwon, O.-J. A comprehensive survey of transformers for computer vision. Drones 2023, 7, 287. [Google Scholar] [CrossRef]
  22. Papoutsis, I.; Bountos, N.I.; Zavras, A.; Michail, D.; Tryfonopoulos, C. Benchmarking and scaling of deep learning models for land cover image classification. ISPRS J. Photogramm. Remote Sens. 2023, 195, 250–268. [Google Scholar] [CrossRef]
  23. Novack, Z.; McAuley, J.; Lipton, Z.C.; Garg, S. Chils: Zero-shot image classification with hierarchical label sets. In Proceedings of the International Conference on Machine Learning, Honolulu, HI, USA, 23–29 July 2023; Volume 202, pp. 26342–26362. [Google Scholar]
  24. Jin, W.; Li, X.; Fatehi, M.; Hamarneh, G. Guidelines and evaluation of clinical explainable AI in medical image analysis. Med. Image Anal. 2023, 84, 102684. [Google Scholar] [CrossRef] [PubMed]
  25. Zhang, M.; Wang, Z.; Wang, X.; Gong, M.; Wu, Y.; Li, H. Features kept generative adversarial network data augmentation strategy for hyperspectral image classification. Pattern Recognit. 2023, 142, 109701. [Google Scholar] [CrossRef]
  26. Yao, D.; Zhi-li, Z.; Xiao-feng, Z.; Wei, C.; Fang, H.; Yao-ming, C.; Cai, W.-W. Deep hybrid: Multi-graph neural network collaboration for hyperspectral image classification. Def. Technol. 2023, 23, 164–176. [Google Scholar] [CrossRef]
  27. Lee, K.; Kim, S.; Lee, E.C. Fast and Accurate Facial Expression Image Classification and Regression Method Based on Knowledge Distillation. Appl. Sci. 2023, 13, 6409. [Google Scholar] [CrossRef]
  28. Liu, S.; Zha, J.; Sun, J.; Li, Z.; Wang, G. EdgeYOLO: An Edge-Real-Time Object Detector. arXiv 2023, arXiv:2302.07483. [Google Scholar]
  29. Zhu, J.; Wu, J.; Bashir, A.K.; Pan, Q.; Wu, Y. Privacy-Preserving Federated Learning of Remote Sensing Image Classification with Dishonest-Majority. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 4685–4698. [Google Scholar] [CrossRef]
  30. Ahn, H.; Son, S.; Roh, J.; Baek, H.; Lee, S.; Chung, Y.; Park, D. SAFP-YOLO: Enhanced Object Detection Speed Using Spatial Attention-Based Filter Pruning. Appl. Sci. 2023, 13, 11237. [Google Scholar] [CrossRef]
  31. Nooralishahi, P.; Rezayiye, R.K.; López, F.; Maldague, X.P. PHM-IRNET: Self-training thermal segmentation approach for thermographic inspection of industrial components. NDT E Int. 2023, 138, 102884. [Google Scholar] [CrossRef]
  32. Chen, Y.; Ge, P.; Wang, G.; Weng, G.; Chen, H. An overview of intelligent image segmentation using active contour models. Intell. Robot. 2023, 3, 23–55. [Google Scholar] [CrossRef]
  33. Stancato, G.; Piga, B.E.A. Image Segmentation and Emotional Analysis of Virtual and Augmented Reality Urban Scenes. In Beyond Digital Representation: Advanced Experiences in AR and AI for Cultural Heritage and Innovative Design; Springer: Berlin/Heidelberg, Germany, 2023; pp. 443–458. [Google Scholar]
  34. Kabiraj, A.; Pal, D.; Ganguly, D.; Chatterjee, K.; Roy, S. Number plate recognition from enhanced super-resolution using generative adversarial network. Multimed. Tools Appl. 2023, 82, 13837–13853. [Google Scholar] [CrossRef]
  35. Jiang, J.; Chen, X.; Tian, G.; Liu, Y. ViG-UNet: Vision Graph Neural Networks for Medical Image Segmentation. arXiv 2023, arXiv:2306.04905. [Google Scholar]
  36. Huang, M.; Dai, W.; Yan, W.; Wang, J. High-Resolution Remote Sensing Image Segmentation Algorithm Based on Improved Feature Extraction and Hybrid Attention Mechanism. Electronics 2023, 12, 3660. [Google Scholar] [CrossRef]
  37. Zhang, J.; Zhang, S.; Shen, X.; Lukasiewicz, T.; Xu, Z. Multi-ConDoS: Multimodal contrastive domain sharing generative adversarial networks for self-supervised medical image segmentation. IEEE Trans. Med. Imaging 2023, 11, 74805–74815. [Google Scholar] [CrossRef]
  38. Xu, G.; Wang, S.; Lukasiewicz, T.; Xu, Z. Adaptive-Masking Policy with Deep Reinforcement Learning for Self-Supervised Medical Image Segmentation. In Proceedings of the 2023 IEEE International Conference on Multimedia and Expo (ICME), Brisbane, Australia, 10–14 July 2023; pp. 2285–2290. [Google Scholar] [CrossRef]
  39. Nassiri, K.; Akhloufi, M. Transformer models used for text-based question answering systems. Appl. Intell. 2023, 53, 10602–10635. [Google Scholar] [CrossRef]
  40. Liu, H.; Huo, G.; Li, Q.; Guan, X.; Tseng, M.-L. Multiscale lightweight 3D segmentation algorithm with attention mechanism: Brain tumor image segmentation. Expert Syst. Appl. 2023, 214, 119166. [Google Scholar] [CrossRef]
  41. Mia, M.S.; Arnob, A.B.H.; Naim+, A.; Voban, A.A.B.; Islam, M.S. ViTs are Everywhere: A Comprehensive Study Showcasing Vision Transformers in Different Domain. arXiv 2023, arXiv:2310.05664. [Google Scholar]
  42. Guo, M.-H.; Lu, C.-Z.; Liu, Z.-N.; Cheng, M.-M.; Hu, S.-M. Visual attention network. Comput. Vis. Media 2023, 9, 733–752. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Moon, H.; Mehmood, I. Editorial on the Special Issue: New Trends in Image Processing III. Appl. Sci. 2023, 13, 12430. https://doi.org/10.3390/app132212430

AMA Style

Moon H, Mehmood I. Editorial on the Special Issue: New Trends in Image Processing III. Applied Sciences. 2023; 13(22):12430. https://doi.org/10.3390/app132212430

Chicago/Turabian Style

Moon, Hyeonjoon, and Irfan Mehmood. 2023. "Editorial on the Special Issue: New Trends in Image Processing III" Applied Sciences 13, no. 22: 12430. https://doi.org/10.3390/app132212430

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop