Next Article in Journal
FishDet-YOLO: Enhanced Underwater Fish Detection with Richer Gradient Flow and Long-Range Dependency Capture through Mamba-C2f
Previous Article in Journal
Improved Control Strategy for Dual-PWM Converter Based on Equivalent Input Disturbance
Previous Article in Special Issue
A UAV Aerial Image Target Detection Algorithm Based on YOLOv7 Improved Model
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Editorial

Applications of Computer Vision, 2nd Edition

by
Eva Cernadas
CiTIUS (Singular Research Center on Intelligent Technologies), University of Santiago de Compostela, 15782 Santiago de Compostela, Spain
Electronics 2024, 13(18), 3779; https://doi.org/10.3390/electronics13183779
Submission received: 14 September 2024 / Accepted: 19 September 2024 / Published: 23 September 2024
(This article belongs to the Special Issue Applications of Computer Vision, 2nd Edition)

1. Introduction to the Applications of Computer Vision

Computer vision (CV) is a broad term mainly used to refer to processing image and video data. CV aims to enable machines to perceive, observe, and understand the physical world as if they have human eyes. While this area of knowledge began to develop during the 1970s and 1980s, the last three decades have been characterized by the field maturing. This progress can be seen in the increasing number of software and hardware products on the market, the significant growth of active applications, and the rise in recent scientific publications on this research area. The first applications of computer vision were in the field of medical imaging and the processing of remote sensing data. Hence, the scientific journals IEEE Transaction on Medical Imaging and IEEE Transactions on Geoscience and Remote Sensing were created by the Institute of Electrical and Electronics Engineers (IEEE) association in 1982 and 1980, respectively, to manage the engineering aspects for medical imaging and satellite data.
Remote sensing (RS) images are obtained using remote sensing technology such as airplanes and satellites under long-distance conditions. The detection of targets in RS images is very important in military applications, urban planning, resource exploration, agriculture, and other fields. CV techniques like content-based image retrieval techniques [1], semantic segmentation [2], scene classification [3], nonsupervised learning [4], and transfer learning [4], among others, have been applied to RS images. One specific application is cropland field identification, which is a key element of precision agriculture [5].
Common imaging techniques like X-ray radiographs, computed tomography (CT), and/or magnetic resonance imaging (MRI) have revolutionized the field of diagnostic medicine, providing non-destructive procedures for examining the interior of our bodies. Due to overlaps between anatomical structures, interpreting medical images is very challenging, even for experienced radiologists. The clinical interest in understanding these medical images explains the interest in developing computer algorithms that can aid experts in their clinical tasks. Across 40 years, the intersection of CV techniques and medical imaging has provided many clinical solutions. Common CV tasks like feature detection, recognition, segmentation, and three-dimensional modeling have been developed for processing different types of medical images and solving specific clinical problems. Some examples are chest radiograph analysis [6]; dental imaging (panoramic X-rays and other imaging modalities), to aid dental experts in diagnosing various dental disorders [7]; brain MRI modalities, for identifying distinct features that characterize autism spectrum disorder [8]; skin lesion analysis from RGB images to diagnosis skin cancer [9]; diagnosing glaucoma by analyzing retinal imaging data [10]; and detecting lung and colorectal cancer using CT imaging [11,12]. Recent advances in robotics now permit the acquisition of more medical images that can help clinicians make diagnoses or guide surgeons, in which the source and detector are positioned by robots with greater precision and accuracy [13,14]. Although X-ray imaging technology has been used in clinical tasks for decades, it has recently been extended to industrial production and security applications, where it can detect anomalies or defects inside products non-destructively and identify prohibited objects inside baggage without opening it [15].
Machine vision systems make use of different processing stages like image pre-processing, target image or video segmentation, feature extraction and selection, object recognition, classification, and 3D modeling, among others. These different types of tasks have typically involved different types of algorithms, with the classical CV techniques using explicitly programmed algorithms to solve specific tasks [16,17]. In recent years, deep learning (DL) models have yielded a new generation of CV methods [18] based on multi-layered neural networks such as convolutional neural networks and transformers, endowing computers with the ability to learn without them being explicitly programmed. The most popular architectures for computer vision are convolutional neural networks (CNNs) [19], which have become the standard DL-based approches for many recognition tasks due to their ability to learn high-level features in their convolutional layers; generative adversarial networks (GANs) [20], which learn from a given training set to generate new data; recurrent neural networks (RNNs), which have the capability to process temporal information and sequential data; different versions of YOLO (You Only Look Once) for object detection [21,22]; and transformers [23], which are primarily based on self-attention mechanisms [24]. These have all found applications in numerous fields, such as medicine [8,25,26,27,28], image generation [29], and remote sensing [2,4], among others. In conclusion, several algorithms have emerged over time, each with its own set of advantages and disadvantages. While DL models have good learnability, they often require a substantial number of real labels for training, provide poor interpretability due to their black-box structure, and require intensive computational resources or specific hardware.
As previously mentioned, the medical and remote sensing fields have used CV techniques for the automation of different tasks extensively. Nevertheless, recent advances in the image acquisition technology available, mainly due to research in optics and digital sensing, as well as increasing computer power, have unleashed new opportunities to apply CV techniques to new types of images, like microscopy imaging [30] or unmanned aerial vehicle (UAV) acquisitions [31]. UAVs are flying robots either remotely controlled by somebody or navigated autonomously using a computer system on board the vehicle or on the ground. They are able to acquire images in complex applications due to their small size, low cost, and high mobility. UAV systems enable the acquisition of real-time environmental data for developing CV applications such as vehicle detection [32] and digital precision agriculture [33,34,35], with the latter involving a variety of tasks, such as weed, crop pest, and disease detection, in order to apply the right practice at the right place, the right time, and the right quantity. Thus, UAVs are versatile, with the capacity for different kinds of sensors to be boarded onto them [36]. In capturing both the spatial and spectral features of an object’s surface, hyperspectral images are also used for agricultural tasks like disease detection, weed, and stress detection; crop monitoring; applying nutrients; soil mineralogy studies; yield estimation; and sorting applications [37,38].
Microscopy imaging has a prominent role in modern biology for the visualization of tissues, cells, proteins, and macromolecular structures at all resolutions. Indeed, biopsy diagnosis is the gold standard for cancer diagnosis in pathology. Machine vision has recently been employed in the biomedical field to detect, measure, and recognize cells and patterns in histopathology images or for target tracking and 3D reconstruction [28,39]. These biomedical applications can be grouped together on the basis of the tissue or organ analyzed—for example, renal pathology [40,41], computational cytology [42], breast cancer [43], oral cancer [44], and intestine pathology [45], among others. However, microscopy imaging has also found applications to pollen identification [46,47], microorganism recognition [48], and estimating the fecundity of fish based on histological images of their gonads [49,50].
CV techniques have also played an important role in the product life cycle across the entire industrial manufacturing process, including product design, modeling and simulation, planning and scheduling, the production process, inspection and quality control, assembly, transportation, and disassembly [51,52]. Equally, they have been applied to a myriad of domains: car parking lot management, detecting the positions of parking spaces [53] or used in autonomous driving [54]; the mushroom industry, for the identification of poisonous mushrooms, plucking cultivated mushrooms covered by the soil, and mechanized grading of mushrooms [55]; continuous monitoring of beehives [56] and beehive products such as honeybee pollen [57]; marine ecosystems, in monitoring fish habitats using underwater videos or images [58] or estimating the fecundity of fish from histological images of their gonads [50,59]; crop disease monitoring [35,60,61]; the identification of insects from digital images [62]; food quality assessments [63,64], covering potatoes [65], fruit damage [66], and dry-cured ham [67,68]; automation within the chicken farming industry [69]; and plant identification [70].
All of these computer vision applications involve the integration of the following elements:
  • Support for data recording: Microscopes; UAVs; satellites; robots; MRI, X-ray, and CT devices; and others.
  • Type of input data: 2D images, videos, or other information, dependent on the high-performance sensors used to perceive the given scenario, which could be RGB cameras; multispectral, hyperspectral, thermal, and infrared sensors; synthetic-aperture radar (SAR) cameras; Light Detection and Ranging or Laser Imaging Detection and Ranging (LiDAR) sensors; or other cameras [71].
  • Machine vision-related aim of the application: Feature detection or recognition, image segmentation, image classification, 3D modeling or reconstruction, object tracking, defect detection, object counting or measurements from images, and visual inspection, among others. The evaluation methodology used in CV techniques is dependent on the aims and application in question.
  • Type of processing: CV methods can be roughly divided into three categories: non-learning-based methods, learning-based methods, and hybrid methods. The first types of methods are usually known as the classical methods, and these rely on unsupervised, manually designed feature extractors or statistical models, in which the output is calculated from direct processing of the input data. Currently, the second types are methods based on deep learning, in which previous training with ground-truth data is needed to compute the output. Hybrid strategies normally combine the extraction of features from the input data with a subsequent machine learning stage.
  • Experimental testing: Using publicly available datasets or private data.
Despite the abundance of works and reviews published in this domain in recent years, many challenges are still open questions. From a computational point of view, future work should focus on designing more efficient algorithms that can operate in real time or run on low-capacity devices such as UAVs. As mentioned, some machine vision techniques require a substantial amount of ground-truth labeled data for training. Transfer learning or unsupervised annotation algorithms have been proposed to alleviate the need for labeled data, addressing domain shift or directly labeling the data, but there is still room for further research on this aspect. At the same time, a considerable number of the studies in the literature on CV only use private data or use public datasets with different experimental setups, complicating comparison between algorithms. So, new data must be made public in order for the field to mature.
For CV applications in which the decision-making involved affects people, there is substantial evidence that AI-based systems take on race-, ethnicity-, culture-, age-, and gender-based biases, among others, that disadvantage minority populations. Gender bias typically intersects with other biases [72], and Natural Language Processing (NLP) and facial analysis and recognition are research fields that feel greater effects of gender bias—for example, gender bias in commercial facial recognition systems [73,74] or the social impact of image generation models [75]. Machine vision systems perpetuated or intensified social inequalities in recent applications of developing systems that integrated NLP and CV [76,77], with biases introduced by both. Biases can be introduced into CV systems in many different ways: the ground-truth labeling of the data, the selection of the data included for training, the algorithm design, and evaluation of the prediction quality, among other design decisions. Gender biases can be imbued into CV systems unintentionally due to our cultural experience or gender stereotypes. Therefore, CV system developers should be aware of gender bias in their future work. Equally, some of the images and videos used in CV research are obtained without the explicit consent of the people photographed. Hence, recently, the IEEE announced that it will no longer allow the use of the Lena image in its publications. Furthermore, in some CV applications, such as emotional computing, people’s right to privacy and intimacy should be socially debated [78].

2. Overview of This Special Issue

This Special Issue called for scientific articles related to the computer vision applications previously covered, and after a double-blind review process, nineteen articles were published. This section provides a brief overview of each contribution in order to encourage further exploration on the part of the reader.
The first contribution, entitled “A UAV Aerial Image Target Detection Algorithm Based on YOLOv7 Improved Model”, proposes an enhanced YOLOv7 model for detecting small targets in UAV images. Experiments were carried out on the UAV aerial photo dataset VisDrone2019 and compared with the YOLOv7 model.
The second contribution, entitled “RN-YOLO: A Small Target Detection Model for Aerial Remote-Sensing Images”, applies a new YOLO model based on YOLOv8, called RN-YOLO, to detecting small targets in RS images. These experiments were conducted on the TGRS-HRRSD and RSOD datasets and compared with the YOLOv8 model.
The third contribution, entitled “Dense Object Detection Based on De-Homogenized Queries”, establishes a new method for dense object detection in images and videos. Experiments were run on the CrowdHuman dataset and compared with other state-of-the-art (SOTA) methods.
The fourth contribution, entitled “Multi-Scale Fusion Uncrewed Aerial Vehicle Detection Based on RT-DETR”, covers an enhanced model of a real-time detection transformer (RT-DETR), a real-time end-to-end object detection model for detecting drones in images. Two available UAV datasets were used for the experiments.
The fifth contribution, entitled “Efficient Vision Transformer YOLOv5 for Accurate and Fast Traffic Sign Detection”, details a new model for detecting traffic signs, which is a vital task in autonomous driving systems. It achieved faster and more accurate results than the YOLOv5 model. Experiments were conducted on the 3L-TT100K traffic sign dataset.
The sixth contribution, entitled “Facial Beauty Prediction Combined with Multi-Task Learning of Adaptive Sharing Policy and Attentional Feature Fusion”, presents a strategy for improving facial attractiveness assessments, involving experimental testing on the LSAFBD and SCUT-FBP5500 databases.
The seventh contribution, entitled “Two-Stage Progressive Learning for Vehicle Re-Identification in Variable Illumination Condition”, elucidates a TSPL framework for recognizing vehicles in images acquired by surveillance cameras with varying viewpoints, levels of illumination, and resolutions. A private large-scale dataset (VERI-DAN) and the Vehicle-1M dataset were used for the experiments, and the framework proposed was compared with other SOTA methods.
Inspired by the separation of luminance and chrominance information in the YCbCr color space, the eighth contribution, entitled “DBENet: Dual-Branch Brightness Enhancement Fusion Network for Low-Light Image Enhancement”, describes a new model for enhancing RGB images with low light, minimizing brightness, color distortion, and noise pollution in the enhanced images. The experiments in this paper made use of multiple publicly available low-light image datasets, and the results were evaluated against those of classical algorithms.
The ninth contribution, entitled “RSLC-Deeplab: A Ground Object Classification Method for High-Resolution Remote Sensing Images”, suggests a semantic segmentation network for accurately segmenting remote sensing images. Experiments conducted using the WHDLD dataset demonstrated its outperformance of the PSP-NET, U-NET, MACU-NET, and DeeplabV3+ networks.
The tenth contribution, entitled “YOLO-CID: Improved YOLOv7 for X-ray Contraband Image Detection”, augments the YOLOv7 method for contraband image detection in X-ray inspection systems in order to detect small objects under occlusion or low contrast. Its results on the PIDray public dataset were an improvement upon the results of the YOLOv7 algorithm.
The eleventh contribution, entitled “Enhancing the Accuracy of an Image Classification Model Using Cross-Modality Transfer Learning”, proposes a cross-modality transfer learning approach to shifting the knowledge when the source and target domains are different, specifically from the text domain to the image domain.
The twelfth contribution, entitled “Three-Dimensional Measurement of Full Profile of Steel Rail Cross-Section Based on Line-Structured Light”, solves the industrial problem of improving railway operation safety by proposing a method for three-dimensional measurement of the cross-sectional profiles of steel rails based on binocular line-structured light. Private data were used in this paper.
The thirteenth contribution, entitled “A Workpiece-Dense Scene Object Detection Method Based on Improved YOLOv5”, optimizes the YOLOv5 method for detecting workpieces in dense images of industrial production lines, using a self-built artifact dataset to compare the results with the original method.
The fourteenth contribution, entitled “Improving the Performance of the Single Shot Multibox Detector for Steel Surface Defects with Context Fusion and Feature Refinement”, devises a method for improving the ability to identify steel surface defects. Experiments were run on the public NEU-DET dataset and compared with other SOTA methods (Faster R-CNN, RetinaNet, and different YOLO methods).
The fifteenth contribution, entitled “Object Detection Algorithm of UAV Aerial Photography Image Based on Anchor-Free Algorithms”, constructs an algorithm for anchor-free target detection in UAV aerial photography images. In experiments performed on the VisDrone dataset, it outperformed the fully convolutional one-stage object detection algorithm.
The sixteenth contribution, entitled “A Vehicle Recognition Model Based on Improved YOLOv5”, in order to increase vehicle driving safety, validated the results of an improved YOLOv5s algorithm for vehicle identification and detection on a self-built dataset against the results of the YOLOv5 method.
The seventeenth contribution, entitled “Material-Aware Path Aggregation Network and Shape Decoupled SIoU for X-ray Contraband Detection”, outlines a method, based on variants of the YOLO model, for detecting and classifying contraband in X-ray baggage images. They evaluated the method proposed on the public X-ray contraband SIXray and OPIXray datasets and conducted a comparison of the results with those of other SOTA X-ray baggage inspection detection methods.
The eighteenth contribution, entitled “Fast Adaptive Binarization of QR Code Images for Automatic Sorting in Logistics Systems”, presents an adaptive binarization method for reading unevenly illuminated QR codes in automatic sorting in logistics systems. The image quality, recognition rate, and computation speed of the proposed method was tested against other SOTA methods on different examples.
The nineteenth contribution, entitled “Surveying Racial Bias in Facial Recognition: Balancing Datasets and Algorithmic Enhancements”, is a review on facial recognition systems that involve specific racial categories, discussing balanced facial recognition datasets, addressing and analyzing the racial bias of the methods, and exploring the interrelation of racial and gender bias.

Funding

This article received no external funding.

Acknowledgments

The Guest Editor of this Special Issue sincerely thanks all the scientists who submitted their research articles, the reviewers who assisted in evaluating these manuscripts, and both the Editorial Board Members and the Editors of Electronics for their overall support. Financial support from the Xunta de Galicia and the European Union (European Regional Development Fund—ERDF), Project ED431G-2019/04, is also acknowledged.

Conflicts of Interest

The author declares no conflicts of interest.

List of Contributions

References

  1. Zhou, W.; Guan, H.; Li, Z.; Shao, Z.; Delavar, M.R. Remote Sensing Image Retrieval in the Past Decade: Achievements, Challenges, and Future Directions. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 1447–1473. [Google Scholar] [CrossRef]
  2. Huang, L.; Jiang, B.; Lv, S.; Liu, Y.; Fu, Y. Deep-Learning-Based Semantic Segmentation of Remote Sensing Images: A Survey. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 8370–8396. [Google Scholar] [CrossRef]
  3. Tan, X.; Xi, B.; Li, J.; Zheng, T.; Li, Y.; Xue, C.; Chanussot, J. Review of Zero-Shot Remote Sensing Image Scene Classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 11274–11289. [Google Scholar] [CrossRef]
  4. Ma, Y.; Chen, S.; Ermon, S.; Lobell, D.B. Transfer learning in environmental remote sensing. Remote Sens. Environ. 2024, 301, 113924. [Google Scholar] [CrossRef]
  5. Xu, F.; Yao, X.; Zhang, K.; Yang, H.; Feng, Q.; Li, Y.; Yan, S.; Gao, B.; Li, S.; Yang, J.; et al. Deep learning in cropland field identification: A review. Comput. Electron. Agric. 2024, 222, 109042. [Google Scholar] [CrossRef]
  6. Van Ginneken, B.; Ter Haar Romeny, B.; Viergever, M. Computer-aided diagnosis in chest radiography: A survey. IEEE Trans. Med. Imaging 2001, 20, 1228–1241. [Google Scholar] [CrossRef]
  7. Priya, J.; Raja, S.K.S.; Kiruthika, S.U. State-of-art technologies, challenges, and emerging trends of computer vision in dental images. Comput. Biol. Med. 2024, 178, 108800. [Google Scholar] [CrossRef]
  8. Alharthi, A.G.; Alzahrani, S.M. Do it the transformer way: A comprehensive review of brain and vision transformers for autism spectrum disorder diagnosis and classification. Comput. Biol. Med. 2023, 167, 107667. [Google Scholar] [CrossRef]
  9. Hasan, M.K.; Ahamad, M.A.; Yap, C.H.; Yang, G. A survey, review, and future trends of skin lesion segmentation and classification. Comput. Biol. Med. 2023, 155, 106624. [Google Scholar] [CrossRef]
  10. Ashtari-Majlan, M.; Dehshibi, M.M.; Masip, D. Glaucoma diagnosis in the era of deep learning: A survey. Expert Syst. Appl. 2024, 256, 124888. [Google Scholar] [CrossRef]
  11. Porto-Álvarez, J.; Barnes, G.T.; Villanueva, A.; García-Figueiras, R.; Baleato-González, S.; Huelga Zapico, E.; Souto-Bayarri, M. Digital Medical X-ray Imaging, CAD in Lung Cancer and Radiomics in Colorectal Cancer: Past, Present and Future. Appl. Sci. 2023, 13, 2218. [Google Scholar] [CrossRef]
  12. González-Castro, V.; Cernadas, E.; Huelga, E.; Fernández-Delgado, M.; Porto, J.; Antunez, J.R.; Souto-Bayarri, M. CT Radiomics in Colorectal Cancer: Detection of KRAS Mutation Using Texture Analysis and Machine Learning. Appl. Sci. 2020, 10, 6214. [Google Scholar] [CrossRef]
  13. Salcudean, S.E.; Moradi, H.; Black, D.G.; Navab, N. Robot-Assisted Medical Imaging: A Review. Proc. IEEE 2022, 110, 951–967. [Google Scholar] [CrossRef]
  14. Schmidt, A.; Mohareri, O.; DiMaio, S.; Yip, M.C.; Salcudean, S.E. Tracking and mapping in medical computer vision: A review. Med. Image Anal. 2024, 94, 103131. [Google Scholar] [CrossRef] [PubMed]
  15. Rafiei, M.; Raitoharju, J.; Iosifidis, A. Computer Vision on X-Ray Data in Industrial Production and Security Applications: A Comprehensive Survey. IEEE Access 2023, 11, 2445–2477. [Google Scholar] [CrossRef]
  16. González, R.C.; Woods, R.E.R.E. Digital Image Processing, 4th ed.; Pearson: New York, NY, USA, 2018. [Google Scholar]
  17. Sonka, M.; Hlavac, V.; Boyle, R. Image Processing, Analysis and Machine Vision, 3rd ed.; PWS: Pacific Grove, CA, USA, 2008. [Google Scholar]
  18. Hassaballah, M.; Awad, A.I. Deep Learning in Computer Vision: Principles and Applications, First Issued in Paperback 2021 ed.; Digital Imaging and Computer Vision Series; CRC Press, Taylor & Francis Group: Boca Raton, FL, USA, 2021. [Google Scholar]
  19. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 444, 436–444. [Google Scholar] [CrossRef]
  20. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:stat.ML/1406.2661. [Google Scholar] [CrossRef]
  21. Zou, Z.; Chen, K.; Shi, Z.; Guo, Y.; Ye, J. Object Detection in 20 Years: A Survey. Proc. IEEE 2023, 111, 257–276. [Google Scholar] [CrossRef]
  22. Chaudhari, S.; Malkan, N.; Momin, A.; Bonde, M. Yolo Real Time Object Detection. Int. J. Comput. Trends Technol. 2020, 68, 70–76. [Google Scholar] [CrossRef]
  23. Han, K.; Wang, Y.; Chen, H.; Chen, X.; Guo, J.; Liu, Z.; Tang, Y.; Xiao, A.; Xu, C.; Xu, Y.; et al. A Survey on Vision Transformer. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 87–110. [Google Scholar] [CrossRef]
  24. Gonçalves, T.; Rio-Torto, I.; Teixeira, L.F.; Cardoso, J.S. A Survey on Attention Mechanisms for Medical Applications: Are we Moving Toward Better Algorithms? IEEE Access 2022, 10, 98909–98935. [Google Scholar] [CrossRef]
  25. Papanastasiou, G.; Dikaios, N.; Huang, J.; Wang, C.; Yang, G. Is Attention all You Need in Medical Image Analysis? A Review. IEEE J. Biomed. Health Inform. 2024, 28, 1398–1411. [Google Scholar] [CrossRef] [PubMed]
  26. Azad, R.; Kazerouni, A.; Heidari, M.; Aghdam, E.K.; Molaei, A.; Jia, Y.; Jose, A.; Roy, R.; Merhof, D. Advances in medical image analysis with vision Transformers: A comprehensive review. Med. Image Anal. 2024, 91, 103000. [Google Scholar] [CrossRef] [PubMed]
  27. Shamshad, F.; Khan, S.; Zamir, S.W.; Khan, M.H.; Hayat, M.; Khan, F.S.; Fu, H. Transformers in medical imaging: A survey. Med. Image Anal. 2023, 88, 102802. [Google Scholar] [CrossRef] [PubMed]
  28. Hu, W.; Li, X.; Li, C.; Li, R.; Jiang, T.; Sun, H.; Huang, X.; Grzegorzek, M.; Li, X. A state-of-the-art survey of artificial neural networks for Whole-slide Image analysis: From popular Convolutional Neural Networks to potential visual transformers. Comput. Biol. Med. 2023, 161, 107034. [Google Scholar] [CrossRef] [PubMed]
  29. Dubey, S.R.; Singh, S.K. Transformer-based Generative Adversarial Networks in Computer Vision: A Comprehensive Survey. IEEE Trans. Artif. Intell. 2024, 1–16. [Google Scholar] [CrossRef]
  30. Kervrann, C.; Acton, S.T.; Olivo-Marin, J.C.; Sorzano, C.; Unser, M. Introduction to the Issue on Advanced Signal Processing in Microscopy and Cell Imaging. IEEE J. Sel. Top. Signal Process. 2016, 10, 3–5. [Google Scholar] [CrossRef]
  31. Han, Y.; Liu, H.; Wang, Y.; Liu, C. A Comprehensive Review for Typical Applications Based Upon Unmanned Aerial Vehicle Platform. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 9654–9666. [Google Scholar] [CrossRef]
  32. Bouguettaya, A.; Zarzour, H.; Kechida, A.; Taberkit, A.M. Vehicle Detection From UAV Imagery With Deep Learning: A Review. IEEE Trans. Neural Networks Learn. Syst. 2022, 33, 6047–6067. [Google Scholar] [CrossRef]
  33. Phang, S.K.; Chiang, T.H.A.; Happonen, A.; Chang, M.M.L. From Satellite to UAV-Based Remote Sensing: A Review on Precision Agriculture. IEEE Access 2023, 11, 127057–127076. [Google Scholar] [CrossRef]
  34. Toscano, F.; Fiorentino, C.; Capece, N.; Erra, U.; Travascia, D.; Scopa, A.; Drosos, M.; D’Antonio, P. Unmanned Aerial Vehicle for Precision Agriculture: A Review. IEEE Access 2024, 12, 69188–69205. [Google Scholar] [CrossRef]
  35. Joshi, P.; Sandhu, K.S.; Singh Dhillon, G.; Chen, J.; Bohara, K. Detection and monitoring wheat diseases using unmanned aerial vehicles (UAVs). Comput. Electron. Agric. 2024, 224, 109158. [Google Scholar] [CrossRef]
  36. López, Y.A.; García-Fernández, M.; Álvarez-Narciandi, G.; Andrés, F.L.H. Unmanned Aerial Vehicle-Based Ground-Penetrating Radar Systems: A review. IEEE Geosci. Remote Sens. Mag. 2022, 10, 66–86. [Google Scholar] [CrossRef]
  37. Ram, B.G.; Oduor, P.; Igathinathane, C.; Howatt, K.; Sun, X. A systematic review of hyperspectral imaging in precision agriculture: Analysis of its current state and future prospects. Comput. Electron. Agric. 2024, 222, 109037. [Google Scholar] [CrossRef]
  38. Shuai, L.; Li, Z.; Chen, Z.; Luo, D.; Mu, J. A research review on deep learning combined with hyperspectral Imaging in multiscale agricultural sensing. Comput. Electron. Agric. 2024, 217, 108577. [Google Scholar] [CrossRef]
  39. He, W.; Liu, T.; Han, Y.; Ming, W.; Du, J.; Liu, Y.; Yang, Y.; Wang, L.; Jiang, Z.; Wang, Y.; et al. A review: The detection of cancer cells in histopathology based on machine vision. Comput. Biol. Med. 2022, 146, 105636. [Google Scholar] [CrossRef]
  40. Cordido, A.; Cernadas, E.; Fernández-Delgado, M.; García-González, M.A. CystAnalyser: A new software tool for the automatic detection and quantification of cysts in Polycystic Kidney and Liver Disease, and other cystic disorders. PLoS Comput. Biol. 2020, 16, e1008337. [Google Scholar] [CrossRef]
  41. Deng, R.; Yang, H.; Jha, A.; Lu, Y.; Chu, P.; Fogo, A.B.; Huo, Y. Map3D: Registration-Based Multi-Object Tracking on 3D Serial Whole Slide Images. IEEE Trans. Med. Imaging 2021, 40, 1924–1933. [Google Scholar] [CrossRef]
  42. Jiang, H.; Zhou, Y.; Lin, Y.; Chan, R.C.; Liu, J.; Chen, H. Deep learning for computational cytology: A survey. Med. Image Anal. 2023, 84, 102691. [Google Scholar] [CrossRef]
  43. Rodríguez-Candela Mateos, M.; Azmat, M.; Santiago-Freijanes, P.; Galán-Moya, E.M.; Fernández-Delgado, M.; Aponte, R.B.; Mosquera, J.; Acea, B.; Cernadas, E.; Mayán, M.D. Software BreastAnalyser for the semi-automatic analysis of breast cancer immunohistochemical images. Sci. Rep. 2024, 14, 2995. [Google Scholar] [CrossRef]
  44. Al-Tarawneh, Z.A.; Pena-Cristóbal, M.; Cernadas, E.; Suarez-Peñaranda, J.M.; Fernández-Delgado, M.; Mbaidin, A.; Gallas-Torreira, M.; Gándara-Vila, P. OralImmunoAnalyser: A software tool for immunohistochemical assessment of oral leukoplakia using image segmentation and classification models. Front. Artif. Intell. 2024, 7, 1324410. [Google Scholar] [CrossRef] [PubMed]
  45. Jing, Y.; Li, C.; Du, T.; Jiang, T.; Sun, H.; Yang, J.; Shi, L.; Gao, M.; Grzegorzek, M.; Li, X. A comprehensive survey of intestine histopathological image analysis using machine vision approaches. Comput. Biol. Med. 2023, 165, 107388. [Google Scholar] [CrossRef] [PubMed]
  46. Rodríguez-Damián, M.; Cernadas, E.; Formella, A.; Fernández-Delgado, M.; Sa-Otero, P.D. Automatic detection and classification of grains of pollen based on shape and texture. IEEE Trans. Syst. Man Cybern. Part C 2006, 36, 531–542. [Google Scholar] [CrossRef]
  47. Li, J.; Cheng, W.; Xu, X.; Zhao, L.; Liu, S.; Gao, Z.; Ye, C.; You, H. How to identify pollen like a palynologist: A prior knowledge-guided deep feature learning for real-world pollen classification. Expert Syst. Appl. 2024, 237, 121392. [Google Scholar] [CrossRef]
  48. Kulwa, F.; Li, C.; Zhao, X.; Cai, B.; Xu, N.; Qi, S.; Chen, S.; Teng, Y. A State-of-the-Art Survey for Microorganism Image Segmentation Methods and Future Potential. IEEE Access 2019, 7, 100243–100269. [Google Scholar] [CrossRef]
  49. Mbaidin, A.; Cernadas, E.; Al-Tarawneh, Z.A.; Fernández-Delgado, M.; Domínguez-Petit, R.; Rábade-Uberos, S.; Hassanat, A. MSCF: Multi-Scale Canny Filter to Recognize Cells in Microscopic Images. Sustainability 2023, 15, 13693. [Google Scholar] [CrossRef]
  50. Mbaidin, A.; Rábade-Uberos, S.; Dominguez-Petit, R.; Villaverde, A.; Gónzalez-Rufino, M.E.; Formella, A.; Fernández-Delgado, M.; Cernadas, E. STERapp: Semiautomatic Software for Stereological Analysis. Application in the Estimation of Fish Fecundity. Electronics 2021, 10, 1432. [Google Scholar] [CrossRef]
  51. Zhou, L.; Zhang, L.; Konz, N. Computer Vision Techniques in Manufacturing. IEEE Trans. Syst. Man, Cybern. Syst. 2023, 53, 105–117. [Google Scholar] [CrossRef]
  52. Ahmad, H.M.; Rahimi, A. Deep learning methods for object detection in smart manufacturing: A survey. J. Manuf. Syst. 2022, 64, 181–196. [Google Scholar] [CrossRef]
  53. de Almeida, P.R.L.; Alves, J.H.; Parpinelli, R.S.; Barddal, J.P. A systematic review on computer vision-based parking lot management applied on public datasets. Expert Syst. Appl. 2022, 198, 116731. [Google Scholar] [CrossRef]
  54. Zhao, J.; Zhao, W.; Deng, B.; Wang, Z.; Zhang, F.; Zheng, W.; Cao, W.; Nan, J.; Lian, Y.; Burke, A.F. Autonomous driving system: A comprehensive survey. Expert Syst. Appl. 2024, 242, 122836. [Google Scholar] [CrossRef]
  55. Yin, H.; Yi, W.; Hu, D. Computer vision and machine learning applied in the mushroom industry: A critical review. Comput. Electron. Agric. 2022, 198, 107015. [Google Scholar] [CrossRef]
  56. Bilik, S.; Zemcik, T.; Kratochvila, L.; Ricanek, D.; Richter, M.; Zambanini, S.; Horak, K. Machine learning and computer vision techniques in continuous beehive monitoring applications: A survey. Comput. Electron. Agric. 2024, 217, 108560. [Google Scholar] [CrossRef]
  57. Carrión, P.; Cernadas, E.; Gálvez, J.F.; Damián, M.; de Sá-Otero, P. Classification of honeybee pollen using a multiscale texture filtering scheme. Mach. Vis. Appl. 2024, 15, 186–193. [Google Scholar] [CrossRef]
  58. Saleh, A.; Sheaves, M.; Jerry, D.; Rahimi Azghadi, M. Applications of deep learning in fish habitat monitoring: A tutorial and survey. Expert Syst. Appl. 2024, 238, 121841. [Google Scholar] [CrossRef]
  59. González-Rufino, E.; Carrión, P.; Cernadas, E.; Fernández-Delgado, M.; Domínguez-Petit, R. Exhaustive comparison of colour texture features and classification methods to discriminate cells categories in histological images of fish ovary. Pattern Recognit. 2013, 46, 2391–2407. [Google Scholar] [CrossRef]
  60. Ariza-Sentís, M.; Vélez, S.; Martínez-Peña, R.; Baja, H.; Valente, J. Object detection and tracking in Precision Farming: A systematic review. Comput. Electron. Agric. 2024, 219, 108757. [Google Scholar] [CrossRef]
  61. Kumar, D.; Kukreja, V. Image segmentation, classification, and recognition methods for wheat diseases: Two Decades’ systematic literature review. Comput. Electron. Agric. 2024, 221, 109005. [Google Scholar] [CrossRef]
  62. De Cesaro Júnior, T.; Rieder, R. Automatic identification of insects from digital images: A survey. Comput. Electron. Agric. 2020, 178, 105784. [Google Scholar] [CrossRef]
  63. Meenu, M.; Kurade, C.; Neelapu, B.C.; Kalra, S.; Ramaswamy, H.S.; Yu, Y. A concise review on food quality assessment using digital image processing. Trends Food Sci. Technol. 2021, 118, 106–124. [Google Scholar] [CrossRef]
  64. Pu, H.; Yu, J.; Sun, D.W.; Wei, Q.; Wang, Z. Feature construction methods for processing and analysing spectral images and their applications in food quality inspection. Trends Food Sci. Technol. 2023, 138, 726–737. [Google Scholar] [CrossRef]
  65. Sanchez, P.D.C.; Hashim, N.; Shamsudin, R.; Mohd Nor, M.Z. Applications of imaging and spectroscopy techniques for non-destructive quality evaluation of potatoes and sweet potatoes: A review. Trends Food Sci. Technol. 2020, 96, 208–221. [Google Scholar] [CrossRef]
  66. Mahanti, N.K.; Pandiselvam, R.; Kothakota, A.; Ishwarya S., P.; Chakraborty, S.K.; Kumar, M.; Cozzolino, D. Emerging non-destructive imaging techniques for fruit damage detection: Image processing and analysis. Trends Food Sci. Technol. 2022, 120, 418–438. [Google Scholar] [CrossRef]
  67. Ávila, M.; Durán, M.; Antequera, T.; Caballero, D.; Palacios-Pérez, T.; Cernadas, E.; Fernández-Delgado, M. Magnetic Resonance Imaging, texture analysis and regression techniques to non-destructively predict the quality characteristics of meat pieces. Eng. Appl. Artif. Intell. 2019, 82, 110–125. [Google Scholar] [CrossRef]
  68. Cernadas, E.; Fernández-Delgado, M.; Fulladosa, E.; Muñoz, I. Automatic marbling prediction of sliced dry-cured ham using image segmentation, texture analysis and regression. Expert Syst. Appl. 2022, 206, 117765. [Google Scholar] [CrossRef]
  69. Yang, D.; Cui, D.; Ying, Y. Development and trends of chicken farming robots in chicken farming tasks: A review. Comput. Electron. Agric. 2024, 221, 108916. [Google Scholar] [CrossRef]
  70. Akhtar, M.S.; Zafar, Z.; Nawaz, R.; Fraz, M.M. Unlocking plant secrets: A systematic review of 3D imaging in plant phenotyping techniques. Comput. Electron. Agric. 2024, 222, 109033. [Google Scholar] [CrossRef]
  71. Gallego, G.; Delbrück, T.; Orchard, G.; Bartolozzi, C.; Taba, B.; Censi, A.; Leutenegger, S.; Davison, A.J.; Conradt, J.; Daniilidis, K.; et al. Event-Based Vision: A Survey. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 154–180. [Google Scholar] [CrossRef]
  72. Shrestha, S.; Das, S. Exploring gender biases in ML and AI academic research through systematic literature review. Front. Artif. Intel. 2022, 5, 976838. [Google Scholar] [CrossRef]
  73. Schwemmer, C.; Knight, C.; Bello-Pardo, E.D.; Oklobdzija, S.; Schoonvelde, M.; Lockhart, J.W. Diagnosing Gender Bias in Image Recognition Systems. Socius 2020, 6, 2378023120967171. [Google Scholar] [CrossRef] [PubMed]
  74. Khalil, A.; Ahmed, S.G.; Khattak, A.M.; Al-Qirim, N. Investigating Bias in Facial Analysis Systems: A Systematic Review. IEEE Access 2020, 8, 130751–130761. [Google Scholar] [CrossRef]
  75. Katirai, A.; Garcia, N.; Ide, K.; Nakashima, Y.; Kishimoto, A. Situating the social issues of image generation models in the model life cycle: A sociotechnical approach. AI Ethics 2024. [Google Scholar] [CrossRef]
  76. Reale-Nosei, G.; Amador-Domínguez, E.; Serrano, E. From vision to text: A comprehensive review of natural image captioning in medical diagnosis and radiology report generation. Med. Image Anal. 2024, 97, 103264. [Google Scholar] [CrossRef] [PubMed]
  77. Nam, W.; Jang, B. A survey on multimodal bidirectional machine learning translation of image and natural language processing. Expert Syst. Appl. 2024, 235, 121168. [Google Scholar] [CrossRef]
  78. Crawford, K. The Atlas of AI; Yale University Press: New Haven, CT, USA, 2021. [Google Scholar]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Cernadas, E. Applications of Computer Vision, 2nd Edition. Electronics 2024, 13, 3779. https://doi.org/10.3390/electronics13183779

AMA Style

Cernadas E. Applications of Computer Vision, 2nd Edition. Electronics. 2024; 13(18):3779. https://doi.org/10.3390/electronics13183779

Chicago/Turabian Style

Cernadas, Eva. 2024. "Applications of Computer Vision, 2nd Edition" Electronics 13, no. 18: 3779. https://doi.org/10.3390/electronics13183779

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop