Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (103)

Search Parameters:
Keywords = CLAHE

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
22 pages, 44861 KiB  
Article
Multi-Scale Fusion Lightweight Target Detection Method for Coal and Gangue Based on EMBS-YOLOv8s
by Lin Gao, Pengwei Yu, Hongjuan Dong and Wenjie Wang
Sensors 2025, 25(6), 1734; https://doi.org/10.3390/s25061734 - 11 Mar 2025
Viewed by 178
Abstract
The accurate detection of coal gangue is an important prerequisite for the intelligent sorting of coal gangue. Aiming at existing coal gangue detection methods, which have problems such as low detection accuracy and complex model structure, a multi-scale fusion lightweight coal gangue target [...] Read more.
The accurate detection of coal gangue is an important prerequisite for the intelligent sorting of coal gangue. Aiming at existing coal gangue detection methods, which have problems such as low detection accuracy and complex model structure, a multi-scale fusion lightweight coal gangue target detection method based on the EMBS-YOLOv8s model is proposed. Firstly, the coal gangue images collected through the visual dark box platform are preprocessed using CLAHE to improve the contrast and clarity of the images. Secondly, the PAN-FAN structure is replaced by the EMBSFPN structure in the neck network. This structure can fully utilize the features of different scales, improve the model’s detection accuracy, and reduce its complexity. Finally, the CIoU loss function is replaced by the Wise-SIoU loss function at the prediction end. This improves the model’s convergence and stability and solves the problem of the imbalance of hard and easy samples in the dataset. The experimental results show that the mean average precision of the EMBS-YOLOv8s model on the self-constructed coal gangue dataset reaches 96.0%, which is 2.1% higher than that of the original YOLOv8s model. The Params, FLOPs, and Size of the model are also reduced by 29.59%, 12.68%, and 28.44%, respectively, relative to those of the original YOLOv8s model. Meanwhile, the detection speed of the EMBS-YOLOv8s model is 93.28 f.s−1, which has certain real-time detection performance. Compared with other YOLO series models, the EMBS-YOLOv8s model can effectively avoid the occurrence of false detection and missed detection phenomena in complex scenes such as low illumination, high noise, and motion blur. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Figure 1

20 pages, 21510 KiB  
Article
Visual Localization Method for Fastener-Nut Disassembly and Assembly Robot Based on Improved Canny and HOG-SED
by Xiangang Cao, Mengzhen Zuo, Guoyin Chen, Xudong Wu, Peng Wang and Yizhe Liu
Appl. Sci. 2025, 15(3), 1645; https://doi.org/10.3390/app15031645 - 6 Feb 2025
Viewed by 657
Abstract
Visual positioning accuracy is crucial for ensuring the successful execution of nut disassembly and assembly tasks by a fastener-nut disassembly and assembly robot. However, disturbances such as on-site lighting changes, abnormal surface conditions of nuts, and complex backgrounds formed by ballast in complex [...] Read more.
Visual positioning accuracy is crucial for ensuring the successful execution of nut disassembly and assembly tasks by a fastener-nut disassembly and assembly robot. However, disturbances such as on-site lighting changes, abnormal surface conditions of nuts, and complex backgrounds formed by ballast in complex railway environments can lead to poor visual positioning accuracy of the fastener nuts, thereby affecting the success rate of the robot’s continuous disassembly and assembly operations. Additionally, the existing method of detecting fasteners first and then positioning nuts has poor applicability in the field. A direct positioning algorithm for spiral rail spikes that combines an improved Canny algorithm with shape feature similarity determination is proposed in response to these issues. Firstly, CLAHE enhances the image, reducing the impact of varying lighting conditions in outdoor work environments on image details. Then, to address the difficulties in extracting the edges of rail spikes caused by abnormal conditions such as water stains, rust, and oil stains on the nuts themselves, the Canny algorithm is improved through three stages, filtering optimization, gradient boosting, and adaptive thresholding, to reduce the impact of edge loss on subsequent rail spike positioning results. Finally, considering the issue of false fitting due to background interference, such as ballast in gradient Hough transformations, the differences in texture and shape features between the rail spike and interference areas are analyzed. The HOG is used to describe the shape features of the area to be screened, and the similarity between the screened area and the standard rail spike template features is compared based on the standard Euclidean distance to determine the rail spike area. Spiral rail spikes are discriminated based on shape features, and the center coordinates of the rail spike are obtained. Experiments were conducted using images collected from the field, and the results showed that the proposed algorithm, when faced with complex environments with multiple interferences, has a correct detection rate higher than 98% and a positioning error mean of 0.9 mm. It exhibits excellent interference resistance and meets the visual positioning accuracy requirements for robot nut disassembly and assembly operations in actual working environments. Full article
(This article belongs to the Section Applied Industrial Technologies)
Show Figures

Figure 1

22 pages, 4113 KiB  
Article
Dual-Modal Approach for Ship Detection: Fusing Synthetic Aperture Radar and Optical Satellite Imagery
by Mahmoud Ahmed, Naser El-Sheimy and Henry Leung
Sensors 2025, 25(2), 329; https://doi.org/10.3390/s25020329 - 8 Jan 2025
Viewed by 719
Abstract
The fusion of synthetic aperture radar (SAR) and optical satellite imagery poses significant challenges for ship detection due to the distinct characteristics and noise profiles of each modality. Optical imagery provides high-resolution information but struggles in adverse weather and low-light conditions, reducing its [...] Read more.
The fusion of synthetic aperture radar (SAR) and optical satellite imagery poses significant challenges for ship detection due to the distinct characteristics and noise profiles of each modality. Optical imagery provides high-resolution information but struggles in adverse weather and low-light conditions, reducing its reliability for maritime applications. In contrast, SAR imagery excels in these scenarios but is prone to noise and clutter, complicating vessel detection. Existing research on SAR and optical image fusion often fails to effectively leverage their complementary strengths, resulting in suboptimal detection outcomes. This research presents a novel fusion framework designed to enhance ship detection by integrating SAR and optical imagery. This framework incorporates a detection system for optical images that utilizes Contrast Limited Adaptive Histogram Equalization (CLAHE) in combination with the YOLOv7 model to improve accuracy and processing speed. For SAR images, a customized Detection Transformer model, SAR-EDT, integrates advanced denoising algorithms and optimized pooling configurations. A fusion module evaluates the overlaps of detected bounding boxes based on intersection over union (IoU) metrics. Fused detections are generated by averaging confidence scores and recalculating bounding box dimensions, followed by robust postprocessing to eliminate duplicates. The proposed framework significantly improves ship detection accuracy across various scenarios. Full article
(This article belongs to the Special Issue Multi-Modal Image Processing Methods, Systems, and Applications)
Show Figures

Figure 1

14 pages, 5970 KiB  
Article
Universal Image Vaccine Against Steganography
by Shiyu Wei, Zichi Wang and Xinpeng Zhang
Symmetry 2025, 17(1), 66; https://doi.org/10.3390/sym17010066 - 2 Jan 2025
Viewed by 610
Abstract
In the past decade, the diversification of steganographic techniques has posed significant threats to information security, necessitating effective countermeasures. Current defenses, mainly reliant on steganalysis, struggle with detection accuracy. While “image vaccines” have been proposed, they often target specific methodologies. This paper introduces [...] Read more.
In the past decade, the diversification of steganographic techniques has posed significant threats to information security, necessitating effective countermeasures. Current defenses, mainly reliant on steganalysis, struggle with detection accuracy. While “image vaccines” have been proposed, they often target specific methodologies. This paper introduces a universal steganographic vaccine to enhance steganalysis accuracy. Our symmetric approach integrates with existing methods to protect images before online dissemination using the Contrast Limited Adaptive Histogram Equalization (CLAHE) algorithm. Experimental results show significant accuracy improvements across traditional and deep learning-based steganalysis, especially at medium-to-high payloads. Specifically, for payloads of 0.1–0.5 bpp, the original detection error rate was reduced from 0.3429 to 0.2346, achieving an overall average reduction of 31.57% for traditional algorithms, while the detection success rate of deep learning-based algorithms can reach 100%. Overall, integrating CLAHE as a universal vaccine significantly advances steganalysis. Full article
Show Figures

Figure 1

19 pages, 3563 KiB  
Article
Impact of Histogram Equalization on the Classification of Retina Lesions from OCT B-Scans
by Tomasz Marciniak and Agnieszka Stankiewicz
Electronics 2024, 13(24), 4996; https://doi.org/10.3390/electronics13244996 - 19 Dec 2024
Viewed by 566
Abstract
Deep learning solutions can be used to classify pathological changes of the human retina visualized in OCT images. Available datasets that can be used to train neural network models include OCT images (B-scans) of classes with selected pathological changes and images of the [...] Read more.
Deep learning solutions can be used to classify pathological changes of the human retina visualized in OCT images. Available datasets that can be used to train neural network models include OCT images (B-scans) of classes with selected pathological changes and images of the healthy retina. These images often require correction due to improper acquisition or intensity variations related to the type of OCT device. This article provides a detailed assessment of the impact of preprocessing on classification efficiency. The histograms of OCT images were examined and, depending on the histogram distribution, incorrect image fragments were removed. At the same time, the impact of histogram equalization using the standard method and the Contrast-Limited Adaptive Histogram Equalization (CLAHE) method was analyzed. The most extensive dataset of Labeled Optical Coherence Tomography (LOCT) images was used for the experimental studies. The impact of changes was assessed for different neural network architectures and various learning parameters, assuming classes of equal size. Comprehensive studies have shown that removing unnecessary white parts from the input image combined with CLAHE improves classification accuracy up to as much as 4.75% depending on the used network architecture and optimizer type. Full article
(This article belongs to the Special Issue Biometrics and Pattern Recognition)
Show Figures

Figure 1

24 pages, 6411 KiB  
Article
CNN-Based Kidney Segmentation Using a Modified CLAHE Algorithm
by Abror Shavkatovich Buriboev, Ahmadjon Khashimov, Akmal Abduvaitov and Heung Seok Jeon
Sensors 2024, 24(23), 7703; https://doi.org/10.3390/s24237703 - 2 Dec 2024
Cited by 1 | Viewed by 978
Abstract
This paper presents an enhanced approach to kidney segmentation using a modified CLAHE preprocessing method, aimed at improving image clarity and CNN performance on the KiTS19 dataset. To assess the impact of the modified CLAHE method, we conducted quality evaluations using the BRISQUE [...] Read more.
This paper presents an enhanced approach to kidney segmentation using a modified CLAHE preprocessing method, aimed at improving image clarity and CNN performance on the KiTS19 dataset. To assess the impact of the modified CLAHE method, we conducted quality evaluations using the BRISQUE metric, comparing the original, standard CLAHE and modified CLAHE versions of the dataset. The BRISQUE score decreased from 28.8 in the original dataset to 21.1 with the modified CLAHE method, indicating a significant improvement in image quality. Furthermore, CNN segmentation accuracy rose from 0.951 with the original dataset to 0.996 with the modified CLAHE method, outperforming the accuracy achieved with standard CLAHE preprocessing (0.969). These results highlight the benefits of the modified CLAHE method in refining image quality and enhancing segmentation performance. This study highlights the value of adaptive preprocessing in medical imaging workflows and shows that CNN-based kidney segmentation accuracy may be greatly increased by altering conventional CLAHE. Our method provides insightful information on optimizing preprocessing for medical imaging applications, leading to more accurate and dependable segmentation results for better clinical diagnosis. Full article
(This article belongs to the Special Issue Machine and Deep Learning in Sensing and Imaging)
Show Figures

Figure 1

9 pages, 3908 KiB  
Proceeding Paper
Automated Glaucoma Detection in Fundus Images Using Comprehensive Feature Extraction and Advanced Classification Techniques
by Vijaya Kumar Velpula, Jyothisri Vadlamudi, Purna Prakash Kasaraneni and Yellapragada Venkata Pavan Kumar
Eng. Proc. 2024, 82(1), 33; https://doi.org/10.3390/ecsa-11-20437 - 25 Nov 2024
Viewed by 345
Abstract
Glaucoma, a primary cause of irreversible blindness, necessitates early detection to prevent significant vision loss. In the literature, fundus imaging is identified as a key tool in diagnosing glaucoma, which captures detailed retina images. However, the manual analysis of these images can be [...] Read more.
Glaucoma, a primary cause of irreversible blindness, necessitates early detection to prevent significant vision loss. In the literature, fundus imaging is identified as a key tool in diagnosing glaucoma, which captures detailed retina images. However, the manual analysis of these images can be time-consuming and subjective. Thus, this paper presents an automated system for glaucoma detection using fundus images, combining diverse feature extraction methods with advanced classifiers, specifically Support Vector Machine (SVM) and AdaBoost. The pre-processing step incorporated image enhancement via Contrast-Limited Adaptive Histogram Equalization (CLAHE) to enhance image quality and feature extraction. This work investigated individual features such as the histogram of oriented gradients (HOG), local binary patterns (LBP), chip histogram features, and the gray-level co-occurrence matrix (GLCM), as well as their various combinations, including HOG + LBP + chip histogram + GLCM, HOG + LBP + chip histogram, and others. These features were utilized with SVM and Adaboost classifiers to improve classification performance. For validation, the ACRIMA dataset, a public fundus image collection comprising 369 glaucoma-affected and 309 normal images, was used in this work, with 80% of the data allocated for training and 20% for testing. The results of the proposed study show that different feature sets yielded varying accuracies with the SVM and Adaboost classifiers. For instance, the combination of LBP + chip histogram achieved the highest accuracy of 99.29% with Adaboost, while the same combination yielded a 65.25% accuracy with SVM. The individual feature LBP alone achieved 97.87% with Adaboost and 98.58% with SVM. Furthermore, the combination of GLCM + LBP provided a 98.58% accuracy with Adaboost and 97.87% with SVM. The results demonstrate that CLAHE and combined feature sets significantly enhance detection accuracy, providing a reliable tool for early and precise glaucoma diagnosis, thus facilitating timely intervention and improved patient outcomes. Full article
Show Figures

Figure 1

19 pages, 3906 KiB  
Article
Adaptive Enhancement of Thermal Infrared Images for High-Voltage Cable Buffer Layer Ablation
by Hao Zhan, Jing Zhang, Yuhao Lan, Fan Zhang, Qinqing Huang, Kai Zhou and Chengde Wan
Processes 2024, 12(11), 2543; https://doi.org/10.3390/pr12112543 - 14 Nov 2024
Cited by 1 | Viewed by 894
Abstract
In recent years, ablation of the buffer layer in high-voltage cables has become a prevalent issue compromising the reliability of power transmission systems. Given the internal location of these faults, direct monitoring and assessment are challenging, resulting in numerous undetected ablation hazards. Previous [...] Read more.
In recent years, ablation of the buffer layer in high-voltage cables has become a prevalent issue compromising the reliability of power transmission systems. Given the internal location of these faults, direct monitoring and assessment are challenging, resulting in numerous undetected ablation hazards. Previous practice has demonstrated that detecting buffer layer ablation through surface temperature distribution changes is feasible, offering a convenient, efficient, and non-destructive approach. However, the variability in heat generation and the subtle temperature differences in thermal infrared images, compounded by noise interference, can impair the accuracy and timeliness of fault detection. To overcome these challenges, this paper introduces an adaptive enhancement method for the thermal infrared imaging of high-voltage cable buffer layer ablation. The method involves an Average Gradient Weighted Guided Filtering (AGWGF) technique to decompose the image into background and detail layers, preventing noise amplification during enhancement. The background layer, containing the primary information, is enhanced using an improved Contrast Limited Adaptive Histogram Equalization (CLAHE) to accentuate temperature differences. The detail layer, rich in high-frequency content, undergoes improved Adaptive Bilateral Filtering (ABF) for noise reduction. The enhanced background and detail layers are then fused and stretched to produce the final enhanced thermal image. To vividly depict temperature variations in the buffer layer, pseudo-color processing is applied to generate color-infrared thermal images. The results indicate that the proposed method’s enhanced images and pseudo-colored infrared thermal images provide a clearer and more intuitive representation of temperature differences compared to the original images, with an average increase of 2.17 in information entropy and 8.38 in average gradient. This enhancement facilitates the detection and assessment of buffer layer ablation faults, enabling the prompt identification of faults. Full article
Show Figures

Figure 1

16 pages, 4589 KiB  
Article
Comparison of Preprocessing Method Impact on the Detection of Soldering Splashes Using Different YOLOv8 Versions
by Peter Klco, Dusan Koniar, Libor Hargas and Marek Paskala
Computation 2024, 12(11), 225; https://doi.org/10.3390/computation12110225 - 12 Nov 2024
Viewed by 744
Abstract
Quality inspection of electronic boards during the manufacturing process is a crucial step, especially in the case of specific and expensive power electronic modules. Soldering splash occurrence decreases the reliability and electric properties of final products. This paper aims to compare different YOLOv8 [...] Read more.
Quality inspection of electronic boards during the manufacturing process is a crucial step, especially in the case of specific and expensive power electronic modules. Soldering splash occurrence decreases the reliability and electric properties of final products. This paper aims to compare different YOLOv8 models (small, medium, and large) with the combination of basic image preprocessing techniques to achieve the best possible performance of the designed algorithm. As preprocessing methods, contrast-limited adaptive histogram equalization (CLAHE) and image color channel manipulation are used. The results show that a suitable combination of the YOLOv8 model and preprocessing methods leads to an increase in the recall parameter. In our inspection task, recall can be considered the most important metric. The results are supported by a standard two-way ANOVA test. Full article
(This article belongs to the Section Computational Engineering)
Show Figures

Figure 1

33 pages, 13566 KiB  
Article
KOC_Net: Impact of the Synthetic Minority Over-Sampling Technique with Deep Learning Models for Classification of Knee Osteoarthritis Using Kellgren–Lawrence X-Ray Grade
by Syeda Nida Hassan, Mudassir Khalil, Humayun Salahuddin, Rizwan Ali Naqvi, Daesik Jeong and Seung-Won Lee
Mathematics 2024, 12(22), 3534; https://doi.org/10.3390/math12223534 - 12 Nov 2024
Viewed by 957
Abstract
One of the most common diseases afflicting humans is knee osteoarthritis (KOA). KOA occurs when the knee joint cartilage breaks down, and knee bones start rubbing together. The diagnosis of KOA is a lengthy process, and missed diagnosis can have serious consequences. Therefore, [...] Read more.
One of the most common diseases afflicting humans is knee osteoarthritis (KOA). KOA occurs when the knee joint cartilage breaks down, and knee bones start rubbing together. The diagnosis of KOA is a lengthy process, and missed diagnosis can have serious consequences. Therefore, the diagnosis of KOA at an initial stage is crucial which prevents the patients from Severe complications. KOA identification using deep learning (DL) algorithms has gained popularity during the past few years. By applying knee X-ray images and the Kellgren–Lawrence (KL) grading system, the objective of this study was to develop a DL model for detecting KOA. This study proposes a novel model based on CNN called knee osteoarthritis classification network (KOC_Net). The KOC_Net model contains 05 convolutional blocks, and each convolutional block has three components such as Convlotuioanl2D, ReLU, and MaxPooling 2D. The KOC_Net model is evaluated on two publicly available benchmark datasets which consist of X-ray images of KOA based on the KL grading system. Additionally, we applied contrast-limited adaptive histogram equalization (CLAHE) methods to enhance the contrast of the images and utilized SMOTE Tomek to deal with the problem of minority classes. For the diagnosis of KOA, the classification performance of the proposed KOC_Net model is compared with baseline deep networks, namely Dense Net-169, Vgg-19, Xception, and Inception-V3. The proposed KOC_Net was able to classify KOA into 5 distinct groups (including Moderate, Minimal, Severe, Doubtful, and Healthy), with an AUC of 96.71%, accuracy of 96.51%, recall of 91.95%, precision of 90.25%, and F1-Score of 96.70%. Dense Net-169, Vgg-19, Xception, and Inception-V3 have relative accuracy rates of 84.97%, 81.08%, 87.06%, and 83.62%. As demonstrated by the results, the KOC_Net model provides great assistance to orthopedics in making diagnoses of KOA. Full article
(This article belongs to the Special Issue Deep Learning Methods for Biomedical and Medical Images)
Show Figures

Figure 1

26 pages, 6796 KiB  
Article
A Hybrid Deep Learning and Machine Learning Approach with Mobile-EfficientNet and Grey Wolf Optimizer for Lung and Colon Cancer Histopathology Classification
by Raquel Ochoa-Ornelas, Alberto Gudiño-Ochoa and Julio Alberto García-Rodríguez
Cancers 2024, 16(22), 3791; https://doi.org/10.3390/cancers16223791 - 11 Nov 2024
Cited by 2 | Viewed by 1836
Abstract
Background: Lung and colon cancers are among the most prevalent and lethal malignancies worldwide, underscoring the urgent need for advanced diagnostic methodologies. This study aims to develop a hybrid deep learning and machine learning framework for the classification of Colon Adenocarcinoma, Colon Benign [...] Read more.
Background: Lung and colon cancers are among the most prevalent and lethal malignancies worldwide, underscoring the urgent need for advanced diagnostic methodologies. This study aims to develop a hybrid deep learning and machine learning framework for the classification of Colon Adenocarcinoma, Colon Benign Tissue, Lung Adenocarcinoma, Lung Benign Tissue, and Lung Squamous Cell Carcinoma from histopathological images. Methods: Current approaches primarily rely on the LC25000 dataset, which, due to image augmentation, lacks the generalizability required for real-time clinical applications. To address this, Contrast Limited Adaptive Histogram Equalization (CLAHE) was applied to enhance image quality, and 1000 new images from the National Cancer Institute GDC Data Portal were introduced into the Colon Adenocarcinoma, Lung Adenocarcinoma, and Lung Squamous Cell Carcinoma classes, replacing augmented images to increase dataset diversity. A hybrid feature extraction model combining MobileNetV2 and EfficientNetB3 was optimized using the Grey Wolf Optimizer (GWO), resulting in the Lung and Colon histopathological classification technique (MEGWO-LCCHC). Cross-validation and hyperparameter tuning with Optuna were performed on various machine learning models, including XGBoost, LightGBM, and CatBoost. Results: The MEGWO-LCCHC technique achieved high classification accuracy, with the lightweight DNN model reaching 94.8%, LightGBM at 93.9%, XGBoost at 93.5%, and CatBoost at 93.3% on the test set. Conclusions: The findings suggest that our approach enhances classification performance and offers improved generalizability for real-world clinical applications. The proposed MEGWO-LCCHC framework shows promise as a robust tool in cancer diagnostics, advancing the application of AI in oncology. Full article
(This article belongs to the Special Issue Image Analysis and Machine Learning in Cancers)
Show Figures

Figure 1

18 pages, 2655 KiB  
Article
Advanced Image Preprocessing and Integrated Modeling for UAV Plant Image Classification
by Girma Tariku, Isabella Ghiglieno, Anna Simonetto, Fulvio Gentilin, Stefano Armiraglio, Gianni Gilioli and Ivan Serina
Drones 2024, 8(11), 645; https://doi.org/10.3390/drones8110645 - 6 Nov 2024
Viewed by 1590
Abstract
The automatic identification of plant species using unmanned aerial vehicles (UAVs) is a valuable tool for ecological research. However, challenges such as reduced spatial resolution due to high-altitude operations, image degradation from camera optics and sensor limitations, and information loss caused by terrain [...] Read more.
The automatic identification of plant species using unmanned aerial vehicles (UAVs) is a valuable tool for ecological research. However, challenges such as reduced spatial resolution due to high-altitude operations, image degradation from camera optics and sensor limitations, and information loss caused by terrain shadows hinder the accurate classification of plant species from UAV imagery. This study addresses these issues by proposing a novel image preprocessing pipeline and evaluating its impact on model performance. Our approach improves image quality through a multi-step pipeline that includes Enhanced Super-Resolution Generative Adversarial Networks (ESRGAN) for resolution enhancement, Contrast-Limited Adaptive Histogram Equalization (CLAHE) for contrast improvement, and white balance adjustments for accurate color representation. These preprocessing steps ensure high-quality input data, leading to better model performance. For feature extraction and classification, we employ a pre-trained VGG-16 deep convolutional neural network, followed by machine learning classifiers, including Support Vector Machine (SVM), random forest (RF), and Extreme Gradient Boosting (XGBoost). This hybrid approach, combining deep learning for feature extraction with machine learning for classification, not only enhances classification accuracy but also reduces computational resource requirements compared to relying solely on deep learning models. Notably, the VGG-16 + SVM model achieved an outstanding accuracy of 97.88% on a dataset preprocessed with ESRGAN and white balance adjustments, with a precision of 97.9%, a recall of 97.8%, and an F1 score of 0.978. Through a comprehensive comparative study, we demonstrate that the proposed framework, utilizing VGG-16 for feature extraction, SVM for classification, and preprocessed images with ESRGAN and white balance adjustments, achieves superior performance in plant species identification from UAV imagery. Full article
(This article belongs to the Section Drones in Ecology)
Show Figures

Figure 1

18 pages, 41079 KiB  
Article
Research on Target Image Classification in Low-Light Night Vision
by Yanfeng Li, Yongbiao Luo, Yingjian Zheng, Guiqian Liu and Jiekai Gong
Entropy 2024, 26(10), 882; https://doi.org/10.3390/e26100882 - 21 Oct 2024
Cited by 2 | Viewed by 1276
Abstract
In extremely dark conditions, low-light imaging may offer spectators a rich visual experience, which is important for both military and civic applications. However, the images taken in ultra-micro light environments usually have inherent defects such as extremely low brightness and contrast, a high [...] Read more.
In extremely dark conditions, low-light imaging may offer spectators a rich visual experience, which is important for both military and civic applications. However, the images taken in ultra-micro light environments usually have inherent defects such as extremely low brightness and contrast, a high noise level, and serious loss of scene details and colors, which leads to great challenges in the research of low-light image and object detection and classification. The low-light night vision image used as the study object in this work has an excessively dim overall picture and very little information about the screen’s features. Three algorithms, HE, AHE, and CLAHE, were used to enhance and highlight the image. The effectiveness of these image enhancement methods is evaluated using metrics such as the peak signal-to-noise ratio and mean square error, and CLAHE was selected after comparison. The target image includes vehicles, people, license plates, and objects. The gray-level co-occurrence matrix (GLCM) was used to extract the texture features of the enhanced images, and the extracted image texture features were used as input to construct a backpropagation (BP) neural network classification model. Then, low-light image classification models were developed based on VGG16 and ResNet50 convolutional neural networks combined with low-light image enhancement algorithms. The experimental results show that the overall classification accuracy of the VGG16 convolutional neural network model is 92.1%. Compared with the BP and ResNet50 neural network models, the classification accuracy was increased by 4.5% and 2.3%, respectively, demonstrating its effectiveness in classifying low-light night vision targets. Full article
Show Figures

Figure 1

36 pages, 4195 KiB  
Review
Artificial Intelligence Tools in Pediatric Urology: A Comprehensive Review of Recent Advances
by Adiba Tabassum Chowdhury, Abdus Salam, Mansura Naznine, Da’ad Abdalla, Lauren Erdman, Muhammad E. H. Chowdhury and Tariq O. Abbas
Diagnostics 2024, 14(18), 2059; https://doi.org/10.3390/diagnostics14182059 - 17 Sep 2024
Cited by 2 | Viewed by 2342
Abstract
Artificial intelligence (AI) is providing novel answers to long-standing clinical problems, and it is quickly changing pediatric urology. This thorough analysis focuses on current developments in AI technologies that improve pediatric urology diagnosis, treatment planning, and surgery results. Deep learning algorithms help detect [...] Read more.
Artificial intelligence (AI) is providing novel answers to long-standing clinical problems, and it is quickly changing pediatric urology. This thorough analysis focuses on current developments in AI technologies that improve pediatric urology diagnosis, treatment planning, and surgery results. Deep learning algorithms help detect problems with previously unheard-of precision in disorders including hydronephrosis, pyeloplasty, and vesicoureteral reflux, where AI-powered prediction models have demonstrated promising outcomes in boosting diagnostic accuracy. AI-enhanced image processing methods have significantly improved the quality and interpretation of medical images. Examples of these methods are deep-learning-based segmentation and contrast limited adaptive histogram equalization (CLAHE). These methods guarantee higher precision in the identification and classification of pediatric urological disorders, and AI-driven ground truth construction approaches aid in the standardization of and improvement in training data, resulting in more resilient and consistent segmentation models. AI is being used for surgical support as well. AI-assisted navigation devices help with difficult operations like pyeloplasty by decreasing complications and increasing surgical accuracy. AI also helps with long-term patient monitoring, predictive analytics, and customized treatment strategies, all of which improve results for younger patients. However, there are practical, ethical, and legal issues with AI integration in pediatric urology that need to be carefully navigated. To close knowledge gaps, more investigation is required, especially in the areas of AI-driven surgical methods and standardized ground truth datasets for pediatric radiologic image segmentation. In the end, AI has the potential to completely transform pediatric urology by enhancing patient care, increasing the effectiveness of treatments, and spurring more advancements in this exciting area. Full article
(This article belongs to the Section Machine Learning and Artificial Intelligence in Diagnostics)
Show Figures

Figure 1

26 pages, 9063 KiB  
Article
Forearm Intravenous Detection and Localization for Autonomous Vein Injection Using Contrast-Limited Adaptive Histogram Equalization Algorithm
by Hany Said, Sherif Mohamed, Omar Shalash, Esraa Khatab, Omar Aman, Ramy Shaaban and Mohamed Hesham
Appl. Sci. 2024, 14(16), 7115; https://doi.org/10.3390/app14167115 - 13 Aug 2024
Cited by 1 | Viewed by 1910
Abstract
Occasionally intravenous insertion forms a challenge to a number of patients. Inserting an IV needle is a difficult task that requires a lot of skill. At the moment, only doctors and medical personnel are allowed to do this because it requires finding the [...] Read more.
Occasionally intravenous insertion forms a challenge to a number of patients. Inserting an IV needle is a difficult task that requires a lot of skill. At the moment, only doctors and medical personnel are allowed to do this because it requires finding the right vein, inserting the needle properly, and carefully injecting fluids or drawing out blood. Even for trained professionals, this can be done incorrectly, which can cause bleeding, infection, or damage to the vein. It is especially difficult to do this on children, elderly people, and people with certain skin conditions. In these cases, the veins are harder to see, so it is less likely to be done correctly the first time and may cause blood clots. In this research, a low-cost embedded system utilizing Near-Infrared (NIR) light technology is developed, and two novel approaches are proposed to detect and select the best candidate veins. The two approaches utilize multiple computer vision tools and are based on contrast-limited adaptive histogram equalization (CLAHE). The accuracy of the proposed algorithm is 91.3% with an average 1.4 s processing time on Raspberry Pi 4 Model B. Full article
Show Figures

Figure 1

Back to TopTop