Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (89)

Search Parameters:
Keywords = semi-automatic annotation

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 1344 KB  
Article
Research on Intelligent Extraction Method of Influencing Factors of Loess Landslide Geological Disasters Based on Soft-Lexicon and GloVe
by Lutong Huang, Yueqin Zhu, Yingfei Li, Tianxiao Yan, Yu Xiao, Dongqi Wei, Ziyao Xing and Jian Li
Appl. Sci. 2025, 15(16), 8879; https://doi.org/10.3390/app15168879 - 12 Aug 2025
Viewed by 298
Abstract
Loess landslide disasters are influenced by a multitude of factors, including slope conditions, triggering mechanisms, and spatial attributes. Extracting these factors from unstructured geological texts is challenging due to nested entities, semantic ambiguity, and rare domain-specific terms. This study proposes a joint extraction [...] Read more.
Loess landslide disasters are influenced by a multitude of factors, including slope conditions, triggering mechanisms, and spatial attributes. Extracting these factors from unstructured geological texts is challenging due to nested entities, semantic ambiguity, and rare domain-specific terms. This study proposes a joint extraction framework guided by a domain ontology that categorizes six types of loess landslide influencing factors, including spatial relationships. The ontology facilitates conceptual classification and semi-automatic nested entity annotation, enabling the construction of a high-quality corpus with eight tag types. The model integrates a Soft-Lexicon mechanism that enhances character-level GloVe embeddings with explicit lexical features, including domain terms, part-of-speech tags, and word boundary indicators derived from a domain-specific lexicon. The resulting hybrid character-level representations are then fed into a BiLSTM-CRF architecture to jointly extract entities, attributes, and multi-level spatial and causal relationships. Extracted results are structured using a content-knowledge model to build a spatially enriched knowledge graph, supporting semantic queries and intelligent reasoning. Experimental results demonstrate improved performance over baseline methods, showcasing the framework’s effectiveness in geohazard information extraction and disaster risk analysis. Full article
(This article belongs to the Special Issue Applications of Big Data and Artificial Intelligence in Geoscience)
Show Figures

Figure 1

17 pages, 1519 KB  
Article
TOM-SSL: Tomato Disease Recognition Using Pseudo-Labelling-Based Semi-Supervised Learning
by Sathiyamohan Nishankar, Thurairatnam Mithuran, Selvarajah Thuseethan, Yakub Sebastian, Kheng Cher Yeo and Bharanidharan Shanmugam
AgriEngineering 2025, 7(8), 248; https://doi.org/10.3390/agriengineering7080248 - 5 Aug 2025
Viewed by 727
Abstract
In the agricultural domain, the availability of labelled data for disease recognition tasks is often limited due to the cost and expertise required for annotation. In this paper, a novel semi-supervised learning framework named TOM-SSL is proposed for automatic tomato leaf disease recognition [...] Read more.
In the agricultural domain, the availability of labelled data for disease recognition tasks is often limited due to the cost and expertise required for annotation. In this paper, a novel semi-supervised learning framework named TOM-SSL is proposed for automatic tomato leaf disease recognition using pseudo-labelling. TOM-SSL effectively addresses the challenge of limited labelled data by leveraging a small labelled subset and confidently pseudo-labelled samples from a large pool of unlabelled data to improve classification performance. Utilising only 10% of the labelled data, the proposed framework with a MobileNetV3-Small backbone achieves the best accuracy at 72.51% on the tomato subset of the PlantVillage dataset and 70.87% on the Taiwan tomato leaf disease dataset across 10 disease categories in PlantVillage and 6 in the Taiwan dataset. While achieving recognition performance on par with current state-of-the-art supervised methods, notably, the proposed approach offers a tenfold enhancement in label efficiency. Full article
Show Figures

Figure 1

19 pages, 43909 KB  
Article
DualBranch-AMR: A Semi-Supervised AMR Method Based on Dual-Student Consistency Regularization with Dynamic Stability Evaluation
by Jiankun Ma, Zhenxi Zhang, Linrun Zhang, Yu Li, Haoyue Tan, Xiaoran Shi and Feng Zhou
Sensors 2025, 25(15), 4553; https://doi.org/10.3390/s25154553 - 23 Jul 2025
Viewed by 393
Abstract
Modulation recognition, as one of the key technologies in the field of wireless communications, holds significant importance in applications such as spectrum resource management, interference suppression, and cognitive radio. While deep learning has substantially improved the performance of Automatic Modulation Recognition (AMR), it [...] Read more.
Modulation recognition, as one of the key technologies in the field of wireless communications, holds significant importance in applications such as spectrum resource management, interference suppression, and cognitive radio. While deep learning has substantially improved the performance of Automatic Modulation Recognition (AMR), it heavily relies on large amounts of labeled data. Given the high annotation costs and privacy concerns, researching semi-supervised AMR methods that leverage readily available unlabeled data for training is of great significance. This study constructs a semi-supervised AMR method based on dual-student. Specifically, we first adopt a dual-branch co-training architecture to fully exploit unlabeled data and effectively learn deep feature representations. Then, we develop a dynamic stability evaluation module using strong and weak augmentation strategies to improve the accuracy of generated pseudo-labels. Finally, based on the dual-student semi-supervised framework and pseudo-label stability evaluation, we propose a stability-guided consistency regularization constraint method and conduct semi-supervised AMR model training. The experimental results demonstrate that the proposed DualBranch-AMR method significantly outperforms traditional supervised baseline approaches on benchmark datasets. With only 5% labeled data, it achieves a recognition accuracy of 55.84%, reaching over 90% of the performance of fully supervised training. This validates the superiority of the proposed method under semi-supervised conditions. Full article
(This article belongs to the Section Communications)
Show Figures

Figure 1

18 pages, 4066 KB  
Article
Video Segmentation of Wire + Arc Additive Manufacturing (WAAM) Using Visual Large Model
by Shuo Feng, James Wainwright, Chong Wang, Jun Wang, Goncalo Rodrigues Pardal, Jian Qin, Yi Yin, Shakirudeen Lasisi, Jialuo Ding and Stewart Williams
Sensors 2025, 25(14), 4346; https://doi.org/10.3390/s25144346 - 11 Jul 2025
Viewed by 733
Abstract
Process control and quality assurance of wire + arc additive manufacturing (WAAM) and automated welding rely heavily on in-process monitoring videos to quantify variables such as melt pool geometry, location and size of droplet transfer, arc characteristics, etc. To enable feedback control based [...] Read more.
Process control and quality assurance of wire + arc additive manufacturing (WAAM) and automated welding rely heavily on in-process monitoring videos to quantify variables such as melt pool geometry, location and size of droplet transfer, arc characteristics, etc. To enable feedback control based upon this information, an automatic and robust segmentation method for monitoring of videos and images is required. However, video segmentation in WAAM and welding is challenging due to constantly fluctuating arc brightness, which varies with deposition and welding configurations. Additionally, conventional computer vision algorithms based on greyscale value and gradient lack flexibility and robustness in this scenario. Deep learning offers a promising approach to WAAM video segmentation; however, the prohibitive time and cost associated with creating a well-labelled, suitably sized dataset have hindered its widespread adoption. The emergence of large computer vision models, however, has provided new solutions. In this study a semi-automatic annotation tool for WAAM videos was developed based upon the computer vision foundation model SAM and the video object tracking model XMem. The tool can enable annotation of the video frames hundreds of times faster than traditional manual annotation methods, thus making it possible to achieve rapid quantitative analysis of WAAM and welding videos with minimal user intervention. To demonstrate the effectiveness of the tool, three cases are demonstrated: online wire position closed-loop control, droplet transfer behaviour analysis, and assembling a dataset for dedicated deep learning segmentation models. This work provides a broader perspective on how to exploit large models in WAAM and weld deposits. Full article
(This article belongs to the Special Issue Sensing and Imaging in Computer Vision)
Show Figures

Figure 1

13 pages, 3291 KB  
Technical Note
Semi-Automated Training of AI Vision Models
by Mathew G. Pelletier, John D. Wanjura and Greg A. Holt
AgriEngineering 2025, 7(7), 225; https://doi.org/10.3390/agriengineering7070225 - 8 Jul 2025
Viewed by 526
Abstract
The adoption of AI vision models in specialized industries is often hindered by the substantial requirement for extensive, manually annotated image datasets. Even when employing transfer learning, robust model development typically necessitates tens of thousands of such images, a process that is time-consuming, [...] Read more.
The adoption of AI vision models in specialized industries is often hindered by the substantial requirement for extensive, manually annotated image datasets. Even when employing transfer learning, robust model development typically necessitates tens of thousands of such images, a process that is time-consuming, costly, and demands consistent expert annotation. This technical note introduces a semi-automated method to significantly reduce this annotation burden. The proposed approach utilizes two general-purpose vision-transformer-to-caption (GP-ViTC) models to generate descriptive text from images. These captions are then processed by a custom-developed semantic classifier (SC), which requires only minimal training to predict the correct image class. This GP-ViTC + SC system demonstrated exemplary classification rates in test cases and can subsequently be used to automatically annotate large image datasets. While the inference speed of the GP-ViTC models is not suited for real-time applications (approximately 10 s per image), this method substantially lessens the labor and expertise required for dataset creation, thereby facilitating the development of new, high-speed, custom AI vision models for niche applications. This work details the approach and its successful application, offering a cost-effective pathway for generating tailored image training sets. Full article
Show Figures

Figure 1

35 pages, 2865 KB  
Article
eyeNotate: Interactive Annotation of Mobile Eye Tracking Data Based on Few-Shot Image Classification
by Michael Barz, Omair Shahzad Bhatti, Hasan Md Tusfiqur Alam, Duy Minh Ho Nguyen, Kristin Altmeyer, Sarah Malone and Daniel Sonntag
J. Eye Mov. Res. 2025, 18(4), 27; https://doi.org/10.3390/jemr18040027 - 7 Jul 2025
Viewed by 806
Abstract
Mobile eye tracking is an important tool in psychology and human-centered interaction design for understanding how people process visual scenes and user interfaces. However, analyzing recordings from head-mounted eye trackers, which typically include an egocentric video of the scene and a gaze signal, [...] Read more.
Mobile eye tracking is an important tool in psychology and human-centered interaction design for understanding how people process visual scenes and user interfaces. However, analyzing recordings from head-mounted eye trackers, which typically include an egocentric video of the scene and a gaze signal, is a time-consuming and largely manual process. To address this challenge, we develop eyeNotate, a web-based annotation tool that enables semi-automatic data annotation and learns to improve from corrective user feedback. Users can manually map fixation events to areas of interest (AOIs) in a video-editing-style interface (baseline version). Further, our tool can generate fixation-to-AOI mapping suggestions based on a few-shot image classification model (IML-support version). We conduct an expert study with trained annotators (n = 3) to compare the baseline and IML-support versions. We measure the perceived usability, annotations’ validity and reliability, and efficiency during a data annotation task. We asked our participants to re-annotate data from a single individual using an existing dataset (n = 48). Further, we conducted a semi-structured interview to understand how participants used the provided IML features and assessed our design decisions. In a post hoc experiment, we investigate the performance of three image classification models in annotating data of the remaining 47 individuals. Full article
Show Figures

Figure 1

23 pages, 2410 KB  
Article
A Semi-Automatic Framework for Practical Transcription of Foreign Person Names in Lithuanian
by Gailius Raškinis, Darius Amilevičius, Danguolė Kalinauskaitė, Artūras Mickus, Daiva Vitkutė-Adžgauskienė, Antanas Čenys and Tomas Krilavičius
Mathematics 2025, 13(13), 2107; https://doi.org/10.3390/math13132107 - 27 Jun 2025
Viewed by 483
Abstract
We present a semi-automatic framework for transcribing foreign personal names into Lithuanian, aimed at reducing pronunciation errors in text-to-speech systems. Focusing on noisy, web-crawled data, the pipeline combines rule-based filtering, morphological normalization, and manual stress annotation—the only non-automated step—to generate training data for [...] Read more.
We present a semi-automatic framework for transcribing foreign personal names into Lithuanian, aimed at reducing pronunciation errors in text-to-speech systems. Focusing on noisy, web-crawled data, the pipeline combines rule-based filtering, morphological normalization, and manual stress annotation—the only non-automated step—to generate training data for character-level transcription models. We evaluate three approaches: a weighted finite-state transducer (WFST), an LSTM-based sequence-to-sequence model with attention, and a Transformer model optimized for character transduction. Results show that word-pair models outperform single-word models, with the Transformer achieving the best performance (19.04% WER) on a cleaned and augmented dataset. Data augmentation via word order reversal proved effective, while combining single-word and word-pair training offered limited gains. Despite filtering, residual noise persists, with 54% of outputs showing some error, though only 11% were perceptually significant. Full article
(This article belongs to the Section E1: Mathematics and Computer Science)
Show Figures

Figure 1

19 pages, 3825 KB  
Article
A Semi-Supervised Attention-Temporal Ensembling Method for Ground Penetrating Radar Target Recognition
by Li Liu, Dajiang Yu, Xiping Zhang, Hang Xu, Jingxia Li, Lijun Zhou and Bingjie Wang
Sensors 2025, 25(10), 3138; https://doi.org/10.3390/s25103138 - 15 May 2025
Viewed by 677
Abstract
Ground penetrating radar (GPR) is an effective and efficient nondestructive testing technology for subsurface investigations. Deep learning-based methods have been successfully used in automatic underground target recognition. However, these methods are mostly based on supervised learning, requiring large amounts of labeled training data [...] Read more.
Ground penetrating radar (GPR) is an effective and efficient nondestructive testing technology for subsurface investigations. Deep learning-based methods have been successfully used in automatic underground target recognition. However, these methods are mostly based on supervised learning, requiring large amounts of labeled training data to guarantee high accuracy and generalization ability, which is a challenge in GPR fields due to time-consuming and labor-intensive data annotation work. To alleviate the demand for abundant labeled data, a semi-supervised deep learning method named attention–temporal ensembling (Attention-TE) is proposed for underground target recognition using GPR B-scan images. This method integrates a semi-supervised temporal ensembling architecture with a triplet attention module to enhance the classification performance. Experimental results of laboratory and field data demonstrate that the proposed method can automatically recognize underground targets with an average accuracy of above 90% using less than 30% of labeled data in the training dataset. Ablation experimental results verify the efficiency of the triplet attention module. Moreover, comparative experimental results validate that the proposed Attention-TE algorithm outperforms the supervised method based on transfer learning and four semi-supervised state-of-the-art methods. Full article
Show Figures

Figure 1

10 pages, 11728 KB  
Article
Real-World Colonoscopy Video Integration to Improve Artificial Intelligence Polyp Detection Performance and Reduce Manual Annotation Labor
by Yuna Kim, Ji-Soo Keum, Jie-Hyun Kim, Jaeyoung Chun, Sang-Il Oh, Kyung-Nam Kim, Young-Hoon Yoon and Hyojin Park
Diagnostics 2025, 15(7), 901; https://doi.org/10.3390/diagnostics15070901 - 1 Apr 2025
Viewed by 1036
Abstract
Background/Objectives: Artificial intelligence (AI) integration in colon polyp detection often exhibits high sensitivity but notably low specificity in real-world settings, primarily due to reliance on publicly available datasets alone. To address this limitation, we proposed a semi-automatic annotation method using real colonoscopy [...] Read more.
Background/Objectives: Artificial intelligence (AI) integration in colon polyp detection often exhibits high sensitivity but notably low specificity in real-world settings, primarily due to reliance on publicly available datasets alone. To address this limitation, we proposed a semi-automatic annotation method using real colonoscopy videos to enhance AI model performance and reduce manual labeling labor. Methods: An integrated AI model was trained and validated on 86,258 training images and 17,616 validation images. Model 1 utilized only publicly available datasets, while Model 2 additionally incorporated images obtained from real colonoscopy videos of patients through a semi-automatic annotation process, significantly reducing the labeling burden on expert endoscopists. Results: The integrated AI model (Model 2) significantly outperformed the public-dataset-only model (Model 1). At epoch 35, Model 2 achieved a sensitivity of 90.6%, a specificity of 96.0%, an overall accuracy of 94.5%, and an F1 score of 89.9%. All polyps in the test videos were successfully detected, demonstrating considerable enhancement in detection performance compared to the public-dataset-only model. Conclusions: Integrating real-world colonoscopy video data using semi-automatic annotation markedly improved diagnostic accuracy while potentially reducing the need for extensive manual annotation typically performed by expert endoscopists. However, the findings need validation through multicenter external datasets to ensure generalizability. Full article
(This article belongs to the Special Issue Application of Artificial Intelligence in Gastrointestinal Disease)
Show Figures

Figure 1

20 pages, 15232 KB  
Article
Swift Transfer of Lactating Piglet Detection Model Using Semi-Automatic Annotation Under an Unfamiliar Pig Farming Environment
by Qi’an Ding, Fang Zheng, Luo Liu, Peng Li and Mingxia Shen
Agriculture 2025, 15(7), 696; https://doi.org/10.3390/agriculture15070696 - 25 Mar 2025
Cited by 2 | Viewed by 454
Abstract
Manual annotation of piglet imagery across varied farming environments is labor-intensive. To address this, we propose a semi-automatic approach within an active learning framework that integrates a pre-annotation model for piglet detection. We further examine how data sample composition influences pre-annotation efficiency to [...] Read more.
Manual annotation of piglet imagery across varied farming environments is labor-intensive. To address this, we propose a semi-automatic approach within an active learning framework that integrates a pre-annotation model for piglet detection. We further examine how data sample composition influences pre-annotation efficiency to enhance the deployment of lactating piglet detection models. Our study utilizes original samples from pig farms in Jingjiang, Suqian, and Sheyang, along with new data from the Yinguang pig farm in Danyang. Using the YOLOv5 framework, we constructed both single and mixed training sets of piglet images, evaluated their performance, and selected the optimal pre-annotation model. This model generated bounding box coordinates on processed new samples, which were subsequently manually refined to train the final model. Results indicate that expanding the dataset and diversifying pigpen scenes significantly improve pre-annotation performance. The best model achieved a test precision of 0.921 on new samples, and after manual calibration, the final model exhibited a training precision of 0.968, a recall of 0.952, and an average precision of 0.979 at the IoU threshold of 0.5. The model demonstrated robust detection under various lighting conditions, with bounding boxes closely conforming to piglet contours, thereby substantially reducing manual labor. This approach is cost-effective for piglet segmentation tasks and offers strong support for advancing smart agricultural technologies. Full article
Show Figures

Figure 1

17 pages, 4646 KB  
Article
Mixed-Supervised Learning for Cell Classification
by Hao Sun, Danqi Guo and Zhao Chen
Sensors 2025, 25(4), 1207; https://doi.org/10.3390/s25041207 - 16 Feb 2025
Viewed by 833
Abstract
Cell classification based on histopathology images is crucial for tumor recognition and cancer diagnosis. Using deep learning, classification accuracy is hugely improved. Semi-supervised learning is an advanced deep learning approach that uses both labeled and unlabeled data. However, complex datasets that comprise diverse [...] Read more.
Cell classification based on histopathology images is crucial for tumor recognition and cancer diagnosis. Using deep learning, classification accuracy is hugely improved. Semi-supervised learning is an advanced deep learning approach that uses both labeled and unlabeled data. However, complex datasets that comprise diverse patterns may drive models towards learning harmful features. Therefore, it is useful to involve human guidance during training. Hence, we propose a mixed-supervised method incorporating semi-supervision and “human-in-the-loop” for cell classification. We design a sample selection mechanism that assigns highly confident unlabeled samples to automatic semi-supervised optimization and unreliable ones for online annotation correction. We use prior human annotations to pretrain the backbone and trustworthy pseudo labels and online human annotations to fine-tune the model for accurate cell classification. Experimental results show that the mixed-supervised model reaches overall accuracies as high as 86.56%, 99.33% and 74.12% on LUSC, BloodCell, and PanNuke datasets, respectively. Full article
Show Figures

Figure 1

23 pages, 8710 KB  
Article
Deep Learning-Based Algorithms for Real-Time Lung Ultrasound Assisted Diagnosis
by Mario Muñoz, Adrián Rubio, Guillermo Cosarinsky, Jorge F. Cruza and Jorge Camacho
Appl. Sci. 2024, 14(24), 11930; https://doi.org/10.3390/app142411930 - 20 Dec 2024
Cited by 2 | Viewed by 2348
Abstract
Lung ultrasound is an increasingly utilized non-invasive imaging modality for assessing lung condition but interpreting it can be challenging and depends on the operator’s experience. To address these challenges, this work proposes an approach that combines artificial intelligence (AI) with feature-based signal processing [...] Read more.
Lung ultrasound is an increasingly utilized non-invasive imaging modality for assessing lung condition but interpreting it can be challenging and depends on the operator’s experience. To address these challenges, this work proposes an approach that combines artificial intelligence (AI) with feature-based signal processing algorithms. We introduce a specialized deep learning model designed and trained to facilitate the analysis and interpretation of lung ultrasound images by automating the detection and location of pulmonary features, including the pleura, A-lines, B-lines, and consolidations. Employing Convolutional Neural Networks (CNNs) trained on a semi-automatically annotated dataset, the model delineates these pulmonary patterns with the objective of enhancing diagnostic precision. Real-time post-processing algorithms further refine prediction accuracy by reducing false-positives and false-negatives, augmenting interpretational clarity and obtaining a final processing rate of up to 20 frames per second with accuracy levels of 89% for consolidation, 92% for B-lines, 66% for A-lines, and 92% for detecting normal lungs compared with an expert opinion. Full article
Show Figures

Figure 1

20 pages, 772 KB  
Communication
SOAMC: A Semi-Supervised Open-Set Recognition Algorithm for Automatic Modulation Classification
by Chengliang Di, Jinwei Ji, Chao Sun and Linlin Liang
Electronics 2024, 13(21), 4196; https://doi.org/10.3390/electronics13214196 - 25 Oct 2024
Cited by 2 | Viewed by 1605
Abstract
Traditional automatic modulation classification methods operate under the closed-set assumption, which proves to be impractical in real-world scenarios due to the diverse nature of wireless technologies and the dynamic characteristics of wireless propagation environments. Open-set environments introduce substantial technical challenges, particularly in terms [...] Read more.
Traditional automatic modulation classification methods operate under the closed-set assumption, which proves to be impractical in real-world scenarios due to the diverse nature of wireless technologies and the dynamic characteristics of wireless propagation environments. Open-set environments introduce substantial technical challenges, particularly in terms of detection effectiveness and computational complexity. To address the limitations of modulation classification and recognition in open-set scenarios, this paper proposes a semi-supervised open-set recognition approach, termed SOAMC (Semi-Supervised Open-Set Automatic Modulation Classification). The primary objective of SOAMC is to accurately classify unknown modulation types, even when only a limited subset of samples is manually labeled. The proposed method consists of three key stages: (1) A signal recognition pre-training model is constructed using data augmentation and adaptive techniques to enhance robustness. (2) Feature extraction and embedding are performed via a specialized extraction network. (3) Label propagation is executed using a graph convolutional neural network (GCN) to efficiently annotate the unlabeled signal samples. Experimental results demonstrate that SOAMC significantly improves classification accuracy, particularly in challenging scenarios with limited amounts of labeled data and high signal similarity. These findings are critical for the practical identification of complex and diverse modulation signals in real-world wireless communication systems. Full article
Show Figures

Figure 1

18 pages, 16454 KB  
Technical Note
Annotated Dataset for Training Cloud Segmentation Neural Networks Using High-Resolution Satellite Remote Sensing Imagery
by Mingyuan He, Jie Zhang, Yang He, Xinjie Zuo and Zebin Gao
Remote Sens. 2024, 16(19), 3682; https://doi.org/10.3390/rs16193682 - 2 Oct 2024
Cited by 2 | Viewed by 2655
Abstract
The integration of satellite data with deep learning has revolutionized various tasks in remote sensing, including classification, object detection, and semantic segmentation. Cloud segmentation in high-resolution satellite imagery is a critical application within this domain, yet progress in developing advanced algorithms for this [...] Read more.
The integration of satellite data with deep learning has revolutionized various tasks in remote sensing, including classification, object detection, and semantic segmentation. Cloud segmentation in high-resolution satellite imagery is a critical application within this domain, yet progress in developing advanced algorithms for this task has been hindered by the scarcity of specialized datasets and annotation tools. This study addresses this challenge by introducing CloudLabel, a semi-automatic annotation technique leveraging region growing and morphological algorithms including flood fill, connected components, and guided filter. CloudLabel v1.0 streamlines the annotation process for high-resolution satellite images, thereby addressing the limitations of existing annotation platforms which are not specifically adapted to cloud segmentation, and enabling the efficient creation of high-quality cloud segmentation datasets. Notably, we have curated the Annotated Dataset for Training Cloud Segmentation (ADTCS) comprising 32,065 images (512 × 512) for cloud segmentation based on CloudLabel. The ADTCS dataset facilitates algorithmic advancement in cloud segmentation, characterized by uniform cloud coverage distribution and high image entropy (mainly 5–7). These features enable deep learning models to capture comprehensive cloud characteristics, enhancing recognition accuracy and reducing ground object misclassification. This contribution significantly advances remote sensing applications and cloud segmentation algorithms. Full article
Show Figures

Figure 1

19 pages, 1112 KB  
Article
Semi-Supervised Interior Decoration Style Classification with Contrastive Mutual Learning
by Lichun Guo, Hao Zeng, Xun Shi, Qing Xu, Jinhui Shi, Kui Bai, Shuang Liang and Wenlong Hang
Mathematics 2024, 12(19), 2980; https://doi.org/10.3390/math12192980 - 25 Sep 2024
Cited by 1 | Viewed by 1258
Abstract
Precisely identifying interior decoration styles holds substantial significance in directing interior decoration practices. Nevertheless, constructing accurate models for the automatic classification of interior decoration styles remains challenging due to the scarcity of expert annotations. To address this problem, we propose a novel pseudo-label-guided [...] Read more.
Precisely identifying interior decoration styles holds substantial significance in directing interior decoration practices. Nevertheless, constructing accurate models for the automatic classification of interior decoration styles remains challenging due to the scarcity of expert annotations. To address this problem, we propose a novel pseudo-label-guided contrastive mutual learning framework (PCML) for semi-supervised interior decoration style classification by harnessing large amounts of unlabeled data. Specifically, PCML introduces two distinct subnetworks and selectively utilizes the diversified pseudo-labels generated by each for mutual supervision, thereby mitigating the issue of confirmation bias. For labeled images, the inconsistent pseudo-labels generated by the two subnetworks are employed to identify images that are prone to misclassification. We then devise an inconsistency-aware relearning (ICR) regularization model to perform a review training process. For unlabeled images, we introduce a class-aware contrastive learning (CCL) regularization to learn their discriminative feature representations using the corresponding pseudo-labels. Since the use of distinct subnetworks reduces the risk of both models producing identical erroneous pseudo-labels, CCL can reduce the possibility of noise data sampling to enhance the effectiveness of contrastive learning. The performance of PCML is evaluated on five interior decoration style image datasets. For the average AUC, accuracy, sensitivity, specificity, precision, and F1 scores, PCML obtains improvements of 1.67%, 1.72%, 3.65%, 1.0%, 4.61%, and 4.66% in comparison with the state-of-the-art method, demonstrating the superiority of our method. Full article
(This article belongs to the Topic AI and Data-Driven Advancements in Industry 4.0)
Show Figures

Figure 1

Back to TopTop