Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (237)

Search Parameters:
Keywords = unsupervised pre-training

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
13 pages, 2561 KB  
Article
Unsupervised Bearing Fault Diagnosis Using Masked Self-Supervised Learning and Swin Transformer
by Pengping Luo and Zhiwei Liu
Machines 2025, 13(9), 792; https://doi.org/10.3390/machines13090792 - 1 Sep 2025
Viewed by 98
Abstract
Bearings are vital to rotating machinery, where undetected faults can cause severe failures. Conventional fault diagnosis methods depend on manual feature engineering and labeled data, struggling with complex industrial conditions. This study introduces an innovative unsupervised framework combining masked self-supervised learning with the [...] Read more.
Bearings are vital to rotating machinery, where undetected faults can cause severe failures. Conventional fault diagnosis methods depend on manual feature engineering and labeled data, struggling with complex industrial conditions. This study introduces an innovative unsupervised framework combining masked self-supervised learning with the Swin Transformer for bearing fault diagnosis. The novel integration leverages masked Auto Encoders to learn robust features from unlabeled vibration signals through reconstruction-based pretraining, while the Swin Transformer’s shifted window attention mechanism enhances efficient capture of fault-related patterns in long-sequence signals. This approach eliminates reliance on labeled data, enabling precise detection of unknown faults. The proposed method achieves 99.53% accuracy on the Paderborn dataset and 100% accuracy on the CWRU dataset significantly, surpassing other unsupervised Auto Encoder-based methods. This method’s innovative design offers high adaptability and substantial potential for predictive maintenance in industrial applications. Full article
Show Figures

Figure 1

22 pages, 1390 KB  
Article
Masked and Clustered Pre-Training for Geosynchronous Satellite Maneuver Detection
by Shu-He Tian, Yu-Qiang Fang, Hua-Fei Diao, Di Luo and Ya-Sheng Zhang
Remote Sens. 2025, 17(17), 2994; https://doi.org/10.3390/rs17172994 - 28 Aug 2025
Viewed by 318
Abstract
Geosynchronous satellite maneuver detection is critical for enhancing space situational awareness and inferring satellite intent. However, traditional methods often require high-quality orbital sequence data and heavily rely on hand-crafted features, limiting their effectiveness in complex real-world environments. While recent neural network-based approaches have [...] Read more.
Geosynchronous satellite maneuver detection is critical for enhancing space situational awareness and inferring satellite intent. However, traditional methods often require high-quality orbital sequence data and heavily rely on hand-crafted features, limiting their effectiveness in complex real-world environments. While recent neural network-based approaches have shown promise, they are typically trained in scene or task-specific settings, resulting in limited generalization and adaptability. To address these challenges, we propose MC-MD, a pre-training framework that integrates Masked and Clustered learning strategies to improve the robustness and transferability of geosynchronous satellite Maneuver Detection. Specifically, we introduce a masked prediction module that applies both time- and frequency-domain masking to help the model capture temporal dynamics more effectively. Meanwhile, a cluster-based module guides the model to learn discriminative representations of different maneuver patterns through unsupervised clustering, mitigating the negative impact of distribution shifts across scenarios. By combining these two strategies, MC-MD captures diverse maneuver behaviors and enhances cross-scenario detection performance. Extensive experiments on both simulated and real-world datasets demonstrate that MCMD achieves significant performance gains over the strongest baseline, with improvements of 8.54% in Precision and 7.8% in F1-Score. Furthermore, reconstructed trajectories analysis shows that MC-MD more accurately aligns with the ground-truth maneuver sequence, highlighting its effectiveness in satellite maneuver detection tasks. Full article
Show Figures

Figure 1

25 pages, 24334 KB  
Article
Unsupervised Knowledge Extraction of Distinctive Landmarks from Earth Imagery Using Deep Feature Outliers for Robust UAV Geo-Localization
by Zakhar Ostrovskyi, Oleksander Barmak, Pavlo Radiuk and Iurii Krak
Mach. Learn. Knowl. Extr. 2025, 7(3), 81; https://doi.org/10.3390/make7030081 - 13 Aug 2025
Viewed by 384
Abstract
Vision-based navigation is a common solution for the critical challenge of GPS-denied Unmanned Aerial Vehicle (UAV) operation, but a research gap remains in the autonomous discovery of robust landmarks from aerial survey imagery needed for such systems. In this work, we propose a [...] Read more.
Vision-based navigation is a common solution for the critical challenge of GPS-denied Unmanned Aerial Vehicle (UAV) operation, but a research gap remains in the autonomous discovery of robust landmarks from aerial survey imagery needed for such systems. In this work, we propose a framework to fill this gap by identifying visually distinctive urban buildings from aerial survey imagery and curating them into a landmark database for GPS-free UAV localization. The proposed framework constructs semantically rich embeddings using intermediate layers from a pre-trained YOLOv11n-seg segmentation network. This novel technique requires no additional training. An unsupervised landmark selection strategy, based on the Isolation Forest algorithm, then identifies objects with statistically unique embeddings. Experimental validation on the VPAIR aerial-to-aerial benchmark shows that the proposed max-pooled embeddings, assembled from selected layers, significantly improve retrieval performance. The top-1 retrieval accuracy for landmarks more than doubled compared to typical buildings (0.53 vs. 0.31), and a Recall@5 of 0.70 is achieved for landmarks. Overall, this study demonstrates that unsupervised outlier selection in a carefully constructed embedding space yields a highly discriminative, computation-friendly set of landmarks suitable for real-time, robust UAV navigation. Full article
(This article belongs to the Special Issue Deep Learning in Image Analysis and Pattern Recognition, 2nd Edition)
Show Figures

Graphical abstract

23 pages, 13529 KB  
Article
A Self-Supervised Contrastive Framework for Specific Emitter Identification with Limited Labeled Data
by Jiaqi Wang, Lishu Guo, Pengfei Liu, Peng Shang, Xiaochun Lu and Hang Zhao
Remote Sens. 2025, 17(15), 2659; https://doi.org/10.3390/rs17152659 - 1 Aug 2025
Viewed by 367
Abstract
Specific Emitter Identification (SEI) is a specialized technique for identifying different emitters by analyzing the unique characteristics embedded in received signals, known as Radio Frequency Fingerprints (RFFs), and SEI plays a crucial role in civilian applications. Recently, various SEI methods based on deep [...] Read more.
Specific Emitter Identification (SEI) is a specialized technique for identifying different emitters by analyzing the unique characteristics embedded in received signals, known as Radio Frequency Fingerprints (RFFs), and SEI plays a crucial role in civilian applications. Recently, various SEI methods based on deep learning have been proposed. However, in real-world scenarios, the scarcity of accurately labeled data poses a significant challenge to these methods, which typically rely on large-scale supervised training. To address this issue, we propose a novel SEI framework based on self-supervised contrastive learning. Our approach comprises two stages: an unsupervised pretraining phase that uses contrastive loss to learn discriminative RFF representations from unlabeled data, and a supervised fine-tuning stage regularized through virtual adversarial training (VAT) to improve generalization under limited labels. This framework enables effective feature learning while mitigating overfitting. To validate the effectiveness of the proposed method, we collected real-world satellite navigation signals using a 40-meter antenna and conducted extensive experiments. The results demonstrate that our approach achieves outstanding SEI performance, significantly outperforming several mainstream SEI methods, thereby highlighting the practical potential of contrastive self-supervised learning in satellite transmitter identification. Full article
Show Figures

Graphical abstract

19 pages, 1088 KB  
Article
The Specialist’s Paradox: Generalist AI May Better Organize Medical Knowledge
by Carlo Galli, Maria Teresa Colangelo, Marco Meleti and Elena Calciolari
Algorithms 2025, 18(7), 451; https://doi.org/10.3390/a18070451 - 21 Jul 2025
Viewed by 432
Abstract
This study investigates the ability of six pre-trained sentence transformers to organize medical knowledge by performing unsupervised clustering on 70 high-level Medical Subject Headings (MeSH) terms across seven medical specialties. We evaluated models from different pre-training paradigms: general-purpose, domain-adapted, and from-scratch domain-specific. The [...] Read more.
This study investigates the ability of six pre-trained sentence transformers to organize medical knowledge by performing unsupervised clustering on 70 high-level Medical Subject Headings (MeSH) terms across seven medical specialties. We evaluated models from different pre-training paradigms: general-purpose, domain-adapted, and from-scratch domain-specific. The results reveal a clear performance hierarchy. A top tier of models, including the general-purpose MPNet and the domain-adapted BioBERT and RoBERTa, produced highly coherent, specialty-aligned clusters (Adjusted Rand Index > 0.80). Conversely, models pre-trained from scratch on specialized corpora, such as PubMedBERT and BioClinicalBERT, performed poorly (Adjusted Rand Index < 0.51), with BioClinicalBERT yielding a disorganized clustering. These findings challenge the assumption that domain-specific pre-training guarantees superior performance for all semantic tasks. We conclude that model architecture, alignment between the pre-training objective and the downstream task, and the nature of the training data are more critical determinants of success for creating semantically coherent embedding spaces for medical concepts. Full article
(This article belongs to the Special Issue Evolution of Algorithms in the Era of Generative AI)
Show Figures

Figure 1

43 pages, 6844 KB  
Article
CORE-ReID V2: Advancing the Domain Adaptation for Object Re-Identification with Optimized Training and Ensemble Fusion
by Trinh Quoc Nguyen, Oky Dicky Ardiansyah Prima, Syahid Al Irfan, Hindriyanto Dwi Purnomo and Radius Tanone
AI Sens. 2025, 1(1), 4; https://doi.org/10.3390/aisens1010004 - 4 Jul 2025
Viewed by 891
Abstract
This study presents CORE-ReID V2, an enhanced framework built upon CORE-ReID V1. The new framework extends its predecessor by addressing unsupervised domain adaptation (UDA) challenges in person ReID and vehicle ReID, with further applicability to object ReID. During pre-training, CycleGAN is employed to [...] Read more.
This study presents CORE-ReID V2, an enhanced framework built upon CORE-ReID V1. The new framework extends its predecessor by addressing unsupervised domain adaptation (UDA) challenges in person ReID and vehicle ReID, with further applicability to object ReID. During pre-training, CycleGAN is employed to synthesize diverse data, bridging image characteristic gaps across different domains. In the fine-tuning, an advanced ensemble fusion mechanism, consisting of the Efficient Channel Attention Block (ECAB) and the Simplified Efficient Channel Attention Block (SECAB), enhances both local and global feature representations while reducing ambiguity in pseudo-labels for target samples. Experimental results on widely used UDA person ReID and vehicle ReID datasets demonstrate that the proposed framework outperforms state-of-the-art methods, achieving top performance in mean average precision (mAP) and Rank-k Accuracy (Top-1, Top-5, Top-10). Moreover, the framework supports lightweight backbones such as ResNet18 and ResNet34, ensuring both scalability and efficiency. Our work not only pushes the boundaries of UDA-based object ReID but also provides a solid foundation for further research and advancements in this domain. Full article
Show Figures

Figure 1

25 pages, 1524 KB  
Article
Detecting Emerging DGA Malware in Federated Environments via Variational Autoencoder-Based Clustering and Resource-Aware Client Selection
by Ma Viet Duc, Pham Minh Dang, Tran Thu Phuong, Truong Duc Truong, Vu Hai and Nguyen Huu Thanh
Future Internet 2025, 17(7), 299; https://doi.org/10.3390/fi17070299 - 3 Jul 2025
Viewed by 545
Abstract
Domain Generation Algorithms (DGAs) remain a persistent technique used by modern malware to establish stealthy command-and-control (C&C) channels, thereby evading traditional blacklist-based defenses. Detecting such evolving threats is especially challenging in decentralized environments where raw traffic data cannot be aggregated due to privacy [...] Read more.
Domain Generation Algorithms (DGAs) remain a persistent technique used by modern malware to establish stealthy command-and-control (C&C) channels, thereby evading traditional blacklist-based defenses. Detecting such evolving threats is especially challenging in decentralized environments where raw traffic data cannot be aggregated due to privacy or policy constraints. To address this, we present FedSAGE, a security-aware federated intrusion detection framework that combines Variational Autoencoder (VAE)-based latent representation learning with unsupervised clustering and resource-efficient client selection. Each client encodes its local domain traffic into a semantic latent space using a shared, pre-trained VAE trained solely on benign domains. These embeddings are clustered via affinity propagation to group clients with similar data distributions and identify outliers indicative of novel threats without requiring any labeled DGA samples. Within each cluster, FedSAGE selects only the fastest clients for training, balancing computational constraints with threat visibility. Experimental results from the multi-zones DGA dataset show that FedSAGE improves detection accuracy by up to 11.6% and reduces energy consumption by up to 93.8% compared to standard FedAvg under non-IID conditions. Notably, the latent clustering perfectly recovers ground-truth DGA family zones, enabling effective anomaly detection in a fully unsupervised manner while remaining privacy-preserving. These foundations demonstrate that FedSAGE is a practical and lightweight approach for decentralized detection of evasive malware, offering a viable solution for secure and adaptive defense in resource-constrained edge environments. Full article
(This article belongs to the Special Issue Security of Computer System and Network)
Show Figures

Figure 1

22 pages, 580 KB  
Article
A Comparative Study of Advanced Transformer Learning Frameworks for Water Potability Analysis Using Physicochemical Parameters
by Enes Algül, Saadin Oyucu, Onur Polat, Hüseyin Çelik, Süleyman Ekşi, Faruk Kurker and Ahmet Aksoz
Appl. Sci. 2025, 15(13), 7262; https://doi.org/10.3390/app15137262 - 27 Jun 2025
Viewed by 3079
Abstract
Keeping drinking water safe is a critical aspect of protecting public health. Traditional laboratory-based methods for evaluating water potability are often time-consuming, costly, and labour-intensive. This paper presents a comparative analysis of four transformer-based deep learning models in the development of automatic classification [...] Read more.
Keeping drinking water safe is a critical aspect of protecting public health. Traditional laboratory-based methods for evaluating water potability are often time-consuming, costly, and labour-intensive. This paper presents a comparative analysis of four transformer-based deep learning models in the development of automatic classification systems for water potability based on physicochemical attributes. The models examined include the enhanced tabular transformer (ETT), feature tokenizer transformer (FTTransformer), self-attention and inter-sample network (SAINT), and tabular autoencoder pretraining enhancement (TAPE). The study utilized an open-access water quality dataset that includes nine key attributes such as pH, hardness, total dissolved solids (TDS), chloramines, sulphate, conductivity, organic carbon, trihalomethanes, and turbidity. The models were evaluated under a unified protocol involving 70–15–15 data partitioning, five-fold cross-validation, fixed random seed, and consistent hyperparameter settings. Among the evaluated models, the enhanced tabular transformer outperforms other models with an accuracy of 95.04% and an F1 score of 0.94. ETT is an advanced model because it can efficiently model high-order feature interactions through multi-head attention and deep hierarchical encoding. Feature importance analysis consistently highlighted chloramines, conductivity, and trihalomethanes as key predictive features across all models. SAINT demonstrated robust generalization through its dual-attention mechanism, while TAPE provided competitive results with reduced computational overhead due to unsupervised pretraining. Conversely, FTTransformer showed limitations, likely due to sensitivity to class imbalance and hyperparameter tuning. The results underscore the potential of transformer-based models, especially ETT, in enabling efficient, accurate, and scalable water quality monitoring. These findings support their integration into real-time environmental health systems and suggest approaches for future research in explainability, domain adaptation, and multimodal fusion. Full article
(This article belongs to the Special Issue Water Treatment: From Membrane Processes to Renewable Energies)
Show Figures

Figure 1

23 pages, 1475 KB  
Article
Large-Language-Model-Enabled Text Semantic Communication Systems
by Zhenyi Wang, Li Zou, Shengyun Wei, Kai Li, Feifan Liao, Haibo Mi and Rongxuan Lai
Appl. Sci. 2025, 15(13), 7227; https://doi.org/10.3390/app15137227 - 26 Jun 2025
Viewed by 1140
Abstract
Large language models (LLMs) have recently demonstrated state-of-the-art performance in various natural language processing (NLP) tasks, achieving near-human levels in multiple language understanding challenges and aligning closely with the core principles of semantic communication Inspired by LLMs’ advancements in semantic processing, we propose [...] Read more.
Large language models (LLMs) have recently demonstrated state-of-the-art performance in various natural language processing (NLP) tasks, achieving near-human levels in multiple language understanding challenges and aligning closely with the core principles of semantic communication Inspired by LLMs’ advancements in semantic processing, we propose LLM-SC, an innovative LLM-enabled semantic communication system framework which applies LLMs directly to the physical layer coding and decoding for the first time. By analyzing the relationship between the training process of LLMs and the optimization objectives of semantic communication, we propose training a semantic encoder through LLMs’ tokenizer training and establishing a semantic knowledge base via the LLMs’ unsupervised pre-training process. This knowledge base facilitates the creation of optimal decoder by providing the prior probability of the transmitted language sequence. Based on this, we derive the optimal decoding criteria for the receiver and introduce beam search algorithm to further reduce complexity. Furthermore, we assert that existing LLMs can be employed directly for LLM-SC without extra re-training or fine-tuning. Simulation results reveal that LLM-SC outperforms conventional DeepSC at signal-to-noise ratios (SNRs) exceeding 3 dB, as it enables error-free transmissions of semantic information under high SNRs while DeepSC fails to do so. In addition to semantic-level performance, LLM-SC demonstrates compatibility with technical-level performance, achieving approximately an 8 dB coding gain for a bit error ratio (BER) of 103 without any channel coding while maintaining the same joint source–channel coding rate as traditional communication systems. Full article
(This article belongs to the Special Issue Recent Advances in AI-Enabled Wireless Communications and Networks)
Show Figures

Figure 1

37 pages, 3049 KB  
Article
English-Arabic Hybrid Semantic Text Chunking Based on Fine-Tuning BERT
by Mai Alammar, Khalil El Hindi and Hend Al-Khalifa
Computation 2025, 13(6), 151; https://doi.org/10.3390/computation13060151 - 16 Jun 2025
Cited by 1 | Viewed by 1385
Abstract
Semantic text chunking refers to segmenting text into coherently semantic chunks, i.e., into sets of statements that are semantically related. Semantic chunking is an essential pre-processing step in various NLP tasks e.g., document summarization, sentiment analysis and question answering. In this paper, we [...] Read more.
Semantic text chunking refers to segmenting text into coherently semantic chunks, i.e., into sets of statements that are semantically related. Semantic chunking is an essential pre-processing step in various NLP tasks e.g., document summarization, sentiment analysis and question answering. In this paper, we propose a hybrid chunking; two-steps semantic text chunking method that combines the effectiveness of unsupervised semantic text chunking based on the similarities between sentences embeddings and the pre-trained language models (PLMs) especially BERT by fine-tuning the BERT on semantic textual similarity task (STS) to provide a flexible and effective semantic text chunking. We evaluated the proposed method in English and Arabic. To the best of our knowledge, there is an absence of an Arabic dataset created to assess semantic text chunking at this level. Therefore, we created an AraWiki50k to evaluate our proposed text chunking method inspired by an existing English dataset. Our experiments showed that exploiting the fine-tuned pre-trained BERT on STS enhances results over unsupervised semantic chunking by an average of 7.4 in the PK metric and by an average of 11.19 in the WindowDiff metric on four English evaluation datasets, and 0.12 in the PK and 2.29 in the WindowDiff for the Arabic dataset. Full article
(This article belongs to the Section Computational Social Science)
Show Figures

Figure 1

20 pages, 2150 KB  
Article
Industrial Image Anomaly Detection via Synthetic-Anomaly Contrastive Distillation
by Junxian Li, Mingxing Li, Shucheng Huang, Gang Wang and Xinjing Zhao
Sensors 2025, 25(12), 3721; https://doi.org/10.3390/s25123721 - 13 Jun 2025
Viewed by 910
Abstract
Industrial image anomaly detection plays a critical role in intelligent manufacturing by automatically identifying defective products through visual inspection. While unsupervised approaches eliminate dependency on annotated anomaly samples, current teacher–student framework-based methods still face two fundamental limitations: insufficient discriminative capability for structural anomalies [...] Read more.
Industrial image anomaly detection plays a critical role in intelligent manufacturing by automatically identifying defective products through visual inspection. While unsupervised approaches eliminate dependency on annotated anomaly samples, current teacher–student framework-based methods still face two fundamental limitations: insufficient discriminative capability for structural anomalies and suboptimal anomaly feature decoupling efficiency. To address these challenges, we propose a Synthetic-Anomaly Contrastive Distillation (SACD) framework for industrial anomaly detection. SACD comprises two pivotal components: (1) a reverse distillation (RD) paradigm whereby a pre-trained teacher network extracts hierarchically structured representations, subsequently guiding the student network with inverse architectural configuration to establish hierarchical feature alignment; (2) a group of feature calibration (FeaCali) modules designed to refine the student’s outputs by eliminating anomalous feature responses. During training, SACD adopts a dual-branch strategy, where one branch encodes multi-scale features from defect-free images, while a Siamese anomaly branch processes synthetically corrupted counterparts. FeaCali modules are trained to strip out a student’s anomalous patterns in anomaly branches, enhancing the student network’s exclusive modeling of normal patterns. We construct a dual-objective optimization integrating cross-model distillation loss and intra-model contrastive loss to train SACD for feature alignment and discrepancy amplification. At the inference stage, pixel-wise anomaly scores are computed through multi-layer feature discrepancies between the teacher’s representations and the student’s refined outputs. Comprehensive evaluations on the MVTec AD and BTAD benchmark demonstrate that our method is effective and superior to current knowledge distillation-based approaches. Full article
(This article belongs to the Section Industrial Sensors)
Show Figures

Figure 1

15 pages, 7136 KB  
Article
Source-Free Domain Adaptation for Cross-Modality Abdominal Multi-Organ Segmentation Challenges
by Xiyu Zhang, Xu Chen, Yang Wang, Dongliang Liu and Yifeng Hong
Information 2025, 16(6), 460; https://doi.org/10.3390/info16060460 - 29 May 2025
Viewed by 545
Abstract
Abdominal organ segmentation in CT images is crucial for accurate diagnosis, treatment planning, and condition monitoring. However, the annotation process is often hindered by challenges such as low contrast, artifacts, and complex organ structures. While unsupervised domain adaptation (UDA) has shown promise in [...] Read more.
Abdominal organ segmentation in CT images is crucial for accurate diagnosis, treatment planning, and condition monitoring. However, the annotation process is often hindered by challenges such as low contrast, artifacts, and complex organ structures. While unsupervised domain adaptation (UDA) has shown promise in addressing these issues by transferring knowledge from a different modality (source domain), its reliance on both source and target data during training presents a practical challenge in many clinical settings due to data privacy concerns. This study aims to develop a cross-modality abdominal multi-organ segmentation model for label-free CT (target domain) data, leveraging knowledge solely from a pre-trained source domain (MRI) model without accessing the source data. To achieve this, we generate source-like images from target-domain images using a one-way image translation approach with the pre-trained model. These synthesized images preserve the anatomical structure of the target, enabling segmentation predictions from the pre-trained model. To further enhance segmentation accuracy, particularly for organ boundaries and small contours, we introduce an auxiliary translation module with an image decoder and multi-level discriminator. The results demonstrate significant improvements across several performance metrics, including the Dice similarity coefficient (DSC) and average symmetric surface distance (ASSD), highlighting the effectiveness of the proposed method. Full article
Show Figures

Figure 1

28 pages, 1007 KB  
Article
Predicting the Event Types in the Human Brain: A Modeling Study Based on Embedding Vectors and Large-Scale Situation Type Datasets in Mandarin Chinese
by Xiaorui Ma and Hongchao Liu
Appl. Sci. 2025, 15(11), 5916; https://doi.org/10.3390/app15115916 - 24 May 2025
Viewed by 490
Abstract
Event types classify Chinese verbs based on the internal temporal structure of events. The categorization of verb event types is the most fundamental classification of concept types represented by verbs in the human brain. Meanwhile, event types exhibit strong predictive capabilities for exploring [...] Read more.
Event types classify Chinese verbs based on the internal temporal structure of events. The categorization of verb event types is the most fundamental classification of concept types represented by verbs in the human brain. Meanwhile, event types exhibit strong predictive capabilities for exploring collocational patterns between words, making them crucial for Chinese teaching. This work focuses on constructing a statistically validated gold-standard dataset, forming the foundation for achieving high accuracy in recognizing verb event types. Utilizing a manually annotated dataset of verbs and aspectual markers’ co-occurrence features, the research conducts hierarchical clustering of Chinese verbs. The resulting dendrogram indicates that verbs can be categorized into three event types—state, activity and transition—based on semantic distance. Two approaches are employed to construct vector matrices: a supervised method that derives word vectors based on linguistic features, and an unsupervised method that uses four models to extract embedding vectors, including Word2Vec, FastText, BERT and ChatGPT. The classification of verb event types is performed using three classifiers: multinomial logistic regression, support vector machines and artificial neural networks. Experimental results demonstrate the superior performance of embedding vectors. Employing the pre-trained FastText model in conjunction with an artificial neural network classifier, the model achieves an accuracy of 98.37% in predicting 3133 verbs, thereby enabling the automatic identification of event types at the level of Chinese verbs and validating the high accuracy and practical value of embedding vectors in addressing complex semantic relationships and classification tasks. This work constructs datasets of considerable semantic complexity, comprising a substantial volume of verbs along with their feature vectors and situation type labels, which can be used for evaluating large language models in the future. Full article
(This article belongs to the Special Issue Application of Artificial Intelligence and Semantic Mining Technology)
Show Figures

Figure 1

13 pages, 8618 KB  
Article
Automated Detection of Canine Babesia Parasite in Blood Smear Images Using Deep Learning and Contrastive Learning Techniques
by Dilip Kumar Baruah, Kuntala Boruah, Nagendra Nath Barman, Abhijit Deka, Arpita Bharali and Lukumoni Buragohain
Parasitologia 2025, 5(2), 23; https://doi.org/10.3390/parasitologia5020023 - 14 May 2025
Viewed by 663
Abstract
This research introduces a novel method that integrates both unsupervised and supervised learning, leveraging SimCLR (Simple Framework for Contrastive Learning of Visual Representations) for self-supervised learning along with different pre-trained models to improve microscopic image classification of Babesia parasite in canines. We focused [...] Read more.
This research introduces a novel method that integrates both unsupervised and supervised learning, leveraging SimCLR (Simple Framework for Contrastive Learning of Visual Representations) for self-supervised learning along with different pre-trained models to improve microscopic image classification of Babesia parasite in canines. We focused on three popular CNN architectures, namely ResNet, EfficientNet, and DenseNet, and evaluated the impact of SimCLR pre-training on their performance. A detailed comparison of the different variants of ResNet, EfficientNet, and Densenet in terms of classification accuracy and training efficiency is presented. Base models such as different variants of the ResNet, EfficientNet, and DenseNet models were utilized within the SimCLR framework. Firstly, the models were pre-trained on unlabeled images, followed by training classifiers on labeled datasets. This approach significantly improved the robustness and accuracy, demonstrating the potential benefits of combining contrastive learning with conventional supervised techniques. The highest accuracy of 97.07% was achieved by Efficientnet_b2. Thus, detection of Babesia or other hemoparasites in microscopic blood smear images could be automated with high accuracy without using a labelled dataset. Full article
Show Figures

Figure 1

24 pages, 3113 KB  
Article
Gradual Geometry-Guided Knowledge Distillation for Source-Data-Free Domain Adaptation
by Yangkuiyi Zhang and Song Tang
Mathematics 2025, 13(9), 1491; https://doi.org/10.3390/math13091491 - 30 Apr 2025
Viewed by 533
Abstract
Due to access to the source data during the transfer phase, conventional domain adaptation works have recently raised safety and privacy concerns. More research attention thus shifts to a more practical setting known as source-data-free domain adaptation (SFDA). The new challenge is how [...] Read more.
Due to access to the source data during the transfer phase, conventional domain adaptation works have recently raised safety and privacy concerns. More research attention thus shifts to a more practical setting known as source-data-free domain adaptation (SFDA). The new challenge is how to obtain reliable semantic supervision in the absence of source domain training data and the labels on the target domain. To that end, in this work, we introduce a novel Gradual Geometry-Guided Knowledge Distillation (G2KD) approach for SFDA. Specifically, to address the lack of supervision, we used local geometry of data to construct a more credible probability distribution over the potential categories, termed geometry-guided knowledge. Then, knowledge distillation was adopted to integrate this extra information for boosting the adaptation. More specifically, first, we constructed a neighborhood geometry for any target data using a similarity comparison on the whole target dataset. Second, based on pre-obtained semantic estimation by clustering, we mined soft semantic representations expressing the geometry-guided knowledge by semantic fusion. Third, using the soften labels, we performed knowledge distillation regulated by the new objective. Considering the unsupervised setting of SFDA, in addition to the distillation loss and student loss, we introduced a mixed entropy regulator that minimized the entropy of individual data as well as maximized the mutual entropy with augmentation data to utilize neighbor relation. Our contribution is that, through local geometry discovery with semantic representation and self-knowledge distillation, the semantic information hidden in the local structures is transformed to effective semantic self-supervision. Also, our knowledge distillation works in a gradual way that is helpful to capture the dynamic variations in the local geometry, mitigating the previous guidance degradation and deviation at the same time. Extensive experiments on five challenging benchmarks confirmed the state-of-the-art performance of our method. Full article
(This article belongs to the Special Issue Robust Perception and Control in Prognostic Systems)
Show Figures

Figure 1

Back to TopTop