Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (1,007)

Search Parameters:
Keywords = label-specific features

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 3049 KB  
Article
Tri-Invariance Contrastive Framework for Robust Unsupervised Person Re-Identification
by Lei Wang, Chengang Liu, Xiaoxiao Wang, Weidong Gao, Xuejian Ge and Shunjie Zhu
Mathematics 2025, 13(21), 3570; https://doi.org/10.3390/math13213570 (registering DOI) - 6 Nov 2025
Abstract
Unsupervised person re-identification (Re-ID) has been proven very effective and it boosts the performance in learning representations from unlabeled data in the dataset. Most current methods have good accuracy, but there are two main problems. First, clustering often generates noisy labels. Second, features [...] Read more.
Unsupervised person re-identification (Re-ID) has been proven very effective and it boosts the performance in learning representations from unlabeled data in the dataset. Most current methods have good accuracy, but there are two main problems. First, clustering often generates noisy labels. Second, features can change because of different camera styles. Noisy labels causes incorrect optimization, which reduces the accuracy of the model. The latter results in inaccurate prediction for samples within the same category that have been captured by different cameras. Despite the significant variations inherent in the vast source data, the principles of invariance and symmetry remain crucial for effective feature recognition. In this paper, we propose a method called Invariance Constraint Contrast Learning (ICCL) to address these two problems. Specifically, we introduce center invariance and instance invariance to reduce the effect of noisy samples. We also use camera invariance to handle feature changes caused by different cameras. Center invariance and instance invariance help decrease the impact of noise. Camera invariance improves the classification accuracy by using a camera-aware classification strategy. We test our method on three common large-scale Re-ID datasets. It clearly improves the accuracy of unsupervised person Re-ID. Specifically, our approach demonstrates its effectiveness by improving mAP by 3.5% on Market-1501, 1.3% on MSMT17 and 3.5% on CUHK03 over state-of-the-art methods. Full article
(This article belongs to the Special Issue Mathematical Computation for Pattern Recognition and Computer Vision)
17 pages, 1515 KB  
Article
CRTSC: Channel-Wise Recalibration and Texture-Structural Consistency Constraint for Anomaly Detection in Medical Chest Images
by Mingfu Xiong, Chong Wang, Hao Cai, Aziz Alotaibi, Saeed Anwar, Abdul Khader Jilani Saudagar, Javier Del Ser and Khan Muhammad
Sensors 2025, 25(21), 6722; https://doi.org/10.3390/s25216722 - 3 Nov 2025
Viewed by 215
Abstract
Unsupervised medical image anomaly detection, which does not need any labels, holds a pivotal role in early disease detection for advancing human intelligent health, and it is among the prominent research endeavors in the realm of biomedical image analysis. Existing deep model-based methods [...] Read more.
Unsupervised medical image anomaly detection, which does not need any labels, holds a pivotal role in early disease detection for advancing human intelligent health, and it is among the prominent research endeavors in the realm of biomedical image analysis. Existing deep model-based methods mainly focus on feature selection and interaction, ignoring the relative position and shape uncertainty of the anomalies themselves, which play an important guiding role in disease diagnosis, hampering performance. To address this issue, our study introduces a novel and effective framework, termed CRTSC, which integrates a channel-wise recalibration module (CRM) along with the texture–structural consistency constraint (TSCC) for anomaly detection in medical chest images acquired from different sensors. Specifically, the CRM adjusts the weight of different medical image feature channels, which are used to establish spatial relationships among anomalous patterns, enhancing the network’s representation and generalization capabilities. The texture–structural consistency constraint is devoted to enhancing the anomaly’s structural (shape) definiteness via evaluating the loss function of similarity between two images and optimizing the model. The two collaborate in an end-to-end fashion to optimize and train the entire framework, thereby enabling anomaly detection in medical chest images. Extensive experiments conducted on the public ZhangLab and CheXpert datasets demonstrate that our method achieves a significant performance improvement compared with the state-of-the-art methods, offering a robust and generalizable solution for sensor-based medical imaging applications. Full article
Show Figures

Figure 1

22 pages, 1208 KB  
Article
Geo-MRC: Dynamic Boundary Inference in Machine Reading Comprehension for Nested Geographic Named Entity Recognition
by Yuting Zhang, Jingzhong Li, Pengpeng Li, Tao Liu, Ping Du and Xuan Hao
ISPRS Int. J. Geo-Inf. 2025, 14(11), 431; https://doi.org/10.3390/ijgi14110431 - 2 Nov 2025
Viewed by 205
Abstract
Geographic Named Entity Recognition (Geo-NER) is a crucial task for extracting geography-related entities from unstructured text, and it plays an essential role in geographic information extraction and spatial semantic understanding. Traditional approaches typically treat Geo-NER as a sequence labeling problem, where each token [...] Read more.
Geographic Named Entity Recognition (Geo-NER) is a crucial task for extracting geography-related entities from unstructured text, and it plays an essential role in geographic information extraction and spatial semantic understanding. Traditional approaches typically treat Geo-NER as a sequence labeling problem, where each token is assigned a single label. However, this formulation struggles to handle nested entities effectively. To overcome this limitation, we propose Geo-MRC, an improved model based on a Machine Reading Comprehension (MRC) framework that reformulates Geo-NER as a question-answering task. The model identifies entities by predicting their start positions, end positions, and lengths, enabling precise detection of overlapping and nested entities. Specifically, it constructs a unified input sequence by concatenating a type-specific question (e.g., “What are the location names in the text?”) with the context. This sequence is encoded using BERT, followed by feature extraction and fusion through Gated Recurrent Units (GRU) and multi-scale 1D convolutions, which improve the model’s sensitivity to both multi-level semantics and local contextual information. Finally, a feed-forward neural network (FFN) predicts whether each token corresponds to the start or end of an entity and estimates the span length, allowing for dynamic inference of entity boundaries. Experimental results on multiple public datasets demonstrate that Geo-MRC consistently outperforms strong baselines, with particularly significant gains on datasets containing nested entities. Full article
Show Figures

Figure 1

28 pages, 1976 KB  
Article
ECG Signal Analysis and Abnormality Detection Application
by Ales Jandera, Yuliia Petryk, Martin Muzelak and Tomas Skovranek
Algorithms 2025, 18(11), 689; https://doi.org/10.3390/a18110689 - 29 Oct 2025
Viewed by 228
Abstract
The electrocardiogram (ECG) signal carries information crucial for health assessment, but its analysis can be challenging due to noise and signal variability; therefore, automated processing focused on noise removal and detection of key features is necessary. This paper introduces an ECG signal analysis [...] Read more.
The electrocardiogram (ECG) signal carries information crucial for health assessment, but its analysis can be challenging due to noise and signal variability; therefore, automated processing focused on noise removal and detection of key features is necessary. This paper introduces an ECG signal analysis and abnormality detection application developed to process single-lead ECG signals. In this study, the Lobachevsky University database (LUDB) was used as the source of ECG signals, as it includes annotated recordings using a multi-class, multi-label taxonomy that covers several diagnostic categories, each with specific diagnoses that reflect clinical ECG interpretation practices. The main aim of the paper is to provide a tool that efficiently filters noisy ECG data, accurately detects the QRS complex, PQ and QT intervals, calculates heart rate, and compares these values with normal ranges based on age and gender. Additionally, a multi-class, multi-label SVM-based model was developed and integrated into the application for heart abnormality diagnostics, i.e., assigning one or several diagnoses from various diagnostic categories. The MATLAB-based application is capable of processing raw ECG signals, allowing the use of ECG records not only from LUDB but also from other databases. Full article
(This article belongs to the Special Issue Algorithms for Computer Aided Diagnosis: 2nd Edition)
Show Figures

Figure 1

37 pages, 36320 KB  
Article
PLISA: An Optical–SAR Remote Sensing Image Registration Method Based on Pseudo-Label Learning and Interactive Spatial Attention
by Yixuan Zhang, Ruiqi Liu, Zeyu Zhang, Limin Shi, Lubin Weng and Lei Hu
Remote Sens. 2025, 17(21), 3571; https://doi.org/10.3390/rs17213571 - 28 Oct 2025
Viewed by 322
Abstract
Multimodal remote sensing image registration faces severe challenges due to geometric and radiometric differences, particularly between optical and synthetic aperture radar (SAR) images. These inherent disparities make extracting highly repeatable cross-modal feature points difficult. Current methods typically rely on image intensity extreme responses [...] Read more.
Multimodal remote sensing image registration faces severe challenges due to geometric and radiometric differences, particularly between optical and synthetic aperture radar (SAR) images. These inherent disparities make extracting highly repeatable cross-modal feature points difficult. Current methods typically rely on image intensity extreme responses or network regression without keypoint supervision for feature point detection. Moreover, they not only lack explicit keypoint annotations as supervision signals but also fail to establish a clear and consistent definition of what constitutes a reliable feature point in cross-modal scenarios. To overcome this limitation, we propose PLISA—a novel heterogeneous image registration method. PLISA integrates two core components: an automated pseudo-labeling module (APLM) and a pseudo-twin interaction network (PTIF). The APLM introduces an innovative labeling strategy that explicitly defines keypoints as corner points, thereby generating consistent pseudo-labels for dual-modality images and effectively mitigating the instability caused by the absence of supervised keypoint annotations. These pseudo-labels subsequently train the PTIF, which adopts a pseudo-twin architecture incorporating a cross-modal interactive attention (CIA) module to effectively reconcile cross-modal commonalities and distinctive characteristics. Evaluations on the SEN1-2 dataset and OSdataset demonstrate PLISA’s state-of-the-art cross-modal feature point repeatability while maintaining robust registration accuracy across a range of challenging conditions, including rotations, scale variations, and SAR-specific speckle noise. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

31 pages, 34773 KB  
Article
Learning Domain-Invariant Representations for Event-Based Motion Segmentation: An Unsupervised Domain Adaptation Approach
by Mohammed Jeryo and Ahad Harati
J. Imaging 2025, 11(11), 377; https://doi.org/10.3390/jimaging11110377 - 27 Oct 2025
Viewed by 282
Abstract
Event cameras provide microsecond temporal resolution, high dynamic range, and low latency by asynchronously capturing per-pixel luminance changes, thereby introducing a novel sensing paradigm. These advantages render them well-suited for high-speed applications such as autonomous vehicles and dynamic environments. Nevertheless, the sparsity of [...] Read more.
Event cameras provide microsecond temporal resolution, high dynamic range, and low latency by asynchronously capturing per-pixel luminance changes, thereby introducing a novel sensing paradigm. These advantages render them well-suited for high-speed applications such as autonomous vehicles and dynamic environments. Nevertheless, the sparsity of event data and the absence of dense annotations are significant obstacles to supervised learning for motion segmentation from event streams. Domain adaptation is also challenging due to the considerable domain shift in intensity images. To address these challenges, we propose a two-phase cross-modality adaptation framework that translates motion segmentation knowledge from labeled RGB-flow data to unlabeled event streams. A dual-branch encoder extracts modality-specific motion and appearance features from RGB and optical flow in the source domain. Using reconstruction networks, event voxel grids are converted into pseudo-image and pseudo-flow modalities in the target domain. These modalities are subsequently re-encoded using frozen RGB-trained encoders. Multi-level consistency losses are implemented on features, predictions, and outputs to enforce domain alignment. Our design enables the model to acquire domain-invariant, semantically rich features through the use of shallow architectures, thereby reducing training costs and facilitating real-time inference with a lightweight prediction path. The proposed architecture, alongside the utilized hybrid loss function, effectively bridges the domain and modality gap. We evaluate our method on two challenging benchmarks: EVIMO2, which incorporates real-world dynamics, high-speed motion, illumination variation, and multiple independently moving objects; and MOD++, which features complex object dynamics, collisions, and dense 1kHz supervision in synthetic scenes. The proposed UDA framework achieves 83.1% and 79.4% accuracy on EVIMO2 and MOD++, respectively, outperforming existing state-of-the-art approaches, such as EV-Transfer and SHOT, by up to 3.6%. Additionally, it is lighter and faster and also delivers enhanced mIoU and F1 Score. Full article
(This article belongs to the Section Image and Video Processing)
Show Figures

Figure 1

17 pages, 38319 KB  
Article
Class-Level Feature Disentanglement for Multi-Label Image Classification
by Yingduo Tong, Zhenyu Lu, Yize Dong and Yonggang Lu
Future Internet 2025, 17(11), 486; https://doi.org/10.3390/fi17110486 - 23 Oct 2025
Viewed by 219
Abstract
Generally, the interpretability of deep neural networks is categorized into a priori and a posteriori interpretability. A priori interpretability involves improving model transparency through deliberate design prior to training. Feature disentanglement is a method for achieving a priori interpretability. Existing disentanglement methods mostly [...] Read more.
Generally, the interpretability of deep neural networks is categorized into a priori and a posteriori interpretability. A priori interpretability involves improving model transparency through deliberate design prior to training. Feature disentanglement is a method for achieving a priori interpretability. Existing disentanglement methods mostly focus on semantic features, such as intrinsic and shared features. These methods distinguish between the background and the main subject, but overlook class-level features in images. To address this, we take a further step by advancing feature disentanglement to the class level. For multi-label image classification tasks, we propose a class-level feature disentanglement method. Specifically, we introduce a multi-head classifier within the feature extraction layer of the backbone network to disentangle features. Each head in this classifier corresponds to a specific class and generates independent predictions, thereby guiding the model to better leverage the intrinsic features of each class and improving multi-label classification precision. Experiments demonstrate that our method significantly enhances performance metrics across various benchmarks while simultaneously achieving a priori interpretability. Full article
(This article belongs to the Special Issue Machine Learning Techniques for Computer Vision)
Show Figures

Figure 1

17 pages, 1775 KB  
Article
AI-Driven Analysis for Real-Time Detection of Unstained Microscopic Cell Culture Images
by Kathrin Hildebrand, Tatiana Mögele, Dennis Raith, Maria Kling, Anna Rubeck, Stefan Schiele, Eelco Meerdink, Avani Sapre, Jonas Bermeitinger, Martin Trepel and Rainer Claus
AI 2025, 6(10), 271; https://doi.org/10.3390/ai6100271 - 18 Oct 2025
Viewed by 608
Abstract
Staining-based assays are widely used for cell analysis but are invasive, alter physiology, and prevent longitudinal monitoring. Label-free, morphology-based approaches could enable real-time, non-invasive drug testing, yet detection of subtle and dynamic changes has remained difficult. We developed a deep learning framework for [...] Read more.
Staining-based assays are widely used for cell analysis but are invasive, alter physiology, and prevent longitudinal monitoring. Label-free, morphology-based approaches could enable real-time, non-invasive drug testing, yet detection of subtle and dynamic changes has remained difficult. We developed a deep learning framework for stain-free monitoring of leukemia cell cultures using automated bright-field microscopy in a semi-automated culture system (AICE3, LABMaiTE, Augsburg, Germany). YOLOv8 models were trained on images from K562, HL-60, and Kasumi-1 cells, using an NVIDIA DGX A100 GPU for training and tested on GPU and CPU environments for real-time performance. Comparative benchmarking with RT-DETR and interpretability analyses using Eigen-CAM and radiomics (RedTell) was performed. YOLOv8 achieved high accuracy (mAP@0.5 > 98%, precision/sensitivity > 97%), with reproducibility confirmed on an independent dataset from a second laboratory and an AICE3 setup. The model distinguished between morphologically similar leukemia lines and reliably classified untreated versus differentiated K562 cells (hemin-induced erythroid and PMA-induced megakaryocytic; >95% accuracy). Incorporation of decitabine-treated cells demonstrated applicability to drug testing, revealing treatment-specific and intermediate phenotypes. Longitudinal monitoring captured culture- and time-dependent drift, enabling separation of temporal from drug-induced changes. Radiomics highlighted interpretable features such as size, elongation, and texture, but with lower accuracy than the deep learning approach. To our knowledge, this is the first demonstration that deep learning resolves subtle, drug-induced, and time-dependent morphological changes in unstained leukemia cells in real time. This approach provides a robust, accessible framework for label-free longitudinal drug testing and establishes a foundation for future autonomous, feedback-driven platforms in precision oncology. Ultimately, this approach may also contribute to more precise and adaptive clinical decision-making, advancing the field of personalized medicine. Full article
(This article belongs to the Special Issue AI in Bio and Healthcare Informatics)
Show Figures

Figure 1

20 pages, 11103 KB  
Data Descriptor
VitralColor-12: A Synthetic Twelve-Color Segmentation Dataset from GPT-Generated Stained-Glass Images
by Martín Montes Rivera, Carlos Guerrero-Mendez, Daniela Lopez-Betancur, Tonatiuh Saucedo-Anaya, Manuel Sánchez-Cárdenas and Salvador Gómez-Jiménez
Data 2025, 10(10), 165; https://doi.org/10.3390/data10100165 - 18 Oct 2025
Viewed by 413
Abstract
The segmentation and classification of color are crucial stages in image processing, computer vision, and pattern recognition, as they significantly impact the results. The diverse, hand-labeled datasets in the literature are applied for monochromatic or color segmentation in specific domains. On the other [...] Read more.
The segmentation and classification of color are crucial stages in image processing, computer vision, and pattern recognition, as they significantly impact the results. The diverse, hand-labeled datasets in the literature are applied for monochromatic or color segmentation in specific domains. On the other hand, synthetic datasets are generated using statistics, artificial intelligence algorithms, or generative artificial intelligence (AI). This last one includes Large Language Models (LLMs), Generative Adversarial Neural Networks (GANs), and Variational Autoencoders (VAEs), among others. In this work, we propose VitralColor-12, a synthetic dataset for color classification and segmentation, comprising twelve colors: black, blue, brown, cyan, gray, green, orange, pink, purple, red, white, and yellow. VitralColor-12 addresses the limitations of color segmentation and classification datasets by leveraging the capabilities of LLMs, including adaptability, variability, copyright-free content, and lower-cost data—properties that are desirable in image datasets. VitralColor-12 includes pixel-level classification and segmentation maps. This makes the dataset broadly applicable and highly variable for a range of computer vision applications. VitralColor-12 utilizes GPT-5 and DALL·E 3 for generating stained-glass images. These images simplify the annotation process, since stained-glass images have isolated colors with distinct boundaries within the steel structure, which provide easy regions to label with a single color per region. Once we obtain the images, we use at least one hand-labeled centroid per color to automatically cluster all pixels based on Euclidean distance and morphological operations, including erosion and dilation. This process enables us to automatically label a classification dataset and generate segmentation maps. Our dataset comprises 910 images, organized into 70 generated images and 12 pixel segmentation maps—one for each color—which include 9,509,524 labeled pixels, 1,794,758 of which are unique. These annotated pixels are represented by RGB, HSL, CIELAB, and YCbCr values, enabling a detailed color analysis. Moreover, VitralColor-12 offers features that address gaps in public resources such as violin diagrams with the frequency of colors across images, histograms of channels per color, 3D color maps, descriptive statistics, and standardized metrics, such as ΔE76, ΔE94, and CIELAB Chromacity, which prove the distribution, applicability, and realistic perceptual structures, including warm, neutral, and cold colors, as well as the high contrast between black and white colors, offering meaningful perceptual clusters, reinforcing its utility for color segmentation and classification. Full article
Show Figures

Figure 1

16 pages, 3378 KB  
Article
Cosine Prompt-Based Class Incremental Semantic Segmentation for Point Clouds
by Lei Guo, Hongye Li, Min Pang, Kaowei Liu, Xie Han and Fengguang Xiong
Algorithms 2025, 18(10), 648; https://doi.org/10.3390/a18100648 - 16 Oct 2025
Viewed by 267
Abstract
Although current 3D semantic segmentation methods have achieved significant success, they suffer from catastrophic forgetting when confronted with dynamic, open environments. To address this issue, class incremental learning is introduced to update models while maintaining a balance between plasticity and stability. In this [...] Read more.
Although current 3D semantic segmentation methods have achieved significant success, they suffer from catastrophic forgetting when confronted with dynamic, open environments. To address this issue, class incremental learning is introduced to update models while maintaining a balance between plasticity and stability. In this work, we propose CosPrompt, a rehearsal-free approach for class incremental semantic segmentation. Specifically, we freeze the prompts for existing classes and incrementally expand and fine-tune the prompts for new classes, thereby generating discriminative and customized features. We employ clamping operations to regulate backward propagation, ensuring smooth training. Furthermore, we utilize the learning without forgetting loss and pseudo-label generation to further mitigate catastrophic forgetting. We conduct comparative and ablation experiments on the S3DIS dataset and ScanNet v2 dataset, demonstrating the effectiveness and feasibility of our method. Full article
(This article belongs to the Section Randomized, Online, and Approximation Algorithms)
Show Figures

Figure 1

18 pages, 3535 KB  
Article
UAV Based Weed Pressure Detection Through Relative Labelling
by Sebastiaan Verbesselt, Rembert Daems, Axel Willekens and Jonathan Van Beek
Remote Sens. 2025, 17(20), 3434; https://doi.org/10.3390/rs17203434 - 15 Oct 2025
Viewed by 429
Abstract
Agricultural management in Europe faces increasing pressure to reduce its environmental footprint. Implementing precision agriculture for weed management could offer a solution and minimize the use of chemical products. High spatial resolution imagery from real time kinematic (RTK) unmanned aerial vehicles (UAV) in [...] Read more.
Agricultural management in Europe faces increasing pressure to reduce its environmental footprint. Implementing precision agriculture for weed management could offer a solution and minimize the use of chemical products. High spatial resolution imagery from real time kinematic (RTK) unmanned aerial vehicles (UAV) in combination with supervised convolutional neural network (CNNs) models have proven successful in making location specific treatments. This site-specific advice limits the amount of herbicide applied to the field to areas that require action, thereby reducing the environmental impact and inputs for the farmer. To develop performant CNN models, there is a need for sufficient high-quality labelled data. To reduce the labelling effort and time, a new labelling method is proposed whereby image subsection pairs are labelled based on their relative differences in weed pressure to train a CNN ordinal regression model. The model is evaluated on detecting weed pressure in potato (Solanum tuberosum L.). Model performance was evaluated on different levels: pairwise accuracy, linearity (Pearson correlation coefficient), rank consistency (Spearman’s (rs) and Kendal (τ) rank correlations coefficients) and binary accuracy. After hyperparameter tuning, a pairwise accuracy of 85.2%, significant linearity (rs = 0.81) and significant rank consistency (rs = 0.87 and τ = 0.69) were found. This suggests that the model is capable of correctly detecting the gradient in weed pressure for the dataset. A maximum binary accuracy and F1-score of 92% and 88% were found for the dataset after thresholding the predicted weed scores into weed versus non-weed images. The model architecture allows us to visualize the intermediate features of the last convolutional block. This allows data analysts to better evaluate if the model “sees” the features of interest (in this case weeds). The results indicate the potential of ordinal regression with relative labels as a fast, lightweight model that predicts weed pressure gradients. Experts have the freedom to decide which threshold value(s) can be used on predicted weed scores depending on the weed, crop and treatment that they want to use for flexible weed control management. Full article
Show Figures

Figure 1

24 pages, 9046 KB  
Article
Novel Multimodal Imaging System for High-Resolution and High-Contrast Tissue Segmentation Based on Chemical Properties
by Björn van Marwick, Felix Lauer, Felix Wühler, Miriam Rittel, Carmen Wängler, Björn Wängler, Carsten Hopf and Matthias Rädle
Sensors 2025, 25(20), 6342; https://doi.org/10.3390/s25206342 - 14 Oct 2025
Viewed by 762
Abstract
Accurate and detailed tissue characterization is a central goal in medical diagnostics, often requiring the combination of multiple imaging modalities. This study presents a multimodal imaging system that integrates mid-infrared (MIR) scanning with fluorescence imaging to enhance the chemical specificity and spatial resolution [...] Read more.
Accurate and detailed tissue characterization is a central goal in medical diagnostics, often requiring the combination of multiple imaging modalities. This study presents a multimodal imaging system that integrates mid-infrared (MIR) scanning with fluorescence imaging to enhance the chemical specificity and spatial resolution in biological samples. A motorized mirror allows rapid switching between MIR and fluorescence modes, enabling efficient, co-registered data acquisition. The MIR modality captures label-free chemical maps based on molecular vibrations, while the fluorescence channel records endogenous autofluorescence for additional biochemical contrast. Applied to mouse brain tissue, the system enabled the clear differentiation of gray matter and white matter, supported by the clustering analysis of spectral features. The addition of autofluorescence imaging further improved anatomical segmentation and revealed fine structural details. In mouse skin, the approach allowed the precise mapping of the layered tissue architecture. These results demonstrate that combining MIR scanning and fluorescence imaging provides complementary, label-free insights into tissue morphology and chemistry. The findings support the utility of this approach as a powerful tool for biomedical research and diagnostic applications, offering a more comprehensive understanding of tissue composition without relying on staining or external markers. Full article
(This article belongs to the Section Biomedical Sensors)
Show Figures

Figure 1

22 pages, 1406 KB  
Article
A GIS-Integrated Framework for Unsupervised Fuzzy Classification of Residential Building Pattern
by Rosa Cafaro, Barbara Cardone, Valeria D’Ambrosio, Ferdinando Di Martino and Vittorio Miraglia
Electronics 2025, 14(20), 4022; https://doi.org/10.3390/electronics14204022 - 14 Oct 2025
Viewed by 254
Abstract
The classification of urban residential settlements through Machine Learning (ML) and Deep Learning (DL) remains a complex task due to the intrinsic heterogeneity of urban environments and the scarcity of large, accurately labeled training datasets. To overcome these limitations, this study introduces a [...] Read more.
The classification of urban residential settlements through Machine Learning (ML) and Deep Learning (DL) remains a complex task due to the intrinsic heterogeneity of urban environments and the scarcity of large, accurately labeled training datasets. To overcome these limitations, this study introduces a novel GIS-based unsupervised classification framework that exploits Fuzzy C-Means (FCM) clustering for the detection and interpretation of urban morphologies. Compared to unsupervised classification approaches that rely on crisp-based clustering algorithms, the proposed FCM-based method more effectively captures heterogeneous urban fabrics where no clear predominance of specific building types exists. Specifically, the method applies fuzzy clustering to census units—considered the fundamental scale of urban analysis—based on construction techniques and building periods. By grouping census areas with similar structural features, the framework provides a flexible, data-driven approach to the characterization of urban settlements. The identification of cluster centroids’ dominant attributes enables a systematic interpretation of the spatial distribution of the built environment, while the subsequent mapping process assigns each cluster a descriptive label reflecting the prevailing building fabric. The generated thematic maps yield critical insights into urban morphology and facilitate evidence-based planning. The framework was validated across ten Italian cities selected for their diverse physical, morphological, and historical characteristics; comparisons with the results of urban zone classifications in these cities conducted by experts show that the proposed method provides accurate results, as the similarity to the classifications made by experts, measured by the use of the Adjusted Rand Index, is always higher than or equal to 0.93; furthermore, it is robust when applied in heterogeneous urban settlements. These results confirm the effectiveness of the method in delineating homogeneous urban areas, thereby offering decision makers a robust instrument to guide targeted interventions on existing building stocks. The proposed framework advances the capacity to analyze urban form, to strategically support renovation and urban regeneration policies, and demonstrates a strong potential for portability, as it can be applied to other cities for urban scale analyses. Full article
(This article belongs to the Special Issue Advances in Algorithm Optimization and Computational Intelligence)
Show Figures

Figure 1

24 pages, 9099 KB  
Article
Dynamic MAML with Efficient Multi-Scale Attention for Cross-Load Few-Shot Bearing Fault Diagnosis
by Qinglei Zhang, Yifan Zhang, Jiyun Qin, Jianguo Duan and Ying Zhou
Entropy 2025, 27(10), 1063; https://doi.org/10.3390/e27101063 - 14 Oct 2025
Viewed by 403
Abstract
Accurate bearing fault diagnosis under various operational conditions presents significant challenges, mainly due to the limited availability of labeled data and the domain mismatches across different operating environments. In this study, an adaptive meta-learning framework (AdaMETA) is proposed, which combines dynamic task-aware model-independent [...] Read more.
Accurate bearing fault diagnosis under various operational conditions presents significant challenges, mainly due to the limited availability of labeled data and the domain mismatches across different operating environments. In this study, an adaptive meta-learning framework (AdaMETA) is proposed, which combines dynamic task-aware model-independent meta-learning (DT-MAML) with efficient multi-scale attention (EMA) modules to enhance the model’s ability to generalize and improve diagnostic performance in small-sample bearing fault diagnosis across different load scenarios. Specifically, a hierarchical encoder equipped with C-EMA is introduced to effectively capture multi-scale fault features from vibration signals, greatly improving feature extraction under constrained data conditions. Furthermore, DT-MAML dynamically adjusts the inner-loop learning rate based on task complexity, promoting efficient adaptation to diverse tasks and mitigating domain bias. Comprehensive experimental evaluations on the CWRU bearing dataset, conducted under carefully designed cross-domain scenarios, demonstrate that AdaMETA achieves superior accuracy (up to 99.26%) and robustness compared to traditional meta-learning and classical diagnostic methods. Additional ablation studies and noise interference experiments further validate the substantial contribution of the EMA module and the dynamic learning rate components. Full article
Show Figures

Figure 1

7 pages, 786 KB  
Proceeding Paper
Enhancing the Precision of Eye Detection with EEG-Based Machine Learning Models
by Masroor Ahmad, Tahir Muhammad Ali and Nunik Destria Arianti
Eng. Proc. 2025, 107(1), 128; https://doi.org/10.3390/engproc2025107128 - 13 Oct 2025
Viewed by 342
Abstract
Achieving a dataset of eye detection comprises a critical task in computer vision and image processing. The primary goal of this dataset is to accurately locate and identify the position of eyes in image or video frames. This process can firstly detect the [...] Read more.
Achieving a dataset of eye detection comprises a critical task in computer vision and image processing. The primary goal of this dataset is to accurately locate and identify the position of eyes in image or video frames. This process can firstly detect the face region and then focus on the eye regions. In this study, 14,980 examples of physiological signal recordings, most likely from EEG or similar sensors, were included in this dataset, which was created for the analysis of neural or sensor-based movement. The constant signals from specific sensor channels are represented by 14 numerical features (AF3, F7, F3, O1, O2, P7, P8, T8, FC5, FC6, etc.). These characteristics record complex changes in signal designs over time, which could suggest shifts in sensor or neuronal activity. Also, the dataset involves a binary target variable called eye detection, and this shows if an eye-related event—such as turning or an open/closed state—is identified during an individual case. The basic label of this dataset is eye detection in human beings, which has instances of (0,1). The eye detection dataset has 14 features and 14,980 instances that can be utilized for training a model. Full article
Show Figures

Figure 1

Back to TopTop