Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (906)

Search Parameters:
Keywords = multi-resolution fusion

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
21 pages, 5521 KB  
Article
AMS-YOLO: Asymmetric Multi-Scale Fusion Network for Cannabis Detection in UAV Imagery
by Xuelin Li, Huanyin Yue, Jianli Liu and Aonan Cheng
Drones 2025, 9(9), 629; https://doi.org/10.3390/drones9090629 (registering DOI) - 6 Sep 2025
Abstract
Cannabis is a strictly regulated plant in China, and its illegal cultivation presents significant challenges for social governance. Traditional manual patrol methods suffer from low coverage efficiency, while satellite imagery struggles to identify illicit plantations due to its limited spatial resolution, particularly for [...] Read more.
Cannabis is a strictly regulated plant in China, and its illegal cultivation presents significant challenges for social governance. Traditional manual patrol methods suffer from low coverage efficiency, while satellite imagery struggles to identify illicit plantations due to its limited spatial resolution, particularly for sparsely distributed and concealed cultivation. UAV remote sensing technology, with its high resolution and mobility, provides a promising solution for cannabis monitoring. However, existing detection methods still face challenges in terms of accuracy and robustness, particularly due to varying target scales, severe occlusion, and background interference. In this paper, we propose AMS-YOLO, a cannabis detection model tailored for UAV imagery. The model incorporates an asymmetric backbone network to improve texture perception by directing the model’s focus towards directional information. Additionally, it features a multi-scale fusion neck structure, incorporating partial convolution mechanisms to effectively improve cannabis detection in small target and complex background scenarios. To evaluate the model’s performance, we constructed a cannabis remote sensing dataset consisting of 1972 images. Experimental results show that AMS-YOLO achieves an mAP of 90.7% while maintaining efficient inference speed, outperforming existing state-of-the-art detection algorithms. This method demonstrates strong adaptability and practicality in complex environments, offering robust technical support for monitoring illegal cannabis cultivation. Full article
Show Figures

Figure 1

20 pages, 2226 KB  
Article
RST-Net: A Semantic Segmentation Network for Remote Sensing Images Based on a Dual-Branch Encoder Structure
by Na Yang, Chuanzhao Tian, Xingfa Gu, Yanting Zhang, Xuewen Li and Feng Zhang
Sensors 2025, 25(17), 5531; https://doi.org/10.3390/s25175531 - 5 Sep 2025
Viewed by 47
Abstract
High-resolution remote sensing images often suffer from inadequate fusion between global and local features, leading to the loss of long-range dependencies and blurred spatial details, while also exhibiting limited adaptability to multi-scale object segmentation. To overcome these limitations, this study proposes RST-Net, a [...] Read more.
High-resolution remote sensing images often suffer from inadequate fusion between global and local features, leading to the loss of long-range dependencies and blurred spatial details, while also exhibiting limited adaptability to multi-scale object segmentation. To overcome these limitations, this study proposes RST-Net, a semantic segmentation network featuring a dual-branch encoder structure. The encoder integrates a ResNeXt-50-based CNN branch for extracting local spatial features and a Shunted Transformer (ST) branch for capturing global contextual information. To further enhance multi-scale representation, the multi-scale feature enhancement module (MSFEM) is embedded in the CNN branch, leveraging atrous and depthwise separable convolutions to dynamically aggregate features. Additionally, the residual dynamic feature fusion (RDFF) module is incorporated into skip connections to improve interactions between encoder and decoder features. Experiments on the Vaihingen and Potsdam datasets show that RST-Net achieves promising performance, with MIoU scores of 77.04% and 79.56%, respectively, validating its effectiveness in semantic segmentation tasks. Full article
Show Figures

Figure 1

23 pages, 1476 KB  
Article
Dynamically Optimized Object Detection Algorithms for Aviation Safety
by Yi Qu, Cheng Wang, Yilei Xiao, Haijuan Ju and Jing Wu
Electronics 2025, 14(17), 3536; https://doi.org/10.3390/electronics14173536 - 4 Sep 2025
Viewed by 175
Abstract
Infrared imaging technology demonstrates significant advantages in aviation safety monitoring due to its exceptional all-weather operational capability and anti-interference characteristics, particularly in scenarios requiring real-time detection of aerial objects such as airport airspace management. However, traditional infrared target detection algorithms face critical challenges [...] Read more.
Infrared imaging technology demonstrates significant advantages in aviation safety monitoring due to its exceptional all-weather operational capability and anti-interference characteristics, particularly in scenarios requiring real-time detection of aerial objects such as airport airspace management. However, traditional infrared target detection algorithms face critical challenges in complex sky backgrounds, including low signal-to-noise ratio (SNR), small target dimensions, and strong background clutter, leading to insufficient detection accuracy and reliability. To address these issues, this paper proposes the AFK-YOLO model based on the YOLO11 framework: it integrates an ADown downsampling module, which utilizes a dual-branch strategy combining average pooling and max pooling to effectively minimize feature information loss during spatial resolution reduction; introduces the KernelWarehouse dynamic convolution approach, which adopts kernel partitioning and a contrastive attention-based cross-layer shared kernel repository to address the challenge of linear parameter growth in conventional dynamic convolution methods; and establishes a feature decoupling pyramid network (FDPN) that replaces static feature pyramids with a dynamic multi-scale fusion architecture, utilizing parallel multi-granularity convolutions and an EMA attention mechanism to achieve adaptive feature enhancement. Experiments demonstrate that the AFK-YOLO model achieves 78.6% mAP on a self-constructed aerial infrared dataset—a 2.4 percentage point improvement over the baseline YOLO11—while meeting real-time requirements for aviation safety monitoring (416.7 FPS), reducing parameters by 6.9%, and compressing weight size by 21.8%. The results demonstrate the effectiveness of dynamic optimization methods in improving the accuracy and robustness of infrared target detection under complex aerial environments, thereby providing reliable technical support for the prevention of mid-air collisions. Full article
(This article belongs to the Special Issue Computer Vision and AI Algorithms for Diverse Scenarios)
Show Figures

Figure 1

17 pages, 1294 KB  
Article
SPARSE-OTFS-Net: A Sparse Robust OTFS Signal Detection Algorithm for 6G Ubiquitous Coverage
by Yunzhi Ling and Jun Xu
Electronics 2025, 14(17), 3532; https://doi.org/10.3390/electronics14173532 - 4 Sep 2025
Viewed by 134
Abstract
With the evolution of 6G technology toward global coverage and multidimensional integration, OTFS modulation has become a research focus due to its advantages in high-mobility scenarios. However, existing OTFS signal detection algorithms face challenges such as pilot contamination, Doppler spread degradation, and diverse [...] Read more.
With the evolution of 6G technology toward global coverage and multidimensional integration, OTFS modulation has become a research focus due to its advantages in high-mobility scenarios. However, existing OTFS signal detection algorithms face challenges such as pilot contamination, Doppler spread degradation, and diverse interference in complex environments. This paper proposes the SPARSE-OTFS-Net algorithm, which establishes a comprehensive signal detection solution by innovatively integrating sparse random pilot design, compressive sensing-based frequency offset estimation with closed-loop cancellation, and joint denoising techniques combining an autoencoder, residual learning, and multi-scale feature fusion. The algorithm employs deep learning to dynamically generate non-uniform pilot distributions, reducing pilot contamination by 60%. Through orthogonal matching pursuit algorithms, it achieves super-resolution frequency offset estimation with tracking errors controlled within 20 Hz, effectively addressing Doppler spread degradation. The multi-stage denoising mechanism of deep neural networks suppresses various interferences while preserving time-frequency domain signal sparsity. Simulation results demonstrate: Under large frequency offset, multipath, and low SNR conditions, multi-kernel convolution technology achieves significant computational complexity reduction while exhibiting outstanding performance in tracking error and weak multipath detection. In 1000 km/h high-speed mobility scenarios, Doppler error estimation accuracy reaches ±25 Hz (approaching the Cramér-Rao bound), with BER performance of 5.0 × 10−6 (7× improvement over single-Gaussian CNN’s 3.5 × 10−5). In 1024-user interference scenarios with BER = 10−5 requirements, SNR demand decreases from 11.4 dB to 9.2 dB (2.2 dB reduction), while maintaining EVM at 6.5% under 1024-user concurrency (compared to 16.5% for conventional MMSE), effectively increasing concurrent user capacity in 6G ultra-massive connectivity scenarios. These results validate the superior performance of SPARSE-OTFS-Net in 6G ultra-massive connectivity applications and provide critical technical support for realizing integrated space–air–ground networks. Full article
(This article belongs to the Section Microwave and Wireless Communications)
Show Figures

Figure 1

22 pages, 11486 KB  
Article
RAP-Net: A Region Affinity Propagation-Guided Semantic Segmentation Network for Plateau Karst Landform Remote Sensing Imagery
by Dongsheng Zhong, Lingbo Cai, Shaoda Li, Wei Wang, Yijing Zhu, Yaning Liu and Ronghao Yang
Remote Sens. 2025, 17(17), 3082; https://doi.org/10.3390/rs17173082 - 4 Sep 2025
Viewed by 175
Abstract
Karst rocky desertification on the Qinghai–Tibet Plateau poses a severe threat to the region’s fragile ecosystem. Accordingly, the rapid and accurate delineation of plateau karst landforms is essential for monitoring ecological degradation and guiding restoration strategies. However, automatic recognition of these landforms in [...] Read more.
Karst rocky desertification on the Qinghai–Tibet Plateau poses a severe threat to the region’s fragile ecosystem. Accordingly, the rapid and accurate delineation of plateau karst landforms is essential for monitoring ecological degradation and guiding restoration strategies. However, automatic recognition of these landforms in remote sensing imagery is hindered by challenges such as blurred boundaries, fragmented targets, and poor intra-region consistency. To address these issues, we propose the Region Affinity Propagation Network (RAP-Net). This framework enhances intra-region consistency, edge sensitivity, and multi-scale context fusion through its core modules: Region Affinity Propagation (RAP), High-Frequency Multi-Scale Attention (HFMSA), and Global–Local Cross Attention (GLCA). In addition, we constructed the Plateau Karst Landform Dataset (PKLD), a high-resolution remote sensing dataset specifically tailored for this task, which provides a standardized benchmark for future studies. On the PKLD, RAP-Net surpasses eight state-of-the-art methods, achieving 3.69–10.31% higher IoU and 3.88–14.28% higher Recall, thereby demonstrating significant improvements in boundary delineation and structural completeness. Moreover, in a cross-regional generalization test on the Mount Genyen area, RAP-Net—trained solely on PKLD without fine-tuning—achieved 2.38% and 1.94% higher IoU and F1-scores, respectively, than the Swin Transformer, confirming its robustness and generalizability in complex, unseen environments. Full article
(This article belongs to the Section Remote Sensing in Geology, Geomorphology and Hydrology)
Show Figures

Figure 1

24 pages, 4010 KB  
Article
MFAFNet: A Multi-Feature Attention Fusion Network for Infrared Small Target Detection
by Zehao Zhao, Weining Chen, Seng Dong, Yaohong Chen and Hao Wang
Remote Sens. 2025, 17(17), 3070; https://doi.org/10.3390/rs17173070 - 3 Sep 2025
Viewed by 244
Abstract
Infrared small target detection is a critical task in remote sensing applications, such as aerial reconnaissance, maritime surveillance, and early-warning systems. However, due to the inherent characteristics of remote sensing imagery, such as complex backgrounds, low contrast, and limited spatial resolution-detecting small-scale, dim [...] Read more.
Infrared small target detection is a critical task in remote sensing applications, such as aerial reconnaissance, maritime surveillance, and early-warning systems. However, due to the inherent characteristics of remote sensing imagery, such as complex backgrounds, low contrast, and limited spatial resolution-detecting small-scale, dim infrared targets remains highly challenging. To address these issues, we propose MFAFNet, a novel Multi-Feature Attention Fusion Network tailored for infrared remote sensing scenarios. The network comprises three key modules: a Feature Interactive Fusion Module (FIFM), a Patch Attention Block (PAB), and an Asymmetric Contextual Fusion Module (ACFM). FIFM enhances target saliency by integrating the original infrared image with two locally enhanced feature maps capturing different receptive field scales. PAB exploits global contextual relationships by computing inter-pixel correlations across multi-scale patches, thus improving detection robustness in cluttered remote scenes. ACFM further refines feature representation by combining shallow spatial details with deep semantic cues, alleviating semantic gaps across feature hierarchies. Experimental results on two public remote sensing datasets, SIRST-Aug and IRSTD-1k, demonstrate that MFAFNet achieves excellent performance, with mean IoU values of 0.7465 and 0.6701, respectively, confirming its effectiveness and generalizability in infrared remote sensing image analysis. Full article
Show Figures

Figure 1

23 pages, 4776 KB  
Article
Category-Guided Transformer for Semantic Segmentation of High-Resolution Remote Sensing Images
by Yue Ni, Jiahang Liu, Hui Zhang, Weijian Chi and Ji Luan
Remote Sens. 2025, 17(17), 3054; https://doi.org/10.3390/rs17173054 - 2 Sep 2025
Viewed by 306
Abstract
High-resolution remote sensing images suffer from large intra-class variance, high inter-class similarity, and significant scale variations, leading to incomplete segmentation and imprecise boundaries. To address these challenges, Transformer-based methods, despite their strong global modeling capability, often suffer from feature confusion, weak detail representation, [...] Read more.
High-resolution remote sensing images suffer from large intra-class variance, high inter-class similarity, and significant scale variations, leading to incomplete segmentation and imprecise boundaries. To address these challenges, Transformer-based methods, despite their strong global modeling capability, often suffer from feature confusion, weak detail representation, and high computational cost. Moreover, existing multi-scale fusion mechanisms are prone to semantic misalignment across levels, hindering effective information integration and reducing boundary clarity. To address these issues, a Category-Guided Transformer (CIGFormer) is proposed. Specifically, the Category-Information-Guided Transformer Module (CIGTM) integrates global and local branches: the global branch combines window-based self-attention (WSAM) and window adaptive pooling self-attention (WAPSAM), using class predictions to enhance global context modeling and reduce intra-class and inter-class confusion; the local branch extracts multi-scale structural features to refine semantic representation and boundaries. In addition, an Adaptive Wavelet Fusion Module (AWFM) is designed, which leverages wavelet decomposition and channel-spatial joint attention for dynamic multi-scale fusion while preserving structural details. Extensive experiments on the ISPRS Vaihingen and Potsdam datasets demonstrate that CIGFormer, with only 21.50 M parameters, achieves outstanding performance in small object recognition, boundary refinement, and complex scene parsing, showing strong potential for practical applications. Full article
(This article belongs to the Section AI Remote Sensing)
Show Figures

Figure 1

16 pages, 2827 KB  
Article
A Dual-Modality CNN Approach for RSS-Based Indoor Positioning Using Spatial and Frequency Fingerprints
by Xiangchen Lai, Yunzhi Luo and Yong Jia
Sensors 2025, 25(17), 5408; https://doi.org/10.3390/s25175408 - 2 Sep 2025
Viewed by 232
Abstract
Indoor positioning systems based on received signal strength (RSS) achieve indoor positioning by leveraging the position-related features inherent in spatial RSS fingerprint images. Their positioning accuracy and robustness are directly influenced by the quality of fingerprint features. However, the inherent spatial low-resolution characteristic [...] Read more.
Indoor positioning systems based on received signal strength (RSS) achieve indoor positioning by leveraging the position-related features inherent in spatial RSS fingerprint images. Their positioning accuracy and robustness are directly influenced by the quality of fingerprint features. However, the inherent spatial low-resolution characteristic of spatial RSS fingerprint images makes it challenging to effectively extract subtle fingerprint features. To address this issue, this paper proposes an RSS-based indoor positioning method that combines enhanced spatial frequency fingerprint representation with fusion learning. First, bicubic interpolation is applied to improve image resolution and reveal finer spatial details. Then, a 2D fast Fourier transform (2D FFT) converts the enhanced spatial images into frequency domain representations to supplement spectral features. These spatial and frequency fingerprints are used as dual-modality inputs for a parallel convolutional neural network (CNN) model with efficient multi-scale attention (EMA) modules. The model extracts modality-specific features and fuses them to generate enriched representations. Each modality—spatial, frequency, and fused—is passed through a dedicated fully connected network to predict 3D coordinates. A coordinate optimization strategy is introduced to select the two most reliable outputs for each axis (x, y, z), and their average is used as the final estimate. Experiments on seven public datasets show that the proposed method significantly improves positioning accuracy, reducing the mean positioning error by up to 47.1% and root mean square error (RMSE) by up to 54.4% compared with traditional and advanced time–frequency methods. Full article
(This article belongs to the Section Navigation and Positioning)
Show Figures

Figure 1

28 pages, 1950 KB  
Review
Remote Sensing Approaches for Water Hyacinth and Water Quality Monitoring: Global Trends, Techniques, and Applications
by Lakachew Y. Alemneh, Daganchew Aklog, Ann van Griensven, Goraw Goshu, Seleshi Yalew, Wubneh B. Abebe, Minychl G. Dersseh, Demesew A. Mhiret, Claire I. Michailovsky, Selamawit Amare and Sisay Asress
Water 2025, 17(17), 2573; https://doi.org/10.3390/w17172573 - 31 Aug 2025
Viewed by 655
Abstract
Water hyacinth (Eichhornia crassipes), native to South America, is a highly invasive aquatic plant threatening freshwater ecosystems worldwide. Its rapid proliferation negatively impacts water quality, biodiversity, and navigation. Remote sensing offers an effective means to monitor such aquatic environments by providing extensive spatial [...] Read more.
Water hyacinth (Eichhornia crassipes), native to South America, is a highly invasive aquatic plant threatening freshwater ecosystems worldwide. Its rapid proliferation negatively impacts water quality, biodiversity, and navigation. Remote sensing offers an effective means to monitor such aquatic environments by providing extensive spatial and temporal coverage with improved resolution. This systematic review examines remote sensing applications for monitoring water hyacinth and water quality in studies published from 2014 to 2024. Seventy-eight peer-reviewed articles were selected from the Web of Science, Scopus, and Google Scholar following strict criteria. The research spans 25 countries across five continents, focusing mainly on lakes (61.5%), rivers (21%), and wetlands (10.3%). Approximately 49% of studies addressed water quality, 42% focused on water hyacinth, and 9% covered both. The Sentinel-2 Multispectral Instrument (MSI) was the most used sensor (35%), followed by the Landsat 8 Operational Land Imager (OLI) (26%). Multi-sensor fusion, especially Sentinel-2 MSI with Unmanned Aerial Vehicles (UAVs), was frequently applied to enhance monitoring capabilities. Detection accuracies ranged from 74% to 98% using statistical, machine learning, and deep learning techniques. Key challenges include limited ground-truth data and inadequate atmospheric correction. The integration of high-resolution sensors with advanced analytics shows strong promise for effective inland water monitoring. Full article
(This article belongs to the Section Ecohydrology)
Show Figures

Figure 1

28 pages, 19672 KB  
Article
A Multi-Fidelity Data Fusion Approach Based on Semi-Supervised Learning for Image Super-Resolution in Data-Scarce Scenarios
by Hongzheng Zhu, Yingjuan Zhao, Ximing Qiao, Jinshuo Zhang, Jingnan Ma and Sheng Tong
Sensors 2025, 25(17), 5373; https://doi.org/10.3390/s25175373 - 31 Aug 2025
Viewed by 430
Abstract
Image super-resolution (SR) techniques can significantly enhance visual quality and information density. However, existing methods often rely on large amounts of paired low- and high-resolution (LR-HR) data, which limits their generalization and robustness when faced with data scarcity, distribution inconsistencies, and missing high-frequency [...] Read more.
Image super-resolution (SR) techniques can significantly enhance visual quality and information density. However, existing methods often rely on large amounts of paired low- and high-resolution (LR-HR) data, which limits their generalization and robustness when faced with data scarcity, distribution inconsistencies, and missing high-frequency details. To tackle the challenges of image reconstruction in data-scarce scenarios, this paper proposes a semi-supervised learning-driven multi-fidelity fusion (SSLMF) method, which integrates multi-fidelity data fusion (MFDF) and semi-supervised learning (SSL) to reduce reliance on high-fidelity data. More specifically, (1) an MFDF strategy is employed to leverage low-fidelity data for global structural constraints, enhancing information compensation; (2) an SSL mechanism is introduced to reduce data dependence by using only a small amount of labeled HR samples along with a large quantity of unlabeled multi-fidelity data. This framework significantly improves data efficiency and reconstruction quality. We first validate the reconstruction accuracy of SSLMF on benchmark functions and then apply it to image reconstruction tasks. The results demonstrate that SSLMF can effectively model both linear and nonlinear relationships among multi-fidelity data, maintaining high performance even with limited high-fidelity samples. Finally, its cross-disciplinary potential is illustrated through an audio restoration case study, offering a novel solution for efficient image reconstruction, especially in data-scarce scenarios where high-fidelity samples are limited. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

17 pages, 16767 KB  
Article
AeroLight: A Lightweight Architecture with Dynamic Feature Fusion for High-Fidelity Small-Target Detection in Aerial Imagery
by Hao Qiu, Xiaoyan Meng, Yunjie Zhao, Liang Yu and Shuai Yin
Sensors 2025, 25(17), 5369; https://doi.org/10.3390/s25175369 - 30 Aug 2025
Viewed by 464
Abstract
Small-target detection in Unmanned Aerial Vehicle (UAV) aerial images remains a significant and unresolved challenge in aerial image analysis, hampered by low target resolution, dense object clustering, and complex, cluttered backgrounds. In order to cope with these problems, we present AeroLight, a novel [...] Read more.
Small-target detection in Unmanned Aerial Vehicle (UAV) aerial images remains a significant and unresolved challenge in aerial image analysis, hampered by low target resolution, dense object clustering, and complex, cluttered backgrounds. In order to cope with these problems, we present AeroLight, a novel and efficient detection architecture that achieves high-fidelity performance in resource-constrained environments. AeroLight is built upon three key innovations. First, we have optimized the feature pyramid at the architectural level by integrating a high-resolution head specifically designed for minute object detection. This design enhances sensitivity to fine-grained spatial details while streamlining redundant and computationally expensive network layers. Second, a Dynamic Feature Fusion (DFF) module is proposed to adaptively recalibrate and merge multi-scale feature maps, mitigating information loss during integration and strengthening object representation across diverse scales. Finally, we enhance the localization precision of irregular-shaped objects by refining bounding box regression using a Shape-IoU loss function. AeroLight is shown to improve mAP50 and mAP50-95 by 7.5% and 3.3%, respectively, on the VisDrone2019 dataset, while reducing the parameter count by 28.8% when compared with the baseline model. Further validation on the RSOD dataset and Huaxing Farm Drone dataset confirms its superior performance and generalization capabilities. AeroLight provides a powerful and efficient solution for real-world UAV applications, setting a new standard for lightweight, high-precision object recognition in aerial imaging scenarios. Full article
(This article belongs to the Section Remote Sensors)
Show Figures

Figure 1

18 pages, 2785 KB  
Article
A Study of Global Hourly Sea Surface Temperature Fusion Based on the Triple-Collocation Fusion Algorithm
by Lan Zhao and Haiyong Ding
Remote Sens. 2025, 17(17), 3014; https://doi.org/10.3390/rs17173014 - 29 Aug 2025
Viewed by 351
Abstract
Sea surface temperature (SST) is vital for climate monitoring and extreme weather forecasting. Existing global SST datasets are typically provided at daily to seasonal resolutions, while hourly data remain limited to regional scales. Polar-orbiting satellites offer global coverage but low temporal resolution, providing [...] Read more.
Sea surface temperature (SST) is vital for climate monitoring and extreme weather forecasting. Existing global SST datasets are typically provided at daily to seasonal resolutions, while hourly data remain limited to regional scales. Polar-orbiting satellites offer global coverage but low temporal resolution, providing only 1–2 observations per day. Geostationary satellites provide high temporal resolution but cover only part of the region. These limitations create a gap in the availability of global, hourly SST data. To address this, we propose a Triple-Collocation (TC)-based fusion algorithm for generating accurate global hourly SST data through multi-source integration. The method includes data preprocessing (quality control and linear interpolation), merging five geostationary SST datasets into two global sets by priority, applying TC fusion to three polar-orbiting datasets, and finally combining all sources via multi-source TC fusion. Results show improved temporal resolution and increased spatial coverage to 32%. The fused dataset achieves high accuracy, with a daily mean Bias below 0.0427 °C, RMSE about 0.5938 °C to 0.6965 °C, and R2 exceeding 0.9879. These outcomes demonstrate the method’s reliability and its potential for supporting climate and environmental research. Full article
(This article belongs to the Section Ocean Remote Sensing)
Show Figures

Figure 1

23 pages, 9065 KB  
Article
Multi-Scale Guided Context-Aware Transformer for Remote Sensing Building Extraction
by Mengxuan Yu, Jiepan Li and Wei He
Sensors 2025, 25(17), 5356; https://doi.org/10.3390/s25175356 - 29 Aug 2025
Viewed by 381
Abstract
Building extraction from high-resolution remote sensing imagery is critical for urban planning and disaster management, yet remains challenging due to significant intra-class variability in architectural styles and multi-scale distribution patterns of buildings. To address these limitations, we propose the Multi-Scale Guided Context-Aware Network [...] Read more.
Building extraction from high-resolution remote sensing imagery is critical for urban planning and disaster management, yet remains challenging due to significant intra-class variability in architectural styles and multi-scale distribution patterns of buildings. To address these limitations, we propose the Multi-Scale Guided Context-Aware Network (MSGCANet), a Transformer-based multi-scale guided context-aware network. Our framework integrates a Contextual Exploration Module (CEM) that synergizes asymmetric and progressive dilated convolutions to hierarchically expand receptive fields, enhancing discriminability for dense building features. We further design a Window-Guided Multi-Scale Attention Mechanism (WGMSAM) to dynamically establish cross-scale spatial dependencies through adaptive window partitioning, enabling precise fusion of local geometric details and global contextual semantics. Additionally, a cross-level Transformer decoder leverages deformable convolutions for spatially adaptive feature alignment and joint channel-spatial modeling. Experimental results show that MSGCANet achieves IoU values of 75.47%, 91.53%, and 83.10%, and F1-scores of 86.03%, 95.59%, and 90.78% on the Massachusetts, WHU, and Inria datasets, respectively, demonstrating robust performance across these datasets. Full article
(This article belongs to the Section Optical Sensors)
Show Figures

Figure 1

24 pages, 17568 KB  
Article
Super-Resolved Pseudo Reference in Dual-Branch Embedding for Blind Ultra-High-Definition Image Quality Assessment
by Jiacheng Gu, Qingxu Meng, Songnan Zhao, Yifan Wang, Shaode Yu and Qiurui Sun
Electronics 2025, 14(17), 3447; https://doi.org/10.3390/electronics14173447 - 29 Aug 2025
Viewed by 317
Abstract
In the Ultra-High-Definition (UHD) domain, blind image quality assessment remains challenging due to the high dimensionality of UHD images, which exceeds the input capacity of deep learning networks. Motivated by the visual discrepancies observed between high- and low-quality images after down-sampling and Super-Resolution [...] Read more.
In the Ultra-High-Definition (UHD) domain, blind image quality assessment remains challenging due to the high dimensionality of UHD images, which exceeds the input capacity of deep learning networks. Motivated by the visual discrepancies observed between high- and low-quality images after down-sampling and Super-Resolution (SR) reconstruction, we propose a SUper-Resolved Pseudo References In Dual-branch Embedding (SURPRIDE) framework tailored for UHD image quality prediction. SURPRIDE employs one branch to capture intrinsic quality features from the original patch input and the other to encode comparative perceptual cues from the SR-reconstructed pseudo-reference. The fusion of the complementary representation, guided by a novel hybrid loss function, enhances the network’s ability to model both absolute and relational quality cues. Key components of the framework are optimized through extensive ablation studies. Experimental results demonstrate that the SURPRIDE framework achieves competitive performance on two UHD benchmarks (AIM 2024 Challenge, PLCC = 0.7755, SRCC = 0.8133, on the testing set; HRIQ, PLCC = 0.882, SRCC = 0.873). Meanwhile, its effectiveness is verified on high- and standard-definition image datasets across diverse resolutions. Future work may explore positional encoding, advanced representation learning, and adaptive multi-branch fusion to align model predictions with human perceptual judgment in real-world scenarios. Full article
Show Figures

Figure 1

24 pages, 1689 KB  
Article
Safeguarding Brand and Platform Credibility Through AI-Based Multi-Model Fake Profile Detection
by Vishwas Chakranarayan, Fadheela Hussain, Fayzeh Abdulkareem Jaber, Redha J. Shaker and Ali Rizwan
Future Internet 2025, 17(9), 391; https://doi.org/10.3390/fi17090391 - 29 Aug 2025
Viewed by 368
Abstract
The proliferation of fake profiles on social media presents critical cybersecurity and misinformation challenges, necessitating robust and scalable detection mechanisms. Such profiles weaken consumer trust, reduce user engagement, and ultimately harm brand reputation and platform credibility. As adversarial tactics and synthetic identity generation [...] Read more.
The proliferation of fake profiles on social media presents critical cybersecurity and misinformation challenges, necessitating robust and scalable detection mechanisms. Such profiles weaken consumer trust, reduce user engagement, and ultimately harm brand reputation and platform credibility. As adversarial tactics and synthetic identity generation evolve, traditional rule-based and machine learning approaches struggle to detect evolving and deceptive behavioral patterns embedded in dynamic user-generated content. This study aims to develop an AI-driven, multi-modal deep learning-based detection system for identifying fake profiles that fuses textual, visual, and social network features to enhance detection accuracy. It also seeks to ensure scalability, adversarial robustness, and real-time threat detection capabilities suitable for practical deployment in industrial cybersecurity environments. To achieve these objectives, the current study proposes an integrated AI system that combines the Robustly Optimized BERT Pretraining Approach (RoBERTa) for deep semantic textual analysis, ConvNeXt for high-resolution profile image verification, and Heterogeneous Graph Attention Networks (Hetero-GAT) for modeling complex social interactions. The extracted features from all three modalities are fused through an attention-based late fusion strategy, enhancing interpretability, robustness, and cross-modal learning. Experimental evaluations on large-scale social media datasets demonstrate that the proposed RoBERTa-ConvNeXt-HeteroGAT model significantly outperforms baseline models, including Support Vector Machine (SVM), Random Forest, and Long Short-Term Memory (LSTM). Performance achieves 98.9% accuracy, 98.4% precision, and a 98.6% F1-score, with a per-profile speed of 15.7 milliseconds, enabling real-time applicability. Moreover, the model proves to be resilient against various types of attacks on text, images, and network activity. This study advances the application of AI in cybersecurity by introducing a highly interpretable, multi-modal detection system that strengthens digital trust, supports identity verification, and enhances the security of social media platforms. This alignment of technical robustness with brand trust highlights the system’s value not only in cybersecurity but also in sustaining platform credibility and consumer confidence. This system provides practical value to a wide range of stakeholders, including platform providers, AI researchers, cybersecurity professionals, and public sector regulators, by enabling real-time detection, improving operational efficiency, and safeguarding online ecosystems. Full article
Show Figures

Figure 1

Back to TopTop