Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (554)

Search Parameters:
Keywords = subpixel

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
20 pages, 7699 KB  
Article
Large-Gradient Displacement Monitoring and Parameter Inversion of Mining Collapse with the Optical Flow Method of Synthetic Aperture Radar Images
by Chuanjiu Zhang and Jie Chen
Remote Sens. 2025, 17(21), 3533; https://doi.org/10.3390/rs17213533 - 25 Oct 2025
Viewed by 290
Abstract
Monitoring large-gradient surface displacement caused by underground mining remains a significant challenge for conventional Synthetic Aperture Radar (SAR)-based techniques. This study introduces optical flow methods to monitor large-gradient displacement in mining areas and conducts a comprehensive comparison with Small Baseline Subset Interferometric SAR [...] Read more.
Monitoring large-gradient surface displacement caused by underground mining remains a significant challenge for conventional Synthetic Aperture Radar (SAR)-based techniques. This study introduces optical flow methods to monitor large-gradient displacement in mining areas and conducts a comprehensive comparison with Small Baseline Subset Interferometric SAR (SBAS-InSAR) and Pixel Offset Tracking (POT) methods. Using 12 high-resolution TerraSAR-X (TSX) SAR images over the Daliuta mining area in Yulin, China, we evaluate the performance of each method in terms of sensitivity to displacement gradients, computational efficiency, and monitoring accuracy. Results indicate that SBAS-InSAR is only capable of detecting displacement at the decimeter level in the Dalinta mining area and is unable to monitor rapid, large-gradient displacement exceeding the meter scale. While POT can detect meter-scale displacements, it suffers from low efficiency and low precision. In contrast, the proposed optical flow method (OFM) achieves sub-pixel accuracy with root mean square errors of 0.17 m (compared to 0.26 m for POT) when validated against Global Navigation Satellite System (GNSS) data while improving computational efficiency by nearly 30 times compared to POT. Furthermore, based on the optical flow results, mining parameters and three-dimensional (3D) displacement fields were successfully inverted, revealing maximum vertical subsidence exceeding 4.4 m and horizontal displacement over 1.5 m. These findings demonstrate that the OFM is a reliable and efficient tool for large-gradient displacement monitoring in mining areas, offering valuable support for hazard assessment and mining management. Full article
Show Figures

Figure 1

13 pages, 12430 KB  
Article
Moiré Reduction Technique for Near-Virtual-Image-Mode Light Field Displays via Aperture Array Modification
by Koichiro Fukano, Toshiki Yura and Yasuhiro Takaki
Appl. Sci. 2025, 15(20), 11031; https://doi.org/10.3390/app152011031 - 14 Oct 2025
Viewed by 268
Abstract
Flat-panel-type light field displays can generate three-dimensional images viewable without glasses; however, they often suffer from a narrow viewing zone, low light efficiency, low resolution, and moiré artifacts. Previously, flat-panel-type light field displays with a near-virtual-image mode were proposed, comprising a lens array [...] Read more.
Flat-panel-type light field displays can generate three-dimensional images viewable without glasses; however, they often suffer from a narrow viewing zone, low light efficiency, low resolution, and moiré artifacts. Previously, flat-panel-type light field displays with a near-virtual-image mode were proposed, comprising a lens array and an aperture array; these displays offered an enhanced viewing zone, increased light efficiency, and improved resolution. In this study, a moiré reduction technique is proposed for near-virtual-image-mode light field displays. In this configuration, moiré artifacts arise from the periodic deformation of virtual subpixel images seen through the lens array, caused by the nonrectangular subpixel structures for the R, G, and B colors of the display panel. To suppress the differences in subpixel shapes, the aperture shapes in the aperture array were modified from straight to zigzag shapes. Zigzag-shaped slits were designed, and their effectiveness in reducing moiré artifacts was evaluated using a diffraction-based moiré analysis technique. Experimental results demonstrated a lower moiré contrast with the designed zigzag slit than with the conventional straight slit, confirming the effectiveness of the proposed technique. Full article
(This article belongs to the Special Issue Optical Imaging and 3D Display Technologies)
Show Figures

Figure 1

19 pages, 1948 KB  
Article
Graph-MambaRoadDet: A Symmetry-Aware Dynamic Graph Framework for Road Damage Detection
by Zichun Tian, Xiaokang Shao and Yuqi Bai
Symmetry 2025, 17(10), 1654; https://doi.org/10.3390/sym17101654 - 5 Oct 2025
Viewed by 495
Abstract
Road-surface distress poses a serious threat to traffic safety and imposes a growing burden on urban maintenance budgets. While modern detectors based on convolutional networks and Vision Transformers achieve strong frame-level performance, they often overlook an essential property of road environments—structural symmetry [...] Read more.
Road-surface distress poses a serious threat to traffic safety and imposes a growing burden on urban maintenance budgets. While modern detectors based on convolutional networks and Vision Transformers achieve strong frame-level performance, they often overlook an essential property of road environments—structural symmetry within road networks and damage patterns. We present Graph-MambaRoadDet (GMRD), a symmetry-aware and lightweight framework that integrates dynamic graph reasoning with state–space modeling for accurate, topology-informed, and real-time road damage detection. Specifically, GMRD employs an EfficientViM-T1 backbone and two DefMamba blocks, whose deformable scanning paths capture sub-pixel crack patterns while preserving geometric symmetry. A superpixel-based graph is constructed by projecting image regions onto OpenStreetMap road segments, encoding both spatial structure and symmetric topological layout. We introduce a Graph-Generating State–Space Model (GG-SSM) that synthesizes sparse sample-specific adjacency in O(M) time, further refined by a fusion module that combines detector self-attention with prior symmetry constraints. A consistency loss promotes smooth predictions across symmetric or adjacent segments. The full INT8 model contains only 1.8 M parameters and 1.5 GFLOPs, sustaining 45 FPS at 7 W on a Jetson Orin Nano—eight times lighter and 1.7× faster than YOLOv8-s. On RDD2022, TD-RD, and RoadBench-100K, GMRD surpasses strong baselines by up to +6.1 mAP50:95 and, on the new RoadGraph-RDD benchmark, achieves +5.3 G-mAP and +0.05 consistency gain. Qualitative results demonstrate robustness under shadows, reflections, back-lighting, and occlusion. By explicitly modeling spatial and topological symmetry, GMRD offers a principled solution for city-scale road infrastructure monitoring under real-time and edge-computing constraints. Full article
(This article belongs to the Section Computer)
Show Figures

Figure 1

25 pages, 12510 KB  
Article
Computer Vision-Based Optical Odometry Sensors: A Comparative Study of Classical Tracking Methods for Non-Contact Surface Measurement
by Ignas Andrijauskas, Marius Šumanas, Andrius Dzedzickis, Wojciech Tanaś and Vytautas Bučinskas
Sensors 2025, 25(19), 6051; https://doi.org/10.3390/s25196051 - 1 Oct 2025
Viewed by 643
Abstract
This article presents a principled framework for selecting and tuning classical computer vision algorithms in the context of optical displacement sensing. By isolating key factors that affect algorithm behavior—such as feed window size and motion step size—the study seeks to move beyond intuition-based [...] Read more.
This article presents a principled framework for selecting and tuning classical computer vision algorithms in the context of optical displacement sensing. By isolating key factors that affect algorithm behavior—such as feed window size and motion step size—the study seeks to move beyond intuition-based practices and provide rigorous, repeatable performance evaluations. Computer vision-based optical odometry sensors offer non-contact, high-precision measurement capabilities essential for modern metrology and robotics applications. This paper presents a systematic comparative analysis of three classical tracking algorithms—phase correlation, template matching, and optical flow—for 2D surface displacement measurement using synthetic image sequences with subpixel-accurate ground truth. A virtual camera system generates controlled test conditions using a multi-circle trajectory pattern, enabling systematic evaluation of tracking performance using 400 × 400 and 200 × 200 pixel feed windows. The systematic characterization enables informed algorithm selection based on specific application requirements rather than empirical trial-and-error approaches. Full article
(This article belongs to the Section Optical Sensors)
Show Figures

Figure 1

28 pages, 32815 KB  
Article
LiteSAM: Lightweight and Robust Feature Matching for Satellite and Aerial Imagery
by Boya Wang, Shuo Wang, Yibin Han, Linfeng Xu and Dong Ye
Remote Sens. 2025, 17(19), 3349; https://doi.org/10.3390/rs17193349 - 1 Oct 2025
Viewed by 397
Abstract
We present a (Light)weight (S)atellite–(A)erial feature (M)atching framework (LiteSAM) for robust UAV absolute visual localization (AVL) in GPS-denied environments. Existing satellite–aerial matching methods struggle with large appearance variations, texture-scarce regions, and limited efficiency for real-time UAV [...] Read more.
We present a (Light)weight (S)atellite–(A)erial feature (M)atching framework (LiteSAM) for robust UAV absolute visual localization (AVL) in GPS-denied environments. Existing satellite–aerial matching methods struggle with large appearance variations, texture-scarce regions, and limited efficiency for real-time UAV applications. LiteSAM integrates three key components to address these issues. First, efficient multi-scale feature extraction optimizes representation, reducing inference latency for edge devices. Second, a Token Aggregation–Interaction Transformer (TAIFormer) with a convolutional token mixer (CTM) models inter- and intra-image correlations, enabling robust global–local feature fusion. Third, a MinGRU-based dynamic subpixel refinement module adaptively learns spatial offsets, enhancing subpixel-level matching accuracy and cross-scenario generalization. The experiments show that LiteSAM achieves competitive performance across multiple datasets. On UAV-VisLoc, LiteSAM attains an RMSE@30 of 17.86 m, outperforming state-of-the-art semi-dense methods such as EfficientLoFTR. Its optimized variant, LiteSAM (opt., without dual softmax), delivers inference times of 61.98 ms on standard GPUs and 497.49 ms on NVIDIA Jetson AGX Orin, which are 22.9% and 19.8% faster than EfficientLoFTR (opt.), respectively. With 6.31M parameters, which is 2.4× fewer than EfficientLoFTR’s 15.05M, LiteSAM proves to be suitable for edge deployment. Extensive evaluations on natural image matching and downstream vision tasks confirm its superior accuracy and efficiency for general feature matching. Full article
Show Figures

Figure 1

34 pages, 9527 KB  
Article
High-Resolution 3D Thermal Mapping: From Dual-Sensor Calibration to Thermally Enriched Point Clouds
by Neri Edgardo Güidi, Andrea di Filippo and Salvatore Barba
Appl. Sci. 2025, 15(19), 10491; https://doi.org/10.3390/app151910491 - 28 Sep 2025
Viewed by 504
Abstract
Thermal imaging is increasingly applied in remote sensing to identify material degradation, monitor structural integrity, and support energy diagnostics. However, its adoption is limited by the low spatial resolution of thermal sensors compared to RGB cameras. This study proposes a modular pipeline to [...] Read more.
Thermal imaging is increasingly applied in remote sensing to identify material degradation, monitor structural integrity, and support energy diagnostics. However, its adoption is limited by the low spatial resolution of thermal sensors compared to RGB cameras. This study proposes a modular pipeline to generate thermally enriched 3D point clouds by fusing RGB and thermal imagery acquired simultaneously with a dual-sensor unmanned aerial vehicle system. The methodology includes geometric calibration of both cameras, image undistortion, cross-spectral feature matching, and projection of radiometric data onto the photogrammetric model through a computed homography. Thermal values are extracted using a custom parser and assigned to 3D points based on visibility masks and interpolation strategies. Calibration achieved 81.8% chessboard detection, yielding subpixel reprojection errors. Among twelve evaluated algorithms, LightGlue retained 99% of its matches and delivered a reprojection accuracy of 18.2% at 1 px, 65.1% at 3 px and 79% at 5 px. A case study on photovoltaic panels demonstrates the method’s capability to map thermal patterns with low temperature deviation from ground-truth data. Developed entirely in Python, the workflow integrates into Agisoft Metashape or other software. The proposed approach enables cost-effective, high-resolution thermal mapping with applications in civil engineering, cultural heritage conservation, and environmental monitoring applications. Full article
Show Figures

Figure 1

21 pages, 10447 KB  
Article
Multi-Focus Imaging and U-Net Segmentation for Mesoscale Asphalt Film Structure Analysis—Method for Characterizing Asphalt Film Structures in RAP
by Ying Wang, Shuming Li, Weina She, Yichen Cai and Hongchao Zhang
Materials 2025, 18(18), 4363; https://doi.org/10.3390/ma18184363 - 18 Sep 2025
Viewed by 389
Abstract
This study presents a high-fidelity image acquisition method for asphalt film structure to address the challenge of capturing mesoscale structures, especially fine mineral filler and asphalt mastic. The method is particularly applied to the analysis of the mortar structure in reclaimed asphalt pavement [...] Read more.
This study presents a high-fidelity image acquisition method for asphalt film structure to address the challenge of capturing mesoscale structures, especially fine mineral filler and asphalt mastic. The method is particularly applied to the analysis of the mortar structure in reclaimed asphalt pavement (RAP) mixtures. A digital camera combined with image stacking and texture suppression techniques was used to develop a reproducible imaging protocol. The resulting sub-pixel images significantly improved clarity and structural integrity, particularly for particles smaller than 0.075 mm. U-Net-based segmentation identified 588,513 aggregate particles—34 times more than in standard images (17,428). Among them, 95% were smaller than 0.075 mm compared to just 45% in standard images. Furthermore, segmentation accuracy reached 99.3% in high-resolution images, surpassing the 98.1% in standard images. These results confirm the method’s strong capability to preserve microscale features and enhance fine particle recognition, making it more effective than conventional imaging approaches. This study bridges physical and digital workflows in asphalt material analysis, offering a scalable, reproducible pipeline for fine-structure identification. The methodology provides foundational support for data-driven pavement modeling, material optimization, and future integration into digital twin frameworks for intelligent infrastructure systems. Full article
(This article belongs to the Special Issue Recent Advances in Reclaimed Asphalt Pavement (RAP) Materials)
Show Figures

Figure 1

22 pages, 6968 KB  
Article
Signatures of Breaking Waves in a Coastal Polynya Covered with Frazil Ice: A High-Resolution Satellite Image Case Study of Terra Nova Bay Polynya
by Katarzyna Bradtke, Wojciech Brodziński and Agnieszka Herman
Remote Sens. 2025, 17(18), 3198; https://doi.org/10.3390/rs17183198 - 16 Sep 2025
Viewed by 512
Abstract
The study focuses on the detection of breaking wave crests in the highly dynamic waters of an Antarctic coastal polynya using high-resolution panchromatic satellite imagery. Accurate assessment of whitecap coverage is crucial for improving our understanding of the interactions between wave generation, air–sea [...] Read more.
The study focuses on the detection of breaking wave crests in the highly dynamic waters of an Antarctic coastal polynya using high-resolution panchromatic satellite imagery. Accurate assessment of whitecap coverage is crucial for improving our understanding of the interactions between wave generation, air–sea heat exchange, and sea ice formation in these complex environments. As open-ocean whitecap detection methods are inadequate in coastal polynyas partially covered with frazil ice, we discuss an approach that exploits specific lighting conditions: the alignment of sunlight with the dominant wind direction and low solar elevation. Under such conditions, steep breaking waves cast pronounced shadows, which are used as the primary indicator of wave crests, particularly in frazil streak zones. The algorithm is optimized to exploit these conditions and minimize false positives along frazil streak boundaries. We applied the algorithm to a WorldView-2 image covering different parts of Terra Nova Bay Polynya (Ross Sea), a dynamic polar coastal zone. This case study demonstrates that the spatial distribution of detected breaking waves is consistent with ice conditions and wind forcing patterns, while also revealing deviations that point to complex wind–wave–ice interactions. Although quantitative validation of satellite-derived whitecaps coverage was not possible due to the lack of in situ data, the method performs reliably under a range of conditions. Limitations of the proposed approach are pointed out and discussed. Finally, the study highlights the risk of misinterpretation of lower-resolution reflectance data in areas where whitecaps and sea ice coexist at subpixel scales. Full article
Show Figures

Figure 1

26 pages, 6338 KB  
Article
Practical Measurements of Vibration Using the Moiré Effect
by Vladimir Saveljev and Gwanghee Heo
Appl. Mech. 2025, 6(3), 66; https://doi.org/10.3390/applmech6030066 - 4 Sep 2025
Viewed by 766
Abstract
Displacement measurement is a critical issue in mechanical engineering. The moiré effect increases the accuracy of contactless measurements. We theoretically estimated the sensitivity threshold of moiré measurements using a digital camera on various objects. The estimated sensitivity threshold can be as low as [...] Read more.
Displacement measurement is a critical issue in mechanical engineering. The moiré effect increases the accuracy of contactless measurements. We theoretically estimated the sensitivity threshold of moiré measurements using a digital camera on various objects. The estimated sensitivity threshold can be as low as a sub-pixel. We confirmed this experimentally in laboratory tests with a static image on a screen and simulated movement with non-integer and fractional amplitudes. Additionally, we provide practical examples, such as displacement measurement tests conducted in laboratories and outdoors. We took simultaneous measurements in two directions. The results can be applied in public safety, particularly for monitoring the condition of engineering structures. Full article
Show Figures

Figure 1

44 pages, 3439 KB  
Review
Conventional to Deep Learning Methods for Hyperspectral Unmixing: A Review
by Jinlin Zou, Hongwei Qu and Peng Zhang
Remote Sens. 2025, 17(17), 2968; https://doi.org/10.3390/rs17172968 - 27 Aug 2025
Viewed by 2323
Abstract
Hyperspectral images often contain many mixed pixels, primarily resulting from their inherent complexity and low spatial resolution. To enhance surface classification and improve sub-pixel target detection accuracy, hyperspectral unmixing technology has consistently become a topical issue. This review provides a comprehensive overview of [...] Read more.
Hyperspectral images often contain many mixed pixels, primarily resulting from their inherent complexity and low spatial resolution. To enhance surface classification and improve sub-pixel target detection accuracy, hyperspectral unmixing technology has consistently become a topical issue. This review provides a comprehensive overview of methodologies for hyperspectral unmixing, from traditional to advanced deep learning approaches. A systematic analysis of various challenges is presented, clarifying underlying principles and evaluating the strengths and limitations of prevalent algorithms. Hyperspectral unmixing is critical for interpreting spectral imagery but faces significant challenges: limited ground-truth data, spectral variability, nonlinear mixing effects, computational demands, and barriers to practical commercialization. Future progress requires bridging the gap to applications through user-centric solutions and integrating multi-modal and multi-temporal data. Research priorities include uncertainty quantification, transfer learning for generalization, neuromorphic edge computing, and developing tuning-free foundation models for cross-scenario robustness. This paper is designed to foster the commercial application of hyperspectral unmixing algorithms and to offer robust support for engineering applications within the hyperspectral remote sensing domain. Full article
(This article belongs to the Special Issue Artificial Intelligence in Hyperspectral Remote Sensing Data Analysis)
Show Figures

Figure 1

18 pages, 2565 KB  
Article
Rock Joint Segmentation in Drill Core Images via a Boundary-Aware Token-Mixing Network
by Seungjoo Lee, Yongjin Kim, Yongseong Kim, Jongseol Park and Bongjun Ji
Buildings 2025, 15(17), 3022; https://doi.org/10.3390/buildings15173022 - 25 Aug 2025
Viewed by 581
Abstract
The precise mapping of rock joint traces is fundamental to the design and safety assessment of foundations, retaining structures, and underground cavities in building and civil engineering. Existing deep learning approaches either impose prohibitive computational demands for on-site deployment or disrupt the topological [...] Read more.
The precise mapping of rock joint traces is fundamental to the design and safety assessment of foundations, retaining structures, and underground cavities in building and civil engineering. Existing deep learning approaches either impose prohibitive computational demands for on-site deployment or disrupt the topological continuity of subpixel lineaments that govern rock mass behavior. This study presents BATNet-Lite, a lightweight encoder–decoder architecture optimized for joint segmentation on resource-constrained devices. The encoder introduces a Boundary-Aware Token-Mixing (BATM) block that separates feature maps into patch tokens and directionally pooled stripe tokens, and a bidirectional attention mechanism subsequently transfers global context to local descriptors while refining stripe features, thereby capturing long-range connectivity with negligible overhead. A complementary Multi-Scale Line Enhancement (MLE) module combines depth-wise dilated and deformable convolutions to yield scale-invariant responses to joints of varying apertures. In the decoder, a Skeletal-Contrastive Decoder (SCD) employs dual heads to predict segmentation and skeleton maps simultaneously, while an InfoNCE-based contrastive loss enforces their topological consistency without requiring explicit skeleton labels. Training leverages a composite focal Tversky and edge IoU loss under a curriculum-thinning schedule, improving edge adherence and continuity. Ablation experiments confirm that BATM, MLE, and SCD each contribute substantial gains in boundary accuracy and connectivity preservation. By delivering topology-preserving joint maps with small parameters, BATNet-Lite facilitates rapid geological data acquisition for tunnel face mapping, slope inspection, and subsurface digital twin development, thereby supporting safer and more efficient building and underground engineering practice. Full article
Show Figures

Figure 1

17 pages, 4223 KB  
Article
Space–Bandwidth Product Extension for Holographic Displays Through Cascaded Wavefront Modulation
by Shenao Zhang, Wenjia Li, Bo Dai, Qi Wang, Songlin Zhuang, Dawei Zhang and Chenliang Chang
Appl. Sci. 2025, 15(17), 9237; https://doi.org/10.3390/app15179237 - 22 Aug 2025
Viewed by 487
Abstract
The immersive experience of holographic displays is fundamentally limited by their space–bandwidth product (SBP), which imposes an inherent trade-off between the field of view (FOV) and eyebox size. This paper proposes a method to extend the SBP by employing cascaded modulation with a [...] Read more.
The immersive experience of holographic displays is fundamentally limited by their space–bandwidth product (SBP), which imposes an inherent trade-off between the field of view (FOV) and eyebox size. This paper proposes a method to extend the SBP by employing cascaded modulation with a dynamic spatial light modulator (SLM) and a passive high-resolution binary random phase mask (BRPM). We find that the key to unlocking this extension of SBP lies in a sophisticated algorithmic optimization, grounded in a physically accurate model of the system. We identify and correct the Nyquist undersampling problem caused by high-frequency scattering in standard diffraction models. Based on this physically accurate model, we employ a gradient descent optimization framework to achieve efficient, end-to-end solving for complex light fields. Simulation and experimental results demonstrate that our method achieves an approximately 16-fold SBP extension (4-fold FOV) while delivering significantly superior reconstructed image quality compared to the traditional Gerchberg–Saxton (GS) algorithm. Furthermore, this study quantitatively reveals the system’s extreme sensitivity to sub-pixel level alignment accuracy, providing critical guidance for the engineering and implementation of our proposed method. Full article
Show Figures

Figure 1

17 pages, 7225 KB  
Article
Placido Sub-Pixel Edge Detection Algorithm Based on Enhanced Mexican Hat Wavelet Transform and Improved Zernike Moments
by Yujie Wang, Jinyu Liang, Yating Xiao, Xinfeng Liu, Jiale Li, Guangyu Cui and Quan Zhang
J. Imaging 2025, 11(8), 267; https://doi.org/10.3390/jimaging11080267 - 11 Aug 2025
Viewed by 537
Abstract
In order to meet the high-precision location requirements of the corneal Placido ring edge in corneal topographic reconstruction, this paper proposes a sub-pixel edge detection algorithm based on multi-scale and multi-position enhanced Mexican Hat Wavelet Transform and improved Zernike moment. Firstly, the image [...] Read more.
In order to meet the high-precision location requirements of the corneal Placido ring edge in corneal topographic reconstruction, this paper proposes a sub-pixel edge detection algorithm based on multi-scale and multi-position enhanced Mexican Hat Wavelet Transform and improved Zernike moment. Firstly, the image undergoes preliminary processing using a multi-scale and multi-position enhanced Mexican Hat Wavelet Transform function. Subsequently, the preliminary edge information extracted is relocated based on the Zernike moments of a 9 × 9 template. Finally, two improved adaptive edge threshold algorithms are employed to determine the actual sub-pixel edge points of the image, thereby realizing sub-pixel edge detection for corneal Placido ring images. Through comparison and analysis of edge extraction results from real human eye images obtained using the algorithm proposed in this paper and those from other existing algorithms, it is observed that the average sub-pixel edge error of other algorithms is 0.286 pixels, whereas the proposed algorithm achieves an average error of only 0.094 pixels. Furthermore, the proposed algorithm demonstrates strong robustness against noise. Full article
(This article belongs to the Section Medical Imaging)
Show Figures

Figure 1

23 pages, 8942 KB  
Article
Optical and SAR Image Registration in Equatorial Cloudy Regions Guided by Automatically Point-Prompted Cloud Masks
by Yifan Liao, Shuo Li, Mingyang Gao, Shizhong Li, Wei Qin, Qiang Xiong, Cong Lin, Qi Chen and Pengjie Tao
Remote Sens. 2025, 17(15), 2630; https://doi.org/10.3390/rs17152630 - 29 Jul 2025
Viewed by 735
Abstract
The equator’s unique combination of high humidity and temperature renders optical satellite imagery highly susceptible to persistent cloud cover. In contrast, synthetic aperture radar (SAR) offers a robust alternative due to its ability to penetrate clouds with microwave imaging. This study addresses the [...] Read more.
The equator’s unique combination of high humidity and temperature renders optical satellite imagery highly susceptible to persistent cloud cover. In contrast, synthetic aperture radar (SAR) offers a robust alternative due to its ability to penetrate clouds with microwave imaging. This study addresses the challenges of cloud-induced data gaps and cross-sensor geometric biases by proposing an advanced optical and SAR image-matching framework specifically designed for cloud-prone equatorial regions. We use a prompt-driven visual segmentation model with automatic prompt point generation to produce cloud masks that guide cross-modal feature-matching and joint adjustment of optical and SAR data. This process results in a comprehensive digital orthophoto map (DOM) with high geometric consistency, retaining the fine spatial detail of optical data and the all-weather reliability of SAR. We validate our approach across four equatorial regions using five satellite platforms with varying spatial resolutions and revisit intervals. Even in areas with more than 50 percent cloud cover, our method maintains sub-pixel edging accuracy under manual check points and delivers comprehensive DOM products, establishing a reliable foundation for downstream environmental monitoring and ecosystem analysis. Full article
Show Figures

Figure 1

24 pages, 8483 KB  
Article
A Weakly Supervised Network for Coarse-to-Fine Change Detection in Hyperspectral Images
by Yadong Zhao and Zhao Chen
Remote Sens. 2025, 17(15), 2624; https://doi.org/10.3390/rs17152624 - 28 Jul 2025
Viewed by 833
Abstract
Hyperspectral image change detection (HSI-CD) provides substantial value in environmental monitoring, urban planning and other fields. In recent years, deep-learning based HSI-CD methods have made remarkable progress due to their powerful nonlinear feature learning capabilities, yet they face several challenges: mixed-pixel phenomenon affecting [...] Read more.
Hyperspectral image change detection (HSI-CD) provides substantial value in environmental monitoring, urban planning and other fields. In recent years, deep-learning based HSI-CD methods have made remarkable progress due to their powerful nonlinear feature learning capabilities, yet they face several challenges: mixed-pixel phenomenon affecting pixel-level detection accuracy; heterogeneous spatial scales of change targets where coarse-grained features fail to preserve fine-grained details; and dependence on high-quality labels. To address these challenges, this paper introduces WSCDNet, a weakly supervised HSI-CD network employing coarse-to-fine feature learning, with key innovations including: (1) A dual-branch detection framework integrating binary and multiclass change detection at the sub-pixel level that enhances collaborative optimization through a cross-feature coupling module; (2) introduction of multi-granularity aggregation and difference feature enhancement module for detecting easily confused regions, which effectively improves the model’s detection accuracy; and (3) proposal of a weakly supervised learning strategy, reducing model sensitivity to noisy pseudo-labels through decision-level consistency measurement and sample filtering mechanisms. Experimental results demonstrate that WSCDNet effectively enhances the accuracy and robustness of HSI-CD tasks, exhibiting superior performance under complex scenarios and weakly supervised conditions. Full article
(This article belongs to the Section Remote Sensing Image Processing)
Show Figures

Figure 1

Back to TopTop