Next Article in Journal
Improved Detection and Location of Small Crop Organs by Fusing UAV Orthophoto Maps and Raw Images
Previous Article in Journal
Flood Mapping and Assessment of Crop Damage Based on Multi-Source Remote Sensing: A Case Study of the “7.27” Rainstorm in Hebei Province, China
Previous Article in Special Issue
Assessing Spatiotemporal Dynamics of Net Primary Productivity in Shandong Province, China (2001–2020) Using the CASA Model and Google Earth Engine: Trends, Patterns, and Driving Factors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Improved Low-Light Image Feature Matching Algorithm Based on the SuperGlue Net Model

1
School of Geomatics Science and Technology, Nanjing Tech University, Nanjing 211816, China
2
Institute of Geospatial Information, Information Engineering University, Zhengzhou 450001, China
3
College of Civil Engineering, Nanjing Forestry University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(5), 905; https://doi.org/10.3390/rs17050905
Submission received: 25 November 2024 / Revised: 1 February 2025 / Accepted: 28 February 2025 / Published: 4 March 2025
(This article belongs to the Special Issue GeoAI and EO Big Data Driven Advances in Earth Environmental Science)

Abstract

:
The SuperGlue algorithm, which integrates deep learning theory with the SuperPoint feature extraction operator and addresses the matching problem using the classical Sinkhorn method, has significantly enhanced matching efficiency and become a prominent research focus. However, existing feature extraction operators often struggle to extract high-quality features from extremely low-light or dark images, resulting in reduced matching accuracy. In this study, we propose a novel feature matching method that combines multi-scale retinex with color restoration (MSRCR) and SuperGlue to address this challenge, enabling effective feature extraction and matching from dark images, successfully addressing the challenges of feature point extraction difficulties, sparse matching points, and low matching accuracy in extreme environments such as nighttime autonomous navigation, mine exploration, and tunnel operations. Our approach first employs the retinex-based MSRCR algorithm to enhance features in original low-light images, followed by utilizing the enhanced image pairs as inputs for SuperGlue feature matching. Experimental results validate the effectiveness of our method, demonstrating that both the quantity of extracted feature points and correctly matched feature points approximately doubles compared to traditional methods, thereby significantly improving matching accuracy in dark images.

1. Introduction

With the advancement of manufacturing industries and artificial intelligence technology, industrial vision technology continues to innovate and expand its applications, maintaining computer vision [1] as a prominent research domain. A critical challenge lies in extracting effective information from images for various applications, including autonomous navigation [2,3,4,5,6], autonomous driving [7,8,9], and robotics [10,11]. Image information extraction primarily encompasses two aspects: feature enhancement and feature matching. Feature enhancement algorithms focus on improving image brightness and contrast while reducing noise to achieve enhanced features. These algorithms include deep learning-based methods [12], retinex-based approaches [13], image inversion techniques [14], dehazing algorithms [15], and histogram equalization (HE) methods [16]. Feature matching algorithms aim to extract feature points from image pairs, establish visual descriptors, and complete feature matching based on optimal path determination. While traditional feature matching algorithms include SIFT [17], ORB [18], and SURF [19], more recent developments, such as the SuperGlue algorithm proposed by Sarlin P. E. et al. [20] in 2020, incorporate deep learning networks. As a fundamental task in machine vision, feature matching quality directly influences the accuracy of applications in autonomous navigation, self-driving vehicles, and robotics. Consequently, recent years have witnessed an increasing research focus on enhancing the accuracy of feature matching algorithms.
Over the years, researchers have invested considerable effort in enhancing the efficiency and accuracy of feature matching algorithms. In 2019, Yu Chen [21] developed a mobile platform visual matching method that integrated FAST and BRIEF algorithms while incorporating an evaluation function. This approach improved the accuracy of conventional ORB feature matching and effectively addressed motion blur issues, though challenges persisted in matching features under extremely low-light conditions. Wang Beiyi [22] introduced an innovative image matching algorithm in 2021 that combined SURF with the Fast Library for Approximate Nearest Neighbors (FLANN). This algorithm employed Fast-Hessian for feature point detection and integrated SURF with FLANN’s RANSAC to eliminate outliers, thereby enhancing both accuracy and real-time performance of feature matching. However, its effectiveness was primarily demonstrated under normal lighting conditions. A significant breakthrough came in 2020 when Paul-Edouard Sarlin [20] introduced the SuperGlue matching algorithm, which achieved remarkable matching accuracy in both indoor and outdoor scenarios. The algorithm comprises two primary modules: an attentional graph neural network and an optimal matching layer. It utilizes SuperPoint for feature point extraction and descriptor establishment, combining these with an attention graph neural network for iterative feature point aggregation. This process enhances feature point quality before processing through the optimal matching layer, which computes the score matrix and employs the classical Sinkhorn algorithm for image matching. While SuperGlue demonstrated superior performance across various conditions, including moderate low-light scenarios, it still faced limitations in extremely dark environments. Building upon these developments, Jiaming Sun [23] proposed the LoFTR algorithm in 2021, inspired by SuperGlue’s attention mechanism. LoFTR employs self-attention and cross-attention within a transformer framework to generate feature descriptors for matching image pairs, establishing initial pixel-level dense matching at a coarse level before refining the assignments at finer scales.
In 2022, Qing Wang [24] advanced the field further with Matchformer, a novel hierarchical feature extraction and matching transformer. By incorporating attention mechanisms throughout a hierarchical encoder and combining self-attention and cross-attention on multi-scale features, Matchformer achieved enhanced matching robustness. In the same year, Jiayuan Li [25] introduced the LNIFT algorithm for spatial-domain multimodal feature matching. This approach first employs normalized filters for feature point detection and description, then detects oriented FAST and ORB keypoints on normalized images, improving keypoint distribution through adaptive non-maximum suppression (ANMS). Furthermore, LNIFT utilizes descriptors such as oriented gradient histograms (HOG) to characterize keypoints on normalized images, resulting in improved matching accuracy. In 2023, Lindenberger P et al. [26] proposed a lightweight and fast matching algorithm called LightGlue. The algorithm features a lightweight network head and improves matching speed by dynamically reducing the number of layers based on the difficulty of image pairs and by early elimination of unmatched points.
In 2024, Yongxian Zhang et al. [27] proposed a multimodal remote sensing image matching method (MICM) utilizing learned features and attention mechanisms to address linear radiation errors and geometric distortion problems. In 2024, Bai Zhu et al. [28] proposed a ground-to-space robust matching algorithm (VDFT) based on viewpoint-invariant deformable feature transformation. The algorithm introduces a ResNet-like block that fuses Deformable Convolutional Networks (DCNs) and Depth-wise Separable Convolutions (DSCs), constructing a learnable and deformable feature network to address viewpoint span issues. They designed an enhanced joint detection and description strategy that improves feature point localization accuracy and representation capability through multi-level feature concatenation. Finally, the matching model is built using self-attention and cross-attention mechanisms to aggregate feature point information and enhance matching precision. While the algorithm addresses matching problems with large viewpoint spans, it is primarily based on research conducted using normal light and resolution imagery. The method has not addressed registration for low-light and highly blurred images, which limits its applicability in scenarios where high-quality reference images cannot be obtained. Additionally, the algorithm is constrained by lighting conditions, making high-precision nighttime registration challenging. In 2024, Jiang H et al. [29] proposed an omnigeneralized matching algorithm called OmniGlue, which enhances model generalization capabilities by introducing foundational model guidance and keypoint location attention guidance.
However, in practical applications, the low-light image pairs we obtain have extremely low brightness. To meet the all-day, all-weather requirements of visual SLAM [30], nighttime navigation and positioning become necessary, thus requiring high-quality feature extraction and matching for nighttime image pairs. But due to the problems of low brightness, low contrast, and high noise in the acquired images, traditional feature extraction operators suffer from issues such as few extracted features and low accuracy. When integrated into existing feature matching algorithms, these problems persist, resulting in fewer matching feature points and low precision. Feature matching algorithms have undergone long-term development, and whether traditional methods like SIFT, ORB, SURF, and FAST, or advanced matching algorithms like SuperGlue, Matchformer, LightGlue, LoFTR, and OmniGlue, they have achieved good matching results in normal light image pairs. However, they still have not solved the image matching problems under extreme conditions such as low-light and nighttime environments, and cannot meet the actual production requirements for matching in ultra-dark environments like nighttime autonomous navigation, tunnels, and mines.
To address the difficulty of feature point extraction in low-light images, in 2021, Risheng Liu et al. [31] proposed a novel method based on Retinex-inspired Architecture Search Unfolding (RUAS), designed to construct a lightweight yet effective enhancement network for handling low-light image enhancement in real-world scenarios. In 2022, Wenhui Wu et al. [32] proposed a deep unfolding network based on retinex theory, where the authors transformed the optimization problem into a learnable network for decomposing low-light images into reflectance and illumination layers. By formulating the decomposition problem as a regularized model with implicit prior conditions, they designed three learning modules responsible for data-based initialization, efficient unfolding optimization, and user-specified illumination enhancement. In 2023, Jiang Hai et al. [33] proposed a network (R2RNet) based on retinex theory for enhancing low-light images, transforming images from real low-light conditions to real normal-light conditions. In 2024, Ji Wu et al. [34] proposed a neural network-based low-light image enhancement network. This approach achieves feature enhancement of dark images, improving the brightness and texture information of the images.
To address the challenges of feature point extraction and matching in extremely low-light conditions, we propose a novel feature matching methodology that integrates multi-scale retinex with color restoration (MSRCR) and SuperGlue. First, the ultra-dark matching image pairs are input into the MSRCR feature enhancement algorithm for extreme feature enhancement of ultra-dark images. MSRCR is a feature enhancement algorithm based on retinex, which divides the image into illumination and reflection components. The reflection component is treated as the true surface information of objects. By using Gaussian filtering to remove the illumination component and then restoring the image’s true information through color recovery, the enhanced images are subsequently input into the SuperGlue feature matching algorithm for feature matching. This approach significantly enhances both the quantity of extracted feature points and the accuracy of matched feature points, thereby improving the efficiency of feature extraction operators and overall matching quality while substantially reducing rotation and translation errors. Our methodology follows a systematic approach: First, the SuperGlue feature matching algorithm is applied to the original low-light image pairs. Subsequently, these images undergo feature enhancement using MSRCR, followed by a second application of SuperGlue for feature matching on the enhanced images. The effectiveness of our proposed method is validated through comprehensive precision evaluation, comparing both the number of feature points and rotational/translational errors before and after enhancement. Comparative experimental results demonstrate the superior performance of our proposed approach.
This article is organized as follows: Section 2 presents a comprehensive analysis of the retinex-based feature enhancement algorithm and examines the SuperGlue feature matching algorithm, culminating in the proposal of our novel integrated approach that combines MSRCR and SuperGlue. Section 3 details the comparative experimental evaluation of the proposed matching method, with particular emphasis on its accuracy in low-light image matching scenarios. Section 4 provides an in-depth discussion of the method’s effectiveness and implications. Finally, Section 5 presents conclusions drawn from the experimental findings and their analysis.

2. Methodology

Contemporary advancements in image processing algorithms have progressed along parallel but distinct paths: enhancement algorithms have primarily focused on restoring authentic image information, improving brightness, and enhancing contrast, while matching algorithms have concentrated on developing novel techniques to improve accuracy and robustness. However, there has been limited exploration of the synergistic potential between these two domains. To address the persistent challenges of low accuracy and mismatching in low-light image scenarios, we propose a novel integrated approach that combines MSRCR and SuperGlue, bridging the gap between enhancement and matching methodologies.
This paper presents a comparative analysis of matching performance before and after enhancement, evaluating the accuracy improvements achieved through the integration of MSRCR and SuperGlue feature matching methods, as illustrated in Figure 1. The methodology comprises the following sequential steps:
Step (1):
The original low-light image pairs are directly processed through the SuperGlue feature matching algorithm to establish baseline matching results.
Step (2):
These same image pairs undergo feature enhancement using the MSRCR algorithm, which restores intrinsic image information while improving brightness and contrast characteristics.
Step (3):
The enhanced image pairs are then processed through the SuperGlue matching algorithm to obtain post-enhancement matching results.
Step (4):
A comprehensive comparative analysis is conducted between the original and post-enhancement matching results to evaluate the accuracy and robustness of the integrated MSRCR-SuperGlue approach.

2.1. Feature Enhancement Theory Based on Retinex

The retinex theory establishes a fundamental framework for image decomposition into two distinct components: illumination and reflectance. The reflectance component, which characterizes the inherent reflectivity of objects, constitutes the high-frequency elements of the image and maintains stability regardless of external lighting conditions. Conversely, the illumination component encompasses the low-frequency elements and is significantly influenced by ambient lighting conditions. By strategically suppressing or eliminating the illumination component, it becomes possible to recover the intrinsic image information. Image enhancement is achieved through the adjustment of brightness and contrast parameters to establish optimal white balance. Through extensive research efforts globally, scholars have continuously refined enhancement theories based on retinex principles. This ongoing optimization process has led to the development of three principal retinex image enhancement algorithms, with their evolutionary progression illustrated in Figure 2.
The single-scale retinex (SSR) algorithm implements image decomposition by separating illumination and reflection components. The core mechanism involves suppressing or eliminating low-frequency components through a methodological approach that enables the reflection image to effectively capture image information. This process is followed by white balance adjustment to enhance the image’s salient features. To achieve this objective, a Gaussian function [35] is employed as the filtering mechanism for the input image. This function serves as an effective normalized low-pass filter, which is specifically designed to filter out the illumination component. The operational principle is structured as follows:
S ( x , y ) = L ( x , y ) R ( x , y )
In the equation, S ( x , y ) represents the original image information, L ( x , y ) represents the illumination component, and R ( x , y ) represents the reflection component.
To facilitate the separation of illumination and reflectance components while optimizing computational efficiency, logarithmic transformation is applied to both sides of the equation, resulting in the following:
log ( S ( x , y ) ) = log ( L ( x , y ) ) + log ( R ( x , y ) )
Introducing Gaussian filtering to compute the illumination component L ( x , y ) , the filtering is computed as follows:
L ( x , y ) = S ( x , y ) G ( x , y )
The illumination component in the original image is finally obtained through Gaussian filtering. The reflection image is derived by filtering the original image to remove the illumination image. White balancing is then applied to the resulting reflection image to enhance low-light images. The computational formulas are as follows:
log ( R ( x , y ) ) = log ( S ( x , y ) ) log [ S ( x , y ) G ( x , y ) ]
The SSR feature enhancement process follows a systematic sequence: initially, logarithmic transformation is applied to the original image function to separate the illumination and reflectance components. The illumination component L ( x , y ) is then extracted through Gaussian blurring, utilizing a Gaussian low-pass filter to isolate low-frequency components. The reflectance component R ( x , y ) is subsequently obtained by subtracting the illumination component from the original image function, ultimately facilitating feature enhancement in low-light conditions.
According to Equation (4), it is evident that for SSR, there exists only one adjustable parameter σ . From the Gaussian curve, it is known that when σ is large, the Gaussian image changes smoothly, which can achieve better color recovery but may lose a significant amount of edge information. Conversely, when σ is small, the Gaussian image is steeper, preserving better edge information but risking color information loss and distortion. Therefore, when using the SSR algorithm for enhancing low-light images, it is necessary to continuously adjust the parameter σ based on environmental changes and specific requirements, significantly increasing workload and reducing efficiency. Thus, there is a need to seek an algorithm capable of balancing image details smoothly while maintaining color fidelity.
To address the issues with the SSR algorithm, Jobson [36] proposed the multi-scale retinex (MSR) algorithm in 1996, building upon SSR’s foundations. This algorithm adopts a weighted summation approach, separating the image information into three channels (R, G, B). It enhances each channel independently and then merges the enhanced images to produce the final low-light feature-enhanced image.
The MSR algorithm can be viewed as a weighted average of multiple SSR algorithms, assigning specific weights to each filtering channel. Ultimately, it achieves feature enhancement through weighted fusion. Its principle is as follows:
r i ( x , y ) = i = 1 N ω k ( log ( S i ( x , y ) ) log ( S i ( x , y ) G k ( x , y ) ) )
In the equation, N represents the number of scales; ω k denotes the weight assigned to each scale. G k ( x , y ) in this context represents the Gaussian filtering function corresponding to the k-th channel, expressed as follows:
G k ( x , y ) = 1 2 π σ k exp ( x 2 + y 2 2 σ k 2 )
In the equation, σ k represents the standard deviation of the Gaussian at the k scale, also known as the scale parameter. The values of each Gaussian function channel can be adjusted as needed to achieve the experimental requirements. Typically, the R, G, and B channels are commonly chosen for this purpose.
In the MSR algorithm, the original image undergoes three separate SSR enhancement operations in the R, G, and B channels. Different Gaussian blur parameters σ are chosen as needed for SSR enhancement in each channel. Subsequently, the outputs from the three channels are weighted and fused to obtain the low-light enhanced image.
While the MSR algorithm enhances low-light images by balancing detail and color information, improving brightness and contrast to a certain extent, it can disrupt the coupling between the original image’s R, G, and B channels. After MSR feature enhancement, the values of the R, G, and B channels in the original image no longer match those in the enhanced image, resulting in color distortion in the MSR-enhanced image.
Jobson [36] introduced the multi-scale retinex with color restoration (MSRCR) algorithm to address color distortion issues that may arise from the multi-scale retinex (MSR) method. The principle behind MSRCR involves the introduction of a color restoration factor.
r i ( x , y ) = i = 1 N c i ( x , y ) ω k ( log ( S i ( x , y ) ) log ( S i ( x , y ) G k ( x , y ) ) )
In the equation, c i ( x , y ) denotes the color restoration factor for the (i)-th channel, expressed as follows:
c i ( x , y ) = β ( log ( α S ( x , y ) ) log ( i ( r , g , b ) S i ( x , y ) ) )
In the equation, α and β represent the gain factors.
In the MSRCR algorithm, first, a Gaussian function is applied for low-pass filtering to remove illumination components. Then, multi-scale retinex (MSR) enhancement is performed to obtain the MSR image. Finally, the color restoration factor is multiplied element-wise with the MSR-enhanced image to recover color information, addressing color distortion issues in the MSR algorithm.

2.2. SuperGlue Feature Matching Algorithm

SuperGlue is a feature matching network proposed by Sarlin [20] in 2020. It takes as input the keypoints and descriptors of two images, which are provided by feature extraction algorithms and deep learning networks. The SuperGlue matching network consists of an attentional graph neural network and an optimal matching layer, integrating feature point positions and visual descriptors extracted by the SuperPoint feature extraction network. The attentional graph neural network encodes keypoints and descriptors into a feature matching vector using a keypoint encoder, which is iteratively aggregated via attention mechanisms to enhance the feature matching vector. The enhanced feature matching vector serves as input to the optimal matching layer, where a score matrix is computed and the optimal feature assignment matrix is iteratively solved using the Sinkhorn algorithm, as shown in Figure 3. Within the attentional graph neural network, feature aggregation is performed using encoders followed by information aggregation facilitated by attention mechanisms.
The attention graph neural network [20] integrates attention mechanisms and convolutional neural networks, and incorporates SuperPoint for feature point extraction.
The aggregation of feature points using a multi-layer perceptron (MLP) aggregates both the positions and descriptors of feature points. This aggregation allows subsequent attention mechanisms to fully consider both appearance and spatial similarity. The formula is
( 0 ) x i = d i + M L P e n c ( P i )
The addition of descriptors and encoded feature points serves as the input to the attention graph neural network.
The attention graph neural network is based on attention mechanisms and principles of human recognition. It iteratively aggregates features within and between graphs using self-attention and cross-attention. Self-attention is employed for intra-graph aggregation, consolidating neighboring and similar feature points within the graph. Cross-attention is utilized to aggregate inter-graph information, identifying relevant points for matching across graphs.
The optimal matching layer [20] is the second major module of the SuperGlue matching network. It takes the feature descriptors obtained from the attentional GNN and inputs them into the optimal matching layer network. It computes a score matrix by solving the shortest path problem and assigns 0s and 1s to achieve feature matching. This matching network calculates a score matrix R M × N to represent potential matches. However, creating separate representations for M × N potential matches is computationally intensive. Therefore, the authors use the inner product of f i A and f i B aggregated by GNN to compute scores. The formula is as follows:
S i , j = < f i A , f i B >
To address occlusion and visibility issues, this paper introduces a garbage bin to collect unmatched points and mismatched points. Finally, the optimal transport problem using the Sinkhorn algorithm is employed to solve the matching problem. The Sinkhorn [37] distance is based on the Earth Mover’s Distance (EMD).
The SuperPoint network [38] is a network designed for feature point extraction and descriptor generation, consisting of two main components: an encoder and a decoder. The architecture diagram of the SuperPoint network is shown in Figure 4.
The encoder is a simple VGG model that takes the original image H W 1 as input, passes it through convolutional layers, pooling layers, and nonlinear activation functions, and outputs the image H / 8 W / 8 128 . The encoder aims to achieve dimensionality reduction through spatial pooling.
The keypoint decoder network decodes by detecting interest points and outputs the probability that a pixel is a keypoint. After dimension reduction by the encoder, it outputs an H / 8 W / 8 128 dimensional tensor. At this point, upsampling is required to restore it to full resolution. Upsampling increases computational costs, so a keypoint detection head is designed to compute x R H x × W c × 65 , and output R H × W represents the size, with the last dimension discarded during the softmax process.
The descriptor head computes x R H x × W c × 65 and outputs R H × W . Using a model similar to UCN, it takes semi-dense descriptors as network inputs. These descriptors undergo bicubic interpolation followed by L2 regularization, resulting in a dense mapping of L2-normalized fixed-length descriptors.

3. Results

3.1. Data Source

The experimental dataset selected in this paper required both camera internal and external parameters and ultra-dark light conditions. However, existing datasets rarely include camera internal and external parameters. Matching datasets are mostly based on high-resolution images, which cannot meet the ultra-dark light conditions. Enhanced datasets also make it difficult to form matching pairs and lack camera internal and external parameters. Therefore, this paper selected the ScanNet indoor dataset that includes camera internal and external parameters, but it still did not meet the ultra-dark light conditions. To meet the experimental requirements, the image exposure was manually adjusted to obtain darker images, with the experimental dataset illustrated in Figure 5.
Analysis of the dataset demonstrates several inherent challenges in low-light images. Specifically, these images exhibit blurred texture information and indistinct features, accompanied by minimal grayscale variation in the vicinity of feature points. These characteristics significantly compromise feature recognition capabilities and impede effective feature point extraction. To overcome these limitations, this study proposes an initial preprocessing step focusing on feature enhancement of low-light images, with particular emphasis on optimizing brightness and contrast parameters.

3.2. Image Enhancement Experiment

To explore the impact of different parameter configurations of the MSRCR algorithm on image matching results, this paper presents comparative experiments to verify the influence of different MSRCR parameters on matching results. Through experiments, it was found that changing MSRCR different parameters has a negligible effect on matching results. Therefore, this paper selected a set of parameter settings with the best performance for experiments, where s i g m a = 300 , s i g m a _ l i s t = [ 15 , 80 , 250 ] , G = 5.0 , b = 25.0 , a l p h a = 125.0 , b e t a = 46.0 , l o w _ c l i p = 0.01 , and h i g h _ c l i p = 0.99 . s i g m a represents the Gaussian filtering parameter, s i g m a _ l i s t represents the scale parameters, G is the gain parameter, b is used to adjust the overall image brightness, a l p h a represents the detail enhancement coefficient, b e t a represents the smoothing control coefficient, and l o w _ c l i p and h i g h _ c l i p represent the lower and upper truncation thresholds. The original low-light images were used as input to the MSRCR algorithm for feature enhancement, improving image brightness and contrast, and enhancing the texture information of the images, as illustrated in Figure 6.
Qualitative analysis of the enhanced images demonstrates that the MSRCR preprocessing significantly improved both photometric and geometric properties of the scenes. Specifically, the enhancement resulted in not only optimized brightness and contrast but also distinctly visible texture features, effectively restoring the intrinsic feature information within the images. This enhancement facilitates more robust keypoint extraction in subsequent processing stages.

3.3. Experiment of Matching Before and After Enhancement

The enhanced and original low-light images were separately input into the SuperGlue algorithm for feature matching, and the matching results are shown in Figure 7.
Comparative analysis between pre- and post-enhancement feature matching results is presented, with the left and right images illustrating the matching outcomes before and after MSRCR enhancement, respectively. Quantitative evaluation across six experimental groups demonstrates that MSRCR enhancement yields substantial improvements in multiple performance metrics. Specifically, the enhanced images exhibit an increased number of detected feature points, reduced rotational and translational errors, and significantly improved matching performance, characterized by both higher participation rates and increased accuracy of feature point correspondences.
Using the x-axis to represent SuperGlue and MSRCR+SuperGlue, and the y-axis to represent the number of correctly matched feature points and the correct matching rate, the density distribution plot is drawn as shown in Figure 8 and Figure 9.
As shown in Figure 8 and Figure 9, the proposed method significantly improves both the number of correctly matched feature points and matching accuracy, solving the issues of low matching quantity and low precision encountered by existing feature matching algorithms such as SuperGlue, Matchformer, LightGlue, LoFTR, and OmniGlue when facing extremely challenging ultra-low-light image matching.

3.4. Accuracy Assessment

This study investigates the efficacy of integrating MSRCR enhancement with the SuperGlue matching framework. The performance evaluation of the proposed methodology encompasses three critical metrics: (1) the comparative analysis of feature point extraction density pre- and post-enhancement, (2) the quantification of rotational and translational error metrics, and (3) the assessment of matching performance in terms of both quantity and precision of correspondences.
The MSRCR algorithm effectively enhances the photometric properties of low-light images through simultaneous optimization of brightness and contrast parameters. This enhancement significantly improves the visibility of intrinsic image details, thereby facilitating more robust feature point extraction.
According to Table 1, from the data of six experimental groups, it can be observed that after MSRCR feature enhancement, there is a significant increase in grayscale variation around feature points, making feature point extraction easier. The number of extracted feature points has increased significantly, to more than 1.5 times that of the feature points extracted from the original low-light images.
As shown in Figure 10, after applying MSRCR feature enhancement to the original low-light images, the number of feature points extracted from both left and right images increased significantly.
The experimental methodology incorporates both image pairs and their corresponding ground truth data, which specify the relative translational and rotational transformations between stereo images. The algorithmic computation of rotation and translation parameters is evaluated against these ground truth values, with the respective discrepancies quantified as ΔR and Δt.
According to the data analysis in Table 2, it can be observed that after enhancement with MSRCR, the relative pose error decreases. In experiments 5 and 6, although the rotation and translation errors cannot be calculated directly from the original image pairs, after feature enhancement, the rotation and translation errors of the matching image pairs can still be determined.
To calculate the matching score, CM represents the number of correctly matched feature points, TM represents the total number of feature points involved in matching, and Mscore represents the matching score.
Analysis of Table 3 and Figure 11 demonstrates that MSRCR feature enhancement yields substantial improvements in both the absolute number of matches and correct correspondence rates across the dataset’s low-light images. The enhancement methodology exhibits particularly notable efficacy in cases of severe degradation, as evidenced by experimental groups 5 and 6, where initial matching attempts yielded no correspondences but post-enhancement processing produced satisfactory matching results. For visualization purposes in the precision analysis, the matching score for the fifth group was normalized to zero in the corresponding line graph.
Quantitative analysis of Figure 12 demonstrates that MSRCR feature enhancement significantly improves matching precision metrics. The proposed methodology exhibits robust performance even under challenging conditions where conventional matching approaches fail, particularly in cases of severely degraded low-light images characterized by poor focus quality. While traditional methods prove ineffective for these degraded images, the proposed MSRCR-enhanced matching framework successfully establishes valid feature correspondences.
We conducted experiments comparing our proposed method with existing advanced matching algorithms LightGlue, LoFTR, and OmniGlue, performing six sets of experiments using the same low-light dataset. In the original OmniGlue algorithm, the threshold was 0.02. To maintain consistency, we set the thresholds for all algorithms to 0.2, which is consistent with the threshold used in our method, and performed statistical analysis on the number of correctly matched feature points, total matching quantity, and matching accuracy.
From Table 4 and Figure 13’s matching accuracy density distribution plot, it can be observed that although existing advanced matching algorithms like LightGlue, LoFTR, and OmniGlue demonstrate strong advantages in well-illuminated high-resolution images with satisfactory matching quantities and precision, they struggle significantly when dealing with noisy ultra-low-light images. In such challenging conditions, these algorithms fail to meet industrial production requirements, exhibiting low matching quantities and poor matching accuracy. In contrast, the proposed method combining MSRCR and SuperGlue shows significant improvements in both matching quantity and accuracy after feature enhancement, clearly outperforming existing algorithms.

4. Discussion

This paper addresses issues such as slow grayscale variations around image feature points in low-light and extremely low-light environments, unclear local features, difficulties in feature point extraction, and low matching accuracy. A novel matching method is proposed that combines the MSRCR feature enhancement algorithm based on retinex with the SuperGlue feature matching algorithm to achieve feature matching under extremely low-light conditions. Building upon the SuperGlue algorithm proposed by Sarlin P. E. et al. in 2020 [20], this paper introduces a feature enhancement module that enhances features prior to matching, thereby addressing the issues of unclear local features and difficult feature point extraction in low-light images. The enhanced low-light images serve as input to the SuperGlue feature matching algorithm, effectively resolving the problems of insufficient feature points and low matching accuracy in low-light image matching. The proposed method consists of two modules: feature enhancement and feature matching. In the feature enhancement module, the MSRCR algorithm based on retinex theory is employed, which decomposes the original image into illumination and reflection components, mimicking human visual perception. By filtering out the illumination component using Gaussian filtering and performing white balancing on the remaining reflection image, the brightness and contrast of the image are enhanced, restoring the true information of objects on the ground and facilitating feature point detection and extraction. In the feature matching module, the SuperGlue algorithm is implemented, which consists of an attention graph neural network and an optimal matching layer. Throughout the matching process, SuperPoint is used to extract feature points and descriptors, which are then encoded into a feature matching vector by an encoder. The attention graph neural network enhances this feature matching vector iteratively using self-attention and cross-attention mechanisms. Subsequently, the optimal matching layer computes matching scores and iteratively uses the Sinkhorn algorithm to determine the optimal feature assignment matrix. By integrating the feature enhancement and feature matching modules, this paper enhances local features by restoring the true surface information of objects using the enhancement algorithm. The SuperGlue feature matching algorithm improves matching accuracy and increases the number of matching feature points, thereby addressing feature matching issues in extremely low-light environments.
Firstly, six pairs of low-light matching images were subjected to feature enhancement experiments. Figure 11 shows the local detailed information before and after enhancement of three of these low-light images. The image on the left is the enhanced low-light image, and the image on the right is the original low-light image. The experimental results indicate the following:
In the original extremely low-light images, texture features exhibit significant attenuation, with predominant dark tonality impeding both object identification and feature point extraction. Following MSRCR enhancement, the algorithm effectively eliminates illumination component interference while preserving crucial reflection information. This enhancement process significantly amplifies detail features and introduces distinct grayscale gradients around feature points, facilitating clear object recognition and surface texture delineation, thus substantially improving the robustness of feature point extraction.
As shown in Figure 14, after enhancement with MSRCR, the brightness and contrast of the images are improved. Gaussian filtering was used to remove the illumination components from the original low-light images, followed by white balance adjustment to enhance the brightness and contrast of the images, making it possible to clearly see the surface texture features of objects.
To validate the effectiveness of the proposed method, six sets of comparative experiments were conducted, with green indicators denoting correct matches and red indicators representing incorrect matches. As illustrated in Figure 15, the analysis compared rotational and translational differences between photographic pairs using both ground truth values and computer-generated matches. Prior to feature enhancement, the system exhibited a rotation error of 40.5° and a translation error of 62.8. Post-enhancement analysis revealed significantly improved performance, with rotation error reducing to 1.0° and translation error to 1.5, representing approximately 1/40 of the original matching errors. The enhancement process demonstrated substantial improvement in matching accuracy, with computed rotational and translational components closely aligning with ground truth values. Quantitative analysis of matching performance revealed that pre-enhancement results yielded 97 correct matches out of 159 total correspondences, corresponding to a matching score of 0.61. Following enhancement, the system achieved 241 correct matches out of 246 total correspondences, resulting in a matching score of 0.98, representing a 27% improvement in matching accuracy. The results demonstrate that feature enhancement not only increased the absolute number of feature point correspondences but also substantially improved the precision of these matches.
As demonstrated in Figure 16, when processing low-light images characterized by poor quality and blurriness, traditional feature matching algorithms, while capable of extracting feature points, generate low-quality correspondences that result in unsuccessful feature matching. This failure prevents the computation of relative rotation and translation parameters between left and right image patches, making error calculation impossible. The proposed methodology employs MSRCR feature enhancement to amplify texture information and improve image brightness and contrast before feeding the enhanced images into the SuperGlue matching algorithm. Experimental results reveal a rotation error of 36.0°, translation error of 40.0, and 36 correct matches out of 42 total correspondences, achieving a matching score of 0.86. Although the matching accuracy is constrained by the inherent low quality and blurriness of the image pairs, the proposed algorithm successfully establishes feature correspondences in scenarios where conventional matching algorithms fail completely.
However, despite the proposed method’s capability to enhance feature point correspondence quantity and matching accuracy, significant challenges persist when processing low-quality, blurry low-light images. While the algorithm successfully establishes matches in such conditions, the inherent image blurriness continues to limit both the quantity and accuracy of feature correspondences. The challenge of improving matching quality for low-quality low-light images remains a substantial research concern. Furthermore, the current implementation maintains MSRCR and SuperGlue as separate algorithmic components rather than integrating them into a unified enhanced matching framework. Future research should focus on developing a cohesive algorithm that seamlessly integrates these complementary approaches into a single enhanced matching solution.

5. Conclusions

This paper addresses the issues of indistinct features, difficult feature point extraction, and low matching accuracy in dark images by proposing a feature matching method that combines MSRCR and SuperGlue. Through comparative experiments before and after dark image enhancement across six groups, the effectiveness of the proposed method was validated. After MSRCR feature enhancement, the method enhanced the texture features of dark images, enabling the extraction of more high-quality feature points, with a significant increase in both correct matches and total matches, and a notable improvement in matching accuracy. The proposed feature matching method combining MSRCR and SuperGlue solves the problems of low matching quantity and accuracy in dark images, meets the operational requirements of visual SLAM in extreme environments, enhances the universality of visual SLAM, provides a new technical approach for nighttime visual navigation and localization, and simultaneously opens up new research directions for the application of intelligent perception systems in complex environments.

Author Contributions

Conceptualization, Y.C. and F.L.; Methodology, Y.C. and F.L.; Software, Q.S.; Validation, F.L., Q.S. and G.S.; Formal analysis, F.L.; Investigation, G.S. and H.Y.; Resources, Y.C. and H.Y.; Data Curation, F.L. and H.Y.; Writing—original draft preparation, F.L.; Writing—review and editing, Y.C., G.S. and J.N.; Visualization, F.L. and G.S.; Supervision, Y.C.; Funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

Higher Education Teaching Reform Research Project of Nanjing Tech University, 2023, Project No.20230073; this work was also supported by the Joint Fund of Henan Collaborative Innovation Center of Geo-Information Technology for Smart Central Plains and Key Laboratory of Spatiotemporal Information Perception and Fusion Technology, Ministry of Land and Resources, No. 212108; this work was also supported by the Postgraduate Research & Practice Innovation Program of Jiangsu Province, No. SJCX24_0511.

Data Availability Statement

The data that support the findings of this study are available from [https://github.com/ScanNet/ScanNet, accessed on 5 March 2024] but restrictions apply to the availability of these data, which were used under license for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of [http://www.scan-net.org/, accessed on 5 March 2024].

Acknowledgments

Fengchao Li was an undergraduate student at the School of Geomatics Science and Technology, Nanjing Tech University when this research was conducted, and is currently pursuing graduate studies at Information Engineering University. Yu Chen, the corresponding author of this paper, served as Fengchao Li’s academic supervisor during this research.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Khan, A.A.; Laghari, A.A.; Awan, S.A. Machine learning in computer vision: A review. EAI Endorsed Trans. Scalable Inf. Syst. 2021, 8, e4. [Google Scholar] [CrossRef]
  2. Zhao, Y.; Yan, L.; Chen, Y.; Dai, J.; Liu, Y. Robust and efffcient trajectory replanning based on guiding path for quadrotor fast autonomous flight. Remote Sens. 2021, 13, 972. [Google Scholar] [CrossRef]
  3. Bijjahalli, S.; Sabatini, R.; Gardi, A. Advances in intelligent and autonomous navigation systems for small UAS. Prog. Aerosp. Sci. 2020, 115, 100617. [Google Scholar] [CrossRef]
  4. Yasuda, Y.D.V.; Martins, L.E.G.; Cappabianco, F.A.M. Autonomous visual navigation for mobile robots: A systematic literature review. ACM Comput. Surv. (CSUR) 2020, 53, 1–34. [Google Scholar] [CrossRef]
  5. Xu, B.; Chen, Y.; Zhang, S.; Wang, J. Improved Point–Line Visual–Inertial Odometry System Using Helmert Variance Component Estimation. Remote Sens. 2020, 12, 2901. [Google Scholar] [CrossRef]
  6. Xu, B.; Li, X.; Wang, J.; Yuen, C.; Li, J. PVI-DSO: Leveraging Planar Regularities for Direct Sparse Visual-Inertial Odometry. IEEE Sens. J. 2023, 23, 17415–17425. [Google Scholar] [CrossRef]
  7. Zhou, T.; Yang, M.; Jiang, K.; Wong, H.; Yang, D. MMW radar-based technologies in autonomous driving: A review. Sensors 2020, 20, 7283. [Google Scholar] [CrossRef]
  8. Cui, Y.; Chen, R.; Chu, W.; Chen, L.; Tian, D.; Li, Y.; Cao, D. Deep learning for image and point cloud fusion in autonomous driving: A review. IEEE Trans. Intell. Transp. Syst. 2021, 23, 722–739. [Google Scholar] [CrossRef]
  9. Liu, L.; Lu, S.; Zhong, R.; Wu, B.; Yao, Y.; Zhang, Q.; Shi, W. Computing systems for autonomous driving: State of the art and challenges. IEEE Internet Things J. 2020, 8, 6469–6486. [Google Scholar] [CrossRef]
  10. Zhao, Y.; Vela, P.A. Good feature matching: Toward accurate, robust vo/vslam with low latency. IEEE Trans. Robot. 2020, 36, 657–675. [Google Scholar] [CrossRef]
  11. Yu, S.; Fu, C.; Gostar, A.K.; Hu, M. A review on map-merging methods for typical map types in multiple-ground-robot SLAM solutions. Sensors 2020, 20, 6988. [Google Scholar] [CrossRef] [PubMed]
  12. Zhao, Z.; Xiong, B.; Wang, L.; Ou, Q.; Yu, L.; Kuang, F. RetinexDIP: A uniffed deep framework for low-light image enhancement. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 1076–1088. [Google Scholar] [CrossRef]
  13. Liu, X.; Zhang, C.; Wang, Y.; Ding, K.; Han, T.; Liu, H.; Tian, Y.; Xu, B.; Ju, M. Low light image enhancement based on multi-scale network fusion. IEEE Access 2022, 10, 127853–127862. [Google Scholar] [CrossRef]
  14. Huang, W.; Li, K.; Xu, M.; Huang, R. Self-Supervised Non-Uniform Low-Light Image Enhancement Combining Image Inversion and Exposure Fusion. Electronics 2023, 12, 4445. [Google Scholar] [CrossRef]
  15. Qiu, Y.; Lu, Y.; Wang, Y.; Jiang, H. IDOD-YOLOV7: Image-dehazing YOLOV7 for object detection in low-light foggy trafffc environments. Sensors 2023, 23, 1347. [Google Scholar] [CrossRef]
  16. Han, Y.; Chen, X.; Zhong, Y.; Huang, Y.; Li, Z.; Han, P.; Li, Q.; Yuan, Z. Low-illumination road image enhancement by fusing retinex theory and histogramequalization. Electronics 2023, 12, 990. [Google Scholar] [CrossRef]
  17. Bellavia, F. SIFT matching by context exposed. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 2445–2457. [Google Scholar] [CrossRef]
  18. Fu, Q.; Yu, H.; Wang, X.; Yang, Z.; He, Y.; Zhang, H.; Mian, A. Fast ORB-SLAM without keypoint descriptors. IEEE Trans. Image Process. 2021, 31, 1433–1446. [Google Scholar] [CrossRef]
  19. Bansal, M.; Kumar, M.; Kumar, M. 2D object recognition: A comparative analysis of SIFT, SURF and ORB featuredescriptors. Multimed. Tools Appl. 2021, 80, 18839–18857. [Google Scholar] [CrossRef]
  20. Sarlin, P.E.; DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superglue: Learning feature matching with graph neural net-works. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4938–4947. [Google Scholar] [CrossRef]
  21. Chen, Y.; Yan, L.; Xu, B.; Liu, Y. Multi-Stage Matching Approach for Mobile Platform Visual Imagery. IEEE Access 2019, 7, 160523–160535. [Google Scholar] [CrossRef]
  22. Wang, B.; Zhang, X.; Wang, W. Feature matching method based on SURF and fast library for approximatenearest neighbor search. Integr. Ferroelectr. 2021, 218, 147–154. [Google Scholar] [CrossRef]
  23. Sun, J.; Shen, Z.; Wang, Y.; Bao, H.; Zhou, X. LoFTR: Detector-free local feature matching with transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 8922–8931. [Google Scholar] [CrossRef]
  24. Wang, Q.; Zhang, J.; Yang, K.; Peng, K.; Stiefelhagen, R. Matchformer: Interleaving attention in transformers for feature matching. In Proceedings of the Asian Conference on Computer Vision, Macao, China, 4–8 December 2022; pp. 2746–2762. [Google Scholar] [CrossRef]
  25. Li, J.; Xu, W.; Shi, P.; Zhang, Y.; Hu, Q. LNIFT: Locally normalized image for rotation invariant multimodal feature matching. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5621314. [Google Scholar] [CrossRef]
  26. Lindenberger, P.; Sarlin, P.E.; Pollefeys, M. Lightglue: Local feature matching at light speed. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 4–6 October 2023; pp. 17627–17638. [Google Scholar]
  27. Zhang, Y.; Lan, C.; Zhang, H.; Ma, G.; Li, H. Multimodal remote sensing image matching via learning features and attention mechanism. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5603620. [Google Scholar] [CrossRef]
  28. Zhu, B.; Ye, Y.; Dai, J.; Peng, T.; Deng, J.; Zhu, Q. VDFT: Robust feature matching of aerial and ground images using viewpoint-invariant deformable feature transformation. ISPRS J. Photogramm. Remote Sens. 2024, 218, 311–325. [Google Scholar] [CrossRef]
  29. Jiang, H.; Karpur, A.; Cao, B.; Huang, Q.; Araujo, A. OmniGlue: Generalizable Feature Matching with Foundation Model Guidance. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 16–22 June 2024; pp. 19865–19875. [Google Scholar]
  30. Kazerouni, I.A.; Fitzgerald, L.; Dooly, G.; Toal, D. A survey of state-of-the-art on visual SLAM. Expert Syst. Appl. 2022, 205, 117734. [Google Scholar] [CrossRef]
  31. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10561–10570. [Google Scholar]
  32. Wu, W.; Weng, J.; Zhang, P.; Wang, X.; Yang, W.; Jiang, J. Uretinex-net: Retinex-based deep unfolding network for low-light image enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5901–5910. [Google Scholar]
  33. Hai, J.; Xuan, Z.; Yang, R.; Hao, Y.; Zou, F.; Lin, F.; Han, S. R2rnet: Low-light image enhancement via real-low to real-normal network. J. Vis. Commun. Image Represent. 2023, 90, 103712. [Google Scholar] [CrossRef]
  34. Wu, J.; Ding, B.; Zhang, B.; Ding, J. A Retinex-based network for image enhancement in low-light environments. PLoS ONE 2024, 19, e0303696. [Google Scholar] [CrossRef]
  35. Ma, J.; Zhou, H.; Zhao, J.; Gao, Y.; Jiang, J.; Tian, J. Robust feature matching for remote sensing image registration via linear adaptive fltering. IEEE Trans. Geosci. Remote Sens. 2020, 59, 1577–1591. [Google Scholar] [CrossRef]
  36. Rahman, Z.; Jobson, D.J.; Woodell, G.A. Multi-scale retinex for color image enhancement. In Proceedings of the 3rd IEEE International Conference on Image Processing, Lausanne, Switzerland, 16–19 September 1996; IEEE: New York, NY, USA, 1996; Volume 3, pp. 1003–1006. [Google Scholar] [CrossRef]
  37. Carlier, G. On the linear convergence of the multimarginal Sinkhorn algorithm. SIAM J. Optim. 2022, 32, 786–794. [Google Scholar] [CrossRef]
  38. DeTone, D.; Malisiewicz, T.; Rabinovich, A. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 224–236. [Google Scholar] [CrossRef]
Figure 1. Enhanced matching algorithm framework.
Figure 1. Enhanced matching algorithm framework.
Remotesensing 17 00905 g001
Figure 2. The development process of the retinex feature enhancement algorithm.
Figure 2. The development process of the retinex feature enhancement algorithm.
Remotesensing 17 00905 g002
Figure 3. SuperGlue network architecture.
Figure 3. SuperGlue network architecture.
Remotesensing 17 00905 g003
Figure 4. SuperPoint network architecture.
Figure 4. SuperPoint network architecture.
Remotesensing 17 00905 g004
Figure 5. Original low-light images.
Figure 5. Original low-light images.
Remotesensing 17 00905 g005
Figure 6. Enhanced images.
Figure 6. Enhanced images.
Remotesensing 17 00905 g006
Figure 7. Comparison of matching experiments before and after enhancement. The green lines indicate correct matches, while the red lines indicate incorrect matches.
Figure 7. Comparison of matching experiments before and after enhancement. The green lines indicate correct matches, while the red lines indicate incorrect matches.
Remotesensing 17 00905 g007
Figure 8. Density distribution plot of correctly matched feature points.
Figure 8. Density distribution plot of correctly matched feature points.
Remotesensing 17 00905 g008
Figure 9. Matching accuracy density distribution plot.
Figure 9. Matching accuracy density distribution plot.
Remotesensing 17 00905 g009
Figure 10. Number of extracted feature points. BEL represents the number of feature points extracted from the left image before MSRCR enhancement, BER represents the number of feature points extracted from the right image before enhancement, AEL represents the number of feature points extracted from the left image after enhancement, and AER represents the number of feature points extracted from the right image after enhancement.
Figure 10. Number of extracted feature points. BEL represents the number of feature points extracted from the left image before MSRCR enhancement, BER represents the number of feature points extracted from the right image before enhancement, AEL represents the number of feature points extracted from the left image after enhancement, and AER represents the number of feature points extracted from the right image after enhancement.
Remotesensing 17 00905 g010
Figure 11. Number of matches. CM represents the number of correctly matched feature points before enhancement, TM represents the total number of matched feature points before enhancement, MS(CM) represents the number of correctly matched feature points after MSRCR enhancement, and MS(TM) represents the total number of feature point matches after MSRCR enhancement.
Figure 11. Number of matches. CM represents the number of correctly matched feature points before enhancement, TM represents the total number of matched feature points before enhancement, MS(CM) represents the number of correctly matched feature points after MSRCR enhancement, and MS(TM) represents the total number of feature point matches after MSRCR enhancement.
Remotesensing 17 00905 g011
Figure 12. Accuracy comparison diagram before and after feature enhancement.
Figure 12. Accuracy comparison diagram before and after feature enhancement.
Remotesensing 17 00905 g012
Figure 13. Precision Density Distribution Plots of the Proposed Method Compared with LightGlue, LoFTR, and OmniGlue.
Figure 13. Precision Density Distribution Plots of the Proposed Method Compared with LightGlue, LoFTR, and OmniGlue.
Remotesensing 17 00905 g013
Figure 14. Details before and after feature enhancement.
Figure 14. Details before and after feature enhancement.
Remotesensing 17 00905 g014
Figure 15. Enhanced feature matching results before and after enhancement.
Figure 15. Enhanced feature matching results before and after enhancement.
Remotesensing 17 00905 g015
Figure 16. Feature matching results before and after enhancing blurry low-light images.
Figure 16. Feature matching results before and after enhancing blurry low-light images.
Remotesensing 17 00905 g016
Table 1. The number of feature points extracted before and after enhancement.
Table 1. The number of feature points extracted before and after enhancement.
Experimental IndexBefore EnhancementAfter Enhancement
Left ImageRight ImageLeft ImageRight Image
15242171024388
2275234506455
3430405665704
4355289538617
5441181813438
659167289240
Table 2. Relative rotation and translation error.
Table 2. Relative rotation and translation error.
Experimental IndexSuperGlueMSRCR+SuperGlue
ΔRΔtΔRΔt
1errorerror36.0°40.0
22.7°1.72.9°0.8
340.5°62.81.0°1.5
411.3°13.96.1°4.1
56.2°11.52.0°3.5
6errorerror55.1°82.5
Table 3. Matching accuracy.
Table 3. Matching accuracy.
Experimental IndexSuperGlueMscoreMSRCR+SuperGlueMscore
CMTMCMTM
102036420.86
261680.91181210.98
3971590.612412460.98
442480.881031060.97
537590.6374930.80
600/570.71
Table 4. Experimental results of the proposed method compared with existing algorithms LightGlue, LoFTR, and OmniGlue.
Table 4. Experimental results of the proposed method compared with existing algorithms LightGlue, LoFTR, and OmniGlue.
AlgorithmExperiment NumberCMTMMscore
Ours12412460.98
274930.8
31181210.98
41031060.97
5570.71
636420.86
LightGlue12150.13
23120.25
3000
4000
5050
6020
LoFTR1170.14
217670.25
3000
4010
5000
6000
OmniGlue1000
2000
3000
4000
5000
6000
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, F.; Chen, Y.; Shi, Q.; Shi, G.; Yang, H.; Na, J. Improved Low-Light Image Feature Matching Algorithm Based on the SuperGlue Net Model. Remote Sens. 2025, 17, 905. https://doi.org/10.3390/rs17050905

AMA Style

Li F, Chen Y, Shi Q, Shi G, Yang H, Na J. Improved Low-Light Image Feature Matching Algorithm Based on the SuperGlue Net Model. Remote Sensing. 2025; 17(5):905. https://doi.org/10.3390/rs17050905

Chicago/Turabian Style

Li, Fengchao, Yu Chen, Qunshan Shi, Ge Shi, Hongding Yang, and Jiaming Na. 2025. "Improved Low-Light Image Feature Matching Algorithm Based on the SuperGlue Net Model" Remote Sensing 17, no. 5: 905. https://doi.org/10.3390/rs17050905

APA Style

Li, F., Chen, Y., Shi, Q., Shi, G., Yang, H., & Na, J. (2025). Improved Low-Light Image Feature Matching Algorithm Based on the SuperGlue Net Model. Remote Sensing, 17(5), 905. https://doi.org/10.3390/rs17050905

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop