Next Article in Journal
Unmasking the True Identity: Unveiling the Secrets of Virtual Private Networks and Proxies
Next Article in Special Issue
Quantum Edge Detection and Convolution Using Paired Transform-Based Image Representation
Previous Article in Journal
Analysis of Cultural Perceptions of the Intangible Cultural Heritage of Chinese Porcelain Inlay: An Investigation Based on Social Media Data
Previous Article in Special Issue
PLC-Fusion: Perspective-Based Hierarchical and Deep LiDAR Camera Fusion for 3D Object Detection in Autonomous Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Thermal Video Enhancement Mamba: A Novel Approach to Thermal Video Enhancement for Real-World Applications

1
Faculty of Mathematics and Mechanics, Yerevan State University, Yerevan 0025, Armenia
2
Department of Computer Science, School of Engineering, City University of New York (CUNY), New York, NY 10031, USA
3
Department of Electrical and Computer Engineering, School of Engineering, Tufts University, Medford, MA 02155, USA
4
Department of Electrical and Computer Engineering, The University of Texas at San Antonio, San Antonio, TX 78249, USA
*
Author to whom correspondence should be addressed.
Information 2025, 16(2), 125; https://doi.org/10.3390/info16020125
Submission received: 22 December 2024 / Revised: 17 January 2025 / Accepted: 7 February 2025 / Published: 9 February 2025
(This article belongs to the Special Issue Emerging Research in Object Tracking and Image Segmentation)

Abstract

:
Object tracking in thermal video is challenging due to noise, blur, and low contrast. We present TVEMamba, a Mamba-based enhancement framework with near-linear complexity that improves tracking in these conditions. Our approach uses a State Space 2D (SS2D) module integrated with Convolutional Neural Networks (CNNs) to filter, sharpen, and highlight important details. Key components include (i) a denoising module to reduce background noise and enhance image clarity, (ii) an optical flow attention module to handle complex motion and reduce blur, and (iii) entropy-based labeling to create a fully labeled thermal dataset for training and evaluation. TVEMamba outperforms existing methods (DCRGC, RLBHE, IE-CGAN, BBCNN) across multiple datasets (BIRDSAI, FLIR, CAMEL, Autonomous Vehicles, Solar Panels) and achieves higher scores on standard quality metrics (EME, BDIM, DMTE, MDIMTE, LGTA). Extensive tests, including ablation studies and convergence analysis, confirm its robustness. Real-world examples, such as tracking humans, animals, and moving objects for self-driving vehicles and remote sensing, demonstrate the practical value of TVEMamba.

1. Introduction

Visually appealing videos are essential not only for human perception but also for advanced computer vision tasks. Unfortunately, many videos are captured under challenging conditions that cause poor visibility, structural degradation, and unpredictable noise. These issues significantly reduce the performance of automated image analysis systems and object-tracking algorithms used in surveillance [1], monitoring [2], intelligent transportation [3], and remote sensing [4]. Object tracking involves identifying and following objects across visible and thermal video frame sequences. Traditional methods often rely on the Kalman filter, assuming linear motion. However, these methods struggle when objects exhibit complex, non-linear motions or when videos suffer from uneven illumination, motion blur, and noise.
Many algorithms have been developed to enhance videos in the visible spectrum (VIS) [5], yet they still face challenges in varying illumination and low-light conditions. Improving low-light or nighttime imagery requires a deep understanding of how light interacts with the scene and how images are formed. In response, thermal imaging has emerged as a reliable solution, especially in environments where VIS images perform poorly. For real-world applications, sensors that operate in different spectrums are often necessary to accommodate changing lighting conditions. Studies have shown that thermal infrared data can improve the accuracy of object tracking, semantic segmentation, saliency detection, and object detection. Thermal videos are also widely used in wildlife monitoring [6], surveillance [7], military operations [8], security [9], and remote sensing [10]. Nevertheless, they pose their own difficulties, such as low contrast, motion blur, and loss of fine details, making it harder to detect targets or enhance infrared imaging technologies, as shown in Figure 1a.
Thermal Video Enhancement (TVE) techniques aim to improve the visual quality of thermal footage for automated processing tasks, such as analysis, detection, segmentation, and recognition. Existing thermal video enhancement methods can be broadly categorized into three main approaches, each with specific limitations: Single-frame Enhancement Methods: Recent works like MdNet [11] and GAPANet [12] focus on enhancing individual frames using multi-scale feature fusion and attention mechanisms. While these methods effectively improve local contrast and details, they often introduce temporal inconsistencies when applied to videos due to frame-by-frame processing. Additionally, they struggle with motion-induced artifacts since they lack temporal context. Learning-based Hybrid Methods: Recent advances include NSCT-based dual CNN models [13,14] and GAN-ViT [15,16], which combine multiple architectural components to address various enhancement aspects. While these approaches promise to handle noise and improve contrast, they often struggle with computational efficiency and real-time processing requirements. Furthermore, their enhancement quality can be inconsistent across different scenes and thermal conditions. Multi-frame Methods: Approaches, such as Shift-Net [17] and AverNet [18], attempt to leverage temporal information through frame alignment and spatial-temporal modules. However, these methods face challenges in rapid motion or complex thermal patterns. Their reliance on explicit frame alignment can fail when dealing with non-rigid motion or thermal variations, resulting in ghosting artifacts and blur in the enhanced output.
Common limitations across existing methods include the following: (1) Loss of fine details during denoising, particularly in regions with subtle thermal gradients, (2) Inability to handle diverse scenarios with varying thermal patterns and motion complexities, (3) Temporal inconsistencies in enhanced video sequences, and (4) Artifacts and distorted features under challenging conditions (see Table 1).
Overcoming these challenges is crucial to achieve robust performance in real-world applications. One potential direction involves incorporating blur-resistant motion deblurring methods that leverage the inherent properties of thermal scenes, providing more reliable and adaptive enhancement.
Enhancing thermal videos is inherently more complex than working with visible light footage. It involves dealing with low image contrast, sensor noise, rapid motion, and limited spatial resolution, which vary with environmental factors, target types, and imaging devices, as shown in Figure 1. Illumination inconsistencies, camera jitter, and atmospheric effects further complicate the task, demanding algorithms that can adapt to different conditions. Addressing these issues can support various applications, from navigation and safety in autonomous systems to reliable object detection in surveillance under unpredictable environments.
This paper addresses critical challenges in thermal video enhancement, where existing methods often struggle with severe sensor noise, motion artifacts, and loss of fine details. These issues are particularly pronounced in thermal imaging due to inherent sensor limitations and the complex nature of infrared radiation capture. Traditional enhancement approaches, primarily based on CNNs, face limitations in handling the temporal aspects of thermal videos and often produce results with temporal inconsistencies or motion blur. The proposed Mamba-based architecture offers two key advantages. First, its selective state space mechanism effectively models long-range dependencies in spatial and temporal dimensions. This enables our model to process longer video sequences while capturing subtle temporal patterns crucial for consistent enhancement. Second, unlike conventional CNNs focusing on local features, Mamba’s state space formulation allows adaptive feature extraction across varying temporal scales, which is essential for handling diverse motion patterns in thermal videos [19].
We integrate this core architecture with two specialized modules: A Basic Denoising (BD) module for addressing thermal sensor noise and an Optical Flow Attention (OFA) module for precise motion handling. The OFA module targets complex motion patterns by adaptively focusing on motion-relevant regions, ensuring temporal consistency even under challenging conditions like rapid movement or varying thermal gradients. This comprehensive approach enables our method to produce enhanced thermal videos with improved clarity, reduced noise, and preserved temporal coherence, making it particularly suitable for critical applications in surveillance, autonomous navigation, and industrial monitoring. The main contributions are as follows:
  • We introduce a novel Mamba model for thermal video enhancement that integrates the SS2D module with CNNs to handle complex motions and challenging lighting conditions. This model includes:
    (a)
    The Basic Denoising module, which reduces noise and improves image quality.
    (b)
    The Optical Flow Attention module, which provides blur-resistant motion deblurring and preserves scene details even under challenging circumstances.
  • We create a labeled thermal video dataset using entropy-based measures to produce meaningful labels for training and evaluation. This dataset includes over three video sequence pairs, with 4k frame pairs.
  • We evaluate the proposed framework on real-world scenarios like wildlife monitoring and autonomous systems. Our experiments cover diverse thermal video datasets, including BIRDSAI [20], FLIR [21], CAMEL [22], Autonomous Vehicles [23], and Solar Panel [24], each presenting unique challenges. Compared to two traditional methods, DCRGC [25] and RLBHE [26], as well as five deep learning-based approaches, IE-CGAN [13], BBCNN [27], IDTransformer [28], AverNet [18], and Shift-Net [17], the presented Mamba model consistently outperforms existing solutions. This is demonstrated through qualitative improvements and quantitative assessments using state-of-the-art thermal image quality measures, such as EME [29], BDIM [30], DMTE [31], MDIMTE [31], LGTA [32] and BIE [33].
The integrated design of the Mamba network combines the adaptability of deep learning with the stability of state space modeling, resulting in enhanced robustness, efficiency, and applicability. This makes it well suited for complex real-world tasks, such as reliable perception for navigation and safety, effective surveillance under challenging conditions, and improved imaging for security, military, and remote sensing applications.
The rest of the paper is organized as follows. Section 2 provides background information on thermal video enhancement and its challenges and reviews related work on existing thermal video enhancement methods, including traditional and deep learning approaches. Section 3 details the proposed TVEMamba framework, outlining its architecture and describing the data generation process. Section 4 presents experimental results, including qualitative and quantitative comparisons across datasets, with an ablation study and object detection performance. Finally, Section 5 summarizes the contributions of this work and highlights potential future research directions.

2. Related Works

2.1. Thermal Imaging Enhancement Models

Thermal image enhancement algorithms are generally divided into two main categories: traditional methods and learning-based methods. Traditional approaches [11,12,13,14,15,16,17,25,26] rely exclusively on patterns learned from unlabeled data. A fundamental technique in this category is Histogram Equalization (HE) [34], which approximately equalizes the cumulative distribution function of the histogram to map pixel intensities. However, HE often over enhances image contrast because it does not have a way to control the level of enhancement. To address this, Adaptive Histogram Equalization (AHE) [26,35] was developed to preserve more image details compared to standard HE techniques. Despite its advantages, AHE can still cause over enhancement in certain image regions due to homogeneous blocks. Contrast adjustment techniques [25] aim to enhance visibility by adjusting intensity values and improving overall contrast. These methods are effective but may perform poorly on images with uneven illumination. Wavelet-based methods [36] decompose an image into different frequency components, allowing separate processing to enhance details and reduce noise. Discrete stationary wavelets are often used in this process. However, these methods require careful parameter selection and can be computationally intensive. Improper management may introduce artifacts into the image. Gradient Field Equalization [37] focuses on improving contrast and reducing noise. While effective in some scenarios, these methods often struggle in complex situations and can sometimes over enhance images, leading to noise amplification and brightness distortion. Frequency-domain-based thermal infrared image enhancement algorithms [38] have also been widely employed. These techniques transform images into the frequency domain, utilize high-pass filters to extract high-frequency components, enhance them, and then convert the images back to the spatial domain.
In contrast, deep learning-based methods leverage neural networks to learn and apply enhancement processes. By utilizing large datasets, these methods effectively address issues like low contrast, noise, and blurred details, making thermal images more suitable for analysis. For instance, MdNet [11] filters thermal radiation effects and enhances features at different scales, addressing noise and low contrast in infrared images based on multi-scale feature fusion. However, it cannot achieve fine processing of features, which limits its ability to fully capture intricate details. Kuang et al. [13] developed a conditional generative adversarial network designed to minimize the amplification of background noise while improving contrast and detail clarity. Shen et al. [28] introduced the Infrared Image Denoising Transformer, a symmetric encoder–decoder framework aimed at mitigating noise, low contrast, and quality degradation in infrared images. The architecture integrates a Convolutional Transposed Self-Attention Block to capture both long-range dependencies and local context, along with Convolutional Gated Linear Units and a Channel Coordinate Attention Block to enhance feature representation. Marnissi et al. [15] proposed the Thermal Enhancement Vision Generative Adversarial Network, which combines a Generative Adversarial Network (GAN) with a Vision Transformer (ViT). This approach uses a U-Net model for image generation and incorporates two ViT modules as discriminators to effectively capture global and local features, improving detail and contextual understanding. The main properties of thermal image enhancement methods, including their strengths and limitations, are summarized in Table 1. This table provides a systematic comparison across key performance metrics.

2.2. Video Enhancement Models

Video enhancement techniques are crucial for improving the quality of low-quality videos, making them clearer and more suitable for applications, such as surveillance, identity verification, traffic monitoring, and object recognition [39,40]. The primary goal is to enhance the video’s visual appearance or provide better representations for automated processing tasks [41]. Like thermal image enhancement methods, video enhancement algorithms can be broadly classified into two main categories: traditional and context-based fusion methods. Table 2 highlights the benefits and limitations of thermal video technology.
Traditional methods include spatial-based domain and transform-based domain techniques [42]. Spatial-based domain methods operate directly on the pixels of video frames and encompass techniques like contrast enhancement, histogram equalization, and tone mapping. Tone mapping is another approach used primarily for high-dynamic-range (HDR) videos. It compresses the luminance levels to displayable ranges on standard devices, enhancing visibility in underexposed areas [43].
Transform-based domain methods modify the frequency components of video frames using techniques like the Discrete Cosine Transform (DCT) and wavelet transforms. Compressed-domain enhancement enhances videos directly in the compressed domain by manipulating transform coefficients, reducing computational complexity and storage requirements [44]. Adjusting DCT coefficients can improve contrast and reduce noise without fully decompressing the video. Wavelet-based methods decompose video frames into different frequency components, allowing separate processing to enhance details and reduce noise [45].
Another significant approach is context-based fusion enhancement, combining information from multiple frames or integrating high-quality background data into low-quality videos [46,47]. This method leverages additional contextual information to enhance video quality, especially under challenging conditions like low light or uneven illumination. Image fusion techniques use Retinex theory to separate illumination and reflectance components, enabling better contrast and detail preservation [48]. By fusing images captured under different lighting conditions, it is possible to enhance the visibility of important scene elements. Motion detection and enhancement utilize algorithms like Gaussian Mixture Models (GMM) to detect moving objects, allowing selective enhancement of these areas [49]. This improves overall visual quality and better detection of important scene elements. Shift-Net [17] introduced an efficient framework leveraging grouped spatial-temporal shift modules to implicitly combine inter-frame information. While the approach demonstrates strong performance, it faces limitations in processing long-sequence videos and exhibits high memory usage. EvLowLight [50] maintains temporal coherence by estimating motion from both events and frames while aligning spatial details across their different resolutions. However, it directly merges features from events and images without addressing noise variations in different local regions. ViWS-Net [16] utilizes degradation messenger tokens to capture specific degradation characteristics and guide the restoration process, effectively addressing unknown types of degradations. In comparison, AverNet [18] introduces a video restoration network specifically designed to handle time-varying unknown degradations. It employs a Prompt-Guided Alignment module for frame alignment and a Prompt-Conditioned Enhancement module for restoring diverse degradations, effectively addressing not only unknown degradation types but also their temporal variations.
Ensuring temporal coherence between frames is crucial to avoid flickering and maintaining visual consistency [51]. In summary, video enhancement techniques play a vital role in improving the clarity and usability of videos.

2.3. State Space Models

The Mamba [52] model is an advanced dynamic state space model (SSM) featuring efficient selection mechanisms, gaining popularity in computer vision. Its main advantage is handling long-range dependencies in data while maintaining linear computational complexity. This significantly improves over traditional transformers, which suffer from quadratic complexity as image sizes increase. Mamba has shown promising results in various visual tasks, such as image classification, feature enhancement, and multi-modal fusion [53]. Due to its efficiency, it is poised as a strong candidate to potentially replace CNNs and transformers as the foundational architecture in visual applications.
Recently, Mamba has been successfully applied to various applications, including image enhancement, video analysis, and object detection [54]. These applications highlight its versatility and significant potential to enhance the accuracy and efficiency of computer vision systems. By effectively managing long-range dependencies and keeping computational demands low, Mamba offers a robust solution for modern visual tasks, paving the way for advancements in the field.
Our work integrated the SSM into visual tasks by following the approach outlined in [55]. The SS2D module in our model consists of three primary operations: Scan Expanding, S6 blocks, and Scan Merging. Initially, the input images undergo the Scan Expanding operation, which systematically unfolds the image from its four corners toward the center. This rearrangement of the spatial structure allows the model to capture features from different spatial regions more effectively. Then, the image is flattened, and the sequence is fed into the S6 module responsible for feature extraction. The operations within the S6 module can be expressed as:
h t = A * h t 1 + B * x t
y t   = C * h t
where, h t represents the latent state at time t, x t represents the input variable, y t is the output, and A ,   B ,   and   C are learnable parameters. The features extracted from the four directions are then summed and merged, and the dimensions of the merged output are adjusted to match the original input size. After processing through the S6 blocks, the Scan Merging operation restores the spatial structure by reorganizing the flattened sequence back into its two-dimensional form. This combination of scan operations enables the SS2D module to effectively capture both local and global features in the image, enhancing feature extraction for our visual tasks.

3. Materials and Methods

3.1. Network Structure

Figure 2 illustrates the presented TVEMamba framework, which consists of three modules: a sharpening and denoising network (SD-Net), a blur-resistant motion estimation network (BRME-Net), and a motion deblurring network (MD-Net). This framework follows three steps to enhance clarity, contrast, and detail in thermal videos.
Step 1: SD-Net (see Figure 2a) improves the sharpness of thermal images and removes noise using a Mamba-based network with an encoder–decoder structure, capturing local and long-range contextual features. Both processes are symmetric and divided into two levels. Each downsampling level consists of a Basic Denoising State Space Model (BDSSM), a downsampling operation, and a convolutional layer with a kernel size of 3 × 3. Similarly, upsampling involves two levels: an upsampling operation, a 1 × 1 convolution applied to the merged features from the corresponding downsampling layer, and a BDSSM. Finally, a 3 × 3 convolution is applied to the image to reduce dimensionality and restore it to grayscale with a single channel. The BDSSM block includes a Basic Denoising (BD) module consisting of four consecutive convolutional layers followed by a residual connection (see Figure 2c), an SS2D module, a normalization layer, and a feed-forward network (FFN), as shown in Figure 2b.
Step 2: Then BRM-Net (see Figure 2a) takes three consecutive input frames from the previous step, T t 1 , T t , and T t + 1 , where T t is the t-th input frame, then estimates the optical flow W t 1 from T t - 1 to T t and W t from T t to T t + 1 . The BRM-Net architecture is based on NeuFlow [56], which computes optical flow between two images. We adopt NeuFlow for BRM-Net due to its comparable accuracy to NeuFlowV2 [57] but with significantly fewer parameters. The architecture follows a global-to-local scheme: global matching is performed on a 1/16 resolution to capture large displacements, followed by refinement at 1/8 resolution using lightweight CNN layers for improved accuracy.
NeuFlow was chosen over traditional non-learning-based methods, such as Lucas–Kanade [58], and TV-L1 [59], because these approaches rely heavily on gradient or intensity differences, which often struggle with challenging scenarios like significant motion blur. While methods like TV-L1 incorporate Total Variation regularization, making them more robust to noise and mild blur compared to simpler methods like Lucas–Kanade, they are still limited in handling significant motion blur and large displacements. Additionally, traditional methods are computationally intensive for high-resolution data due to their iterative optimization processes and rely on fixed mathematical models that require manual tuning to adapt to diverse data distributions. NeuFlow, in contrast, demonstrates robustness in handling large displacements and ambiguities while maintaining computational efficiency. Its ability to capture global motion through low-resolution matching and refine details using lightweight CNN layers makes it particularly suitable for challenging environmental conditions, where images often exhibit significant blur and complex motion.
Step3: The MD-Net (see Figure 2a) is applied in the final stage of the enhancement process. It takes as input the frames T t 1 ,   T t , T t + 1 , and the optical flows W t 1 and W t to generate a blur-free, corrected T t frame (see Figure 1b). Both stages of TVEMamba are built on a U-Net-based encoder–decoder architecture with the integration of vision Mamba. Similar to SD-Net, MD-Net employs a two-level encoder–decoder structure; however, instead of BDSSM, we apply an Attention-Based State Space Model (ABSSM), which includes an Optical Flow Attention (OFA) module, an SS2D module, a normalization layer, and a feed-forward network (FFN) (see Figure 2b). The OFA module uses two branches to generate attention weights based on both local and global features. The branches share a similar structure, utilizing convolutional layers, but the global branch uses dilated convolutions to capture more global features. The outputs from both branches are concatenated, and the final convolution layer generates a weight map for T t 1 , T t , T t + 1 frames, as shown in Figure 2c.

3.2. Training and Dataset

3.2.1. Dataset Generation

The experiments were conducted on the FLIR dataset [21]. Since no thermal video dataset with ground-truth labels is available, we generated synthetic labels using two methods, as shown in Figure 3. First, we applied a sharpening technique [60] with the following parameters: patch radius r = 11, epsilon ε = 0.01, scale s = 1, and kappa k where k was chosen from a range of 4 to 6, incremented in steps of 0.2, based on the entropy measure for thermal images [33]. However, thermal images are inherently noisy, and the sharpening process amplifies this noise. We applied a denoising method [61] to address this, utilizing a recurrent network with non-local operations for image restoration. Two denoising models were employed, using noise levels of 15 and 25. Images denoised at noise level 15 retained a small amount of noise, while those at noise level 25 were noise-free but lost small details. To balance noise reduction and detail preservation, we merged the two denoised images using the following formula:
I L a b e l   = c * I n o i s e = 25 + ( 1 c ) *   I n o i s e = 15
where c = 0.6. This approach allowed us to generate high-quality synthetic labels. The data generation process took approximately six days. The dataset contains three videos, totaling 4224 image frames with a resolution of 640 × 512.

3.2.2. Sharpening and Denoising Network

SD-Net was trained for 250 epochs using the Adam optimizer with an initial learning rate of 1 × 10−4. The experiments were conducted on the above-described dataset. Simple augmentation techniques were applied to increase generalizability, such as horizontal and vertical flips and random cropping to 256 × 256-pixel patches. All experiments were performed on an NVIDIA RTX 4090 GPU with 24 GB of memory. The network was trained using the following loss function:
L S D = M S E ( I I n p u t ,   I L a b e l )
where MSE [62] is the Mean Squared Error, which minimizes the difference between the predicted and labeled images, I I n p u t represents the input image and I L a b e l represents the sharp generated label image.

3.2.3. Blur-Resistant Motion Estimation Network

NeuFlow was initially trained using blur-free datasets that provide ground-truth optical flow maps, such as Sintel [63], KITTI [64], and HD1K [65]. However, thermal videos often contain blur, resulting in inaccurate optical flow estimations. Unfortunately, no available datasets offer ground-truth optical flow maps for blurry images. To overcome this limitation, we fine-tuned our BRM-Net using a blurred video dataset [66] that contains pairs of sharp and blurred videos. The blurred video dataset was created by capturing sharp videos at high frame rates and averaging adjacent frames to simulate blur. The dataset contains 71 pairs of blurry videos and corresponding sharp versions, providing 6708 pairs of 1280 × 720 blurred and sharp frames. We generated optical flow maps for training using only the sharp frames, employing the NeuFlow model. However, during optical flow generation, the model failed to generate accurate flow for certain videos. The accuracy of optical flow predictions can depend on the dataset on which the optical flow method was initially trained. The failed images may have a different distribution than the training dataset. After filtering out videos with inaccurate flow estimates, the final dataset for training consisted of 46 videos, totaling 4369 frames. Inspired by [67], we used the same estimated optical flow map for the following pairs:
( T t 1 b l u r ,   T t b l u r ) ,   ( T t 1 b l u r ,   T t s h a r p ) , ( T t 1 s h a r p ,   T t b l u r ) , ( T t 1 s h a r p ,   T t s h a r p ) .
The network was trained for 300 epochs using the Adam optimizer, with an initial learning rate of 1 × 10−4. The input images were cropped to 256 × 256 pixels. The network was trained using the following loss function:
L B R = M S E ( W t b l u r , b l u r ,   W t ) + M S E ( W t b l u r , s h a r p ,   W t ) + M S E ( W t s h a r p , b l u r ,   W t ) + M S E ( W t s h a r p , s h a r p , W t )  
where, W t * , * , is optical flow predicted by BRM-Net using ( T t 1 * ,   T t * ) frames and W t is the label generated by the same network using only sharp frames, where * { b l u r , s h a r p } .

3.2.4. Motion Deblurring Network

The MD-Net was trained over 300 epochs using the Adam optimizer, with an initial learning rate of 5 × 10 4 . For training, we utilized the same dataset as in BRM-Net, which consists of 46 pairs of blurry videos and their corresponding sharp videos. We used data augmentation techniques, including horizontal and vertical flips and random cropping to 256 × 256 pixel patches. The network was trained using the following loss function:
L M D = M S E ( T t ,   T t s h a r p )
where MSE is the Mean Squared Error, which minimizes the difference between the predicted and target images, T t represents the network’s prediction, and T t s h a r p represents the sharp ground-truth frame.

4. Results

This Section presents the experimental results of the proposed TVEMamba framework, positioning it alongside several established enhancement methods. For comparison, we selected two representative traditional approaches, DCRGC [25] and RLBHE [26], and five deep learning-based methods, IE-CGAN [13], BBCNN [27], IDTransformer [28], AverNet [18], and Shift-Net [17]. These methods were chosen due to their relevance in addressing the common challenges of thermal imaging, such as low contrast and noise, and their demonstrated effectiveness in various thermal image enhancement tasks.
By conducting a comparative analysis, we aim to identify each technique’s unique contributions and potential limitations, including their ability to maintain image integrity, enhance critical features, and avoid common artifacts. While some existing methods may offer marginal improvements in contrast, they often come at the cost of increased noise or halo effects that undermine the overall image quality. In contrast, the TVEMamba achieves a favorable balance between visual clarity and structural fidelity, providing a more stable and versatile foundation for tasks like object detection. This combination of enhancement quality and robustness under varying lighting conditions highlights the practical value of TVEMamba in real-world thermal imaging scenarios.

4.1. Qualitative Comparision

Figure 4 presents a qualitative comparison of the TVEMamba framework against several established thermal image enhancement methods applied to a variety of datasets, including BIRDSAI [20], FLIR [21], CAMEL [22], Autonomous Vehicles [23] and Solar Panels [24]. In some cases, DCRGC can achieve balanced contrast, but it frequently introduces halo artifacts and amplifies noise (Figure 4a,c). These issues become especially noticeable in images that are too dark or bright, ultimately reducing their overall visual quality and usefulness. RLBHE applies smaller contrast adjustments but is highly sensitive to bright areas, making already bright regions even more intense while failing to reduce existing noise. IE-CGAN, while aiming to reduce noise in underexposed scenes, often produces overly dark images that obscure subtle details critical for further analysis (Figure 4c,d). Conversely, when dealing with overexposed images, it tends to over brighten them, causing the loss of important features (Figure 4b). BBCNN, on the other hand, does not sufficiently highlight dark defects, resulting in the loss of essential information. It also struggles to maintain consistent brightness and contrast across different scenes, occasionally introducing artifacts. Furthermore, its tendency to add excessive sharpness can distort the natural appearance of the images, potentially leading to misinterpretation. AverNet can effectively remove small amounts of noise without significantly changing the contrast (Figure 4c,d). However, it struggles in more complex and challenging situations, where it fails to enhance subtle details. This leaves more complicated noisy or hidden features untouched, reducing its effectiveness under such conditions. Shift-Net and IDTransformer handle more complex noise than AverNet and are capable of addressing some challenging scenarios. However, both methods fail to improve contrast effectively and cannot manage motion blur adequately. As seen in Figure 4a,b, this limitation results in images becoming blurrier and losing critical details, particularly in scenes with dynamic elements or intricate textures.
In comparison, TVEMamba achieves a more balanced enhancement. It preserves structural details, maintains natural brightness and contrast levels, and reduces noise without introducing distracting artifacts. Even under difficult scenarios, such as low-light conditions, moving elements, or complex textures, TVEMamba consistently produces stable, clear, and visually coherent frames. Additionally, Figure 5 provides a detailed view of how our TVEMamba framework preserves and refines subtle features in the image. The improvement in edge sharpness, textural fidelity, and balance of contrast and brightness is visible, underscoring the method’s ability to recover important scene details. Moreover, a simple colorization technique enhances the image interpretability, allowing distinct elements to stand out more clearly.
Finally, Figure 6 illustrates the effectiveness of our TVEMamba framework on a low-quality solar panel video dataset. This figure displays five sequential frames from the original video (top row) and their enhanced counterparts generated by TVEMamba (bottom row). The input frames suffer from poor visibility and a lack of detail, making it difficult to identify subtle structural features. In contrast, the enhanced frames significantly improve clarity, contrast, and edge definition. TVEMamba highlights critical details and ensures smooth temporal consistency across frames, an essential factor for video analysis applications. This example further demonstrates the robustness and adaptability of our model in handling challenging real-world scenarios.

4.2. Quantitative Comparision

To assess the effectiveness of the TVEMamba framework, we used six non-reference image quality metrics as follows: (i) Measure of Enhancement (EME) [29], which evaluates image contrast entropy on a block basis rather than individual pixels. This metric is essential for assessing enhanced images and highlighting contrast variations within blocks. (ii) Block Distribution-Based Information Measure (BDIM) [30], which quantifies the information in image blocks by examining local contrast and structural details. It ensures that fine details are preserved and enhanced effectively. (iii) Density-based Measure of Thermal-image Enhancement (DMTE) [31], which incorporates elements of the human visual system with density-based analysis. (iv) Global Contrast Measure of Enhancement (MDIMTE) [31], which combines features related to human vision, information theory, and distribution-based metrics. This measure provides a holistic assessment of enhancement quality by focusing on overall contrast improvements that align with human perception and effective information distribution. (v) Local and Global Thermal Assessment (LGTA) [32], which integrates both local and global features to evaluate thermal image quality comprehensively. By combining block-level analysis with global intensity distribution, it closely aligns with human perception, offering nuanced insights into image clarity and enhancement. (vi) Block-Wise Image Entropy (BIE) [33], which captures the informational content of thermal images by combining local and global data, is specifically designed to assess thermal images. Unlike traditional entropy-based measures, BIE evaluates image quality and enhancement by preserving fine structural details and ensuring a balanced distribution of information across the image.
High scores across these metrics indicate superior enhancement and a more natural visual appearance. Table 3 presents the comparative analysis, showcasing TVEMamba’s performance against existing methods. Our approach outperforms both traditional and CNN-based methods, achieving the highest average scores across all metrics. These results demonstrate TVEMamba’s outstanding ability to enhance thermal images while preserving critical details and maintaining a realistic appearance.

4.3. Evaluation Metrics for Object Detection

To validate our approach for thermal video enhancement, we employed object detection methods and evaluated their performance using standard metrics: mAP0.5 and mAP0.5:0.95. These metrics are widely used to assess object detection models’ localization and classification accuracy. Precision and recall are fundamental components of these evaluations, where precision measures the proportion of true positive samples in all of the predicted positive samples, and recall is used to measure the proportion of true positive samples in all of the predicted positive samples. These can be mathematically expressed as:
  p r e c i s i o n = T P T P + F P
  r e c a l l = T P T P + F N
where, T P (True Positive) represents the number of objects that are correctly recognized as belonging to the target class, F P (False Positives) refers to the number of instances where non-target objects are incorrectly identified as belonging to the target class, and F N (False Negatives) indicates the number of instances where target objects are incorrectly classified as non-target objects.
We use Intersection over Union (IoU), which measures the overlap between the predicted bounding box and the ground-truth bounding box to assess the accuracy of the predicted bounding boxes. Mathematically, IoU is defined as the ratio of the intersection of the two bounding boxes to their union:
  IoU = | B g B p | | B g B p |
where, B g and B p represent the ground truth and predicted bounding boxes, respectively. A prediction is a true positive if the IoU exceeds a predefined threshold, such as 0.5.
Building on these concepts, the mean average precision (mAP) is the primary evaluation metric for object detection. The mAP calculates the average precision (AP) for each class and then averages the values across all classes. For mAP0.5, the IoU threshold is fixed at 0.5, meaning that predicted bounding boxes are required to have at least 50% overlap with the ground truth. For a more rigorous evaluation, mAP0.5:0.95 averages the precision over multiple IoU thresholds, ranging from 0.5 to 0.95 in steps of 0.05, providing a more comprehensive measure of detection accuracy. This approach ensures that both the localization quality of the bounding boxes and the ability to classify objects correctly are accounted for. A higher mAP value indicates better overall performance of the object detection model. Using these metrics, we can effectively measure the improvements in object detection accuracy achieved through our thermal video enhancement method.

4.4. Ablation Study

We conducted a series of ablation experiments to assess the contribution of each module in the proposed TVEMamba. Specifically, we trained the network with and without the SD-Net, BRME-Net, and MD-Net to evaluate their individual impacts. Additionally, to analyze the effects of the BD and OFA blocks, we trained the network with and without incorporating these blocks to evaluate their impact on performance. Only the BD block was retained in one variant, while the OFA block was removed with the optical flow estimation module. In another variant, only the OFA block was retained while the BD block was removed. As shown in Table 4, the highest values were achieved when all modules were integrated, indicating that their combination significantly improves the model’s ability to enhance overall image quality.
Also, Figure 7 shows that the highest values for each measure were achieved when both blocks were integrated, indicating that their combination significantly improves the model’s ability to enhance overall image quality. Additionally, Figure 8 provides qualitative comparisons that illustrate the network’s performance for each module. Furthermore, to evaluate the effectiveness of our TVEMamba framework on downstream computer vision tasks, we utilized object detection experiments using two datasets: BIRDSAI and FLIR. The BIRDSAI dataset contains thermal videos of elephants captured for wildlife monitoring. To assess our enhancement method under different scenarios, we designed two labeling schemes: first, with two classes, “Elephant” and “Unknown” (including all non-elephant objects), and second, with three classes, “Elephant”, “Human”, and “Unknown”. This allowed us to evaluate the model’s ability to distinguish between multiple object categories in enhanced thermal videos. The FLIR dataset, commonly used for autonomous driving and surveillance applications, focuses on two classes: “Pedestrian” and “Car”. We employed two object detection architectures: YOLOR [68], which combines implicit and explicit knowledge within a single model for efficient detection, and Hyper-YOLO [69], incorporating hyperparameter optimization and architectural improvements. For each dataset and model, we trained on both the original thermal images and the enhanced images produced by the TVEMamba framework. As shown in Table 5, the notable improvements observed on the BIRDSAI dataset suggest that our enhancement method particularly benefits datasets with lower initial image quality due to low contrast and noise. In addition, Figure 9 illustrates YOLOR predictions on both original and enhanced BIRDSAI frames, clearly demonstrating improved detection accuracy after enhancement. There was no significant improvement in object detection performance for the FLIR dataset. However, the lack of deterioration confirms that our method does not introduce artifacts or distortions that could negatively impact detection. These findings indicate that the proposed TVEMamba framework benefits applications requiring reliable object detection in challenging thermal environments, such as wildlife monitoring and surveillance under adverse conditions.

5. Discussion

This paper introduces the Mamba model, a novel thermal video enhancement method that leverages a State Space 2D module integrated with a Convolutional Neural Network. The Mamba model addresses major challenges, such as low contrast, motion blur, and noise by incorporating the Basic Denoising and Optical Flow Attention modules. Simulation results demonstrated across multiple datasets, including BIRDSAI, FLIR, CAMEL, Autonomous Vehicles, and Solar Panel, highlight the Mamba model’s ability to outperform both traditional and deep learning-based methods, resulting in higher-quality thermal videos suitable for a wide range of applications. Through this integration of state space modeling and deep learning, the Mamba network adapts to diverse lighting conditions and varying motion patterns, making it well suited for practical use cases. Applications include surveillance, where robust detection and tracking under challenging conditions are essential, autonomous systems, where reliable perception ensures safety and navigation, and remote sensing, where improved thermal imaging can aid critical monitoring tasks.
Future research directions involve extending the Mamba model’s capabilities. First, we aim to enhance data association in trackers by leveraging richer thermal information. Second, we will explore the fusion of visible and non-visible spectral data to improve tracking accuracy under different lighting and environmental conditions. Finally, optimizing the model for real-time performance will facilitate its deployment in time-sensitive applications. Advancing state-of-the-art thermal video enhancement will unlock the full potential of thermal imaging technologies, ultimately improving the performance and reliability of a broad spectrum of computer vision tasks. Finally, we will develop a GUI-based image enhancement, object detection, and tracking framework that can run through the cloud environment.

Author Contributions

Conceptualization, S.A.; methodology, S.H.; software, S.H.; validation, S.H., A.G. and S.A.; formal analysis, S.H., A.G. and K.P.; investigation, S.H., S.A., A.G. and K.P.; resources, S.H.; data curation, S.H. and K.P.; writing—original draft preparation, S.H.; writing—review and editing, S.H., S.A., A.G. and K.P.; visualization, S.H.; supervision, S.A., A.G. and K.P.; project administration, S.A.; funding acquisition, S.A. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the Higher Education and Science Committee, in the frames of the project № 25FAST-1B001.

Data Availability Statement

Due to the nature of this research, the participants did not agree to share their data publicly; therefore, the supporting data are not available.

Acknowledgments

The first author thanks Sarkis and Nune Sepetjians for their support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shidik, G.; Noersasongko, E.; Nugraha, A.; Andono, P.N.; Jumanto, J.; Kusuma, E.J. A systematic review of intelligence video surveillance: Trends, techniques, frameworks, and datasets. IEEE Access 2019, 7, 170457–170473. [Google Scholar] [CrossRef]
  2. Karpuzov, S.; Petkov, G.; Ilieva, S.; Petkov, A.; Kalitzin, S. Object tracking based on optical flow reconstruction of motion-group parameters. Information 2024, 15, 296. [Google Scholar] [CrossRef]
  3. Alsrehin, N.; Klaib, A.; Magableh, A. Intelligent transportation and control systems using data mining and machine learning techniques: A comprehensive study. IEEE Access 2019, 7, 49830–49857. [Google Scholar] [CrossRef]
  4. Zhang, L.; Xiong, N.; Gao, W.; Wu, P. Improved detection method for micro-targets in remote sensing images. Information 2024, 15, 108. [Google Scholar] [CrossRef]
  5. Dong, Y.; Pan, W.D. A survey on compression domain image and video data processing and analysis techniques. Information 2023, 14, 184. [Google Scholar] [CrossRef]
  6. Yoshida, E.; Kato, S.; Sato, T.; Suzuki, T.; Koyama, H.; Kato, S. Proposal and prototyping on wildlife tracking system using infrared sensors. In Proceedings of the International Conference on Information Networking (ICOIN), Jeju-si, Republic of Korea, 7–10 January 2022; pp. 292–297. [Google Scholar]
  7. Kim, E.; Kim, W.; Park, J.; Yeo, K. Human detection in infrared image using daytime model-based transfer learning for military surveillance system. In Proceedings of the 14th International Conference on Information and Communication Technology Convergence (ICTC), Jeju Island, Republic of Korea, 11–13 October 2023; pp. 1306–1308. [Google Scholar]
  8. Yuan, C.-J.; Lan, S.-J.; Yan, G.-D.; Wang, D.; Lu, J.-H.; Meng, Q.-F. Application of near-infrared spectroscopy in rapid determination of adenosine and polysaccharide in Cordyceps militaris. In Proceedings of the Fifth International Conference on Natural Computation, Tianjin, China, 14–16 August 2009; pp. 578–582. [Google Scholar]
  9. Alheeti, K.; McDonald-Maier, K. An intelligent security system for autonomous cars based on infrared sensors. In Proceedings of the 23rd Internat. Conference on Automation and Computing (ICAC), Huddersfield, UK, 7–8 September 2017; pp. 1–5. [Google Scholar]
  10. Mo, F.; Li, H.; Yao, X.; Wang, Q.; Jing, Q.; Zhang, L. Intelligent onboard processing and multichannel transmission technology for infrared remote sensing data. In Proceedings of the IEEE International Geoscience and Remote Sensing Symposium, Yokohama, Japan, 28 July–2 August 2019; pp. 9063–9066. [Google Scholar]
  11. Wu, H.; Chen, B.; Guo, Z.; He, C.; Luo, S. Mini-infrared thermal imaging system image denoising with multi-head feature fusion and detail enhancement network. Opt. Laser Technol. 2024, 179, 111311. [Google Scholar] [CrossRef]
  12. Shi, Y.; Zhang, H.; Li, J.; Wang, X.; Zhao, Q.; Liu, T. GAPANet: Group alternate perceived attention network for optical imaging infrared thermal radiation effect correction. Opt. Express 2024, 32, 35888–35902. [Google Scholar] [CrossRef]
  13. Kuang, X.; Sui, X.; Liu, Y.; Chen, Q.; Gu, G. Single infrared image enhancement using a deep convolutional neural network. Neurocomputing 2019, 332, 119–128. [Google Scholar] [CrossRef]
  14. Xu, Z.; Zhao, H.; Zheng, Y.; Guo, H.; Li, S.; Lyu, Z. A dual nonsubsampled contourlet network for synthesis images and infrared thermal images denoising. PeerJ Comput. Sci. 2024, 10, e1817. [Google Scholar] [CrossRef] [PubMed]
  15. Marnissi, M.; Fathallah, A. Revolutionizing thermal imaging: GAN-based vision transformers for image enhancement. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Kuala Lumpur, Malaysia, 8–12 October 2023; pp. 2735–2739. [Google Scholar]
  16. Yang, Y.; Aviles-Rivero, A.I.; Fu, H.; Liu, Y.; Wang, W.; Zhu, L. Video adverse-weather-component suppression network via weather messenger and adversarial backpropagation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–5 October 2023; pp. 13200–13210. [Google Scholar]
  17. Li, D.; Shi, X.; Zhang, Y.; Cheung, K.C.; See, S.; Wang, X.; Qin, H.; Li, H. A simple baseline for video restoration with grouped spatial-temporal shift. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 9822–9832. [Google Scholar]
  18. Zhao, H.; Chen, L.; Wang, M.; Li, J.; Xu, T.; Zhou, Q. AverNet: All-in-one video restoration for time-varying unknown degradations. In Proceedings of the Thirty-Eighth Annual Conference on Neural Information Processing Systems (NeurIPS), Montréal, Canada, 11–16 December 2024. [Google Scholar]
  19. Park, J.; Kim, H.; Lee, J.; Choi, Y.; Song, M. Videomamba: Spatio-temporal selective state space model. In Proceedings of the European Conference on Computer Vision (ECCV), Cham, Switzerland, 4–8 October 2024; Springer: Cham, Switzerland, 2024. [Google Scholar]
  20. Bondi, E.; Jain, R.; Aggrawal, P.; Anand, S.; Hannaford, R.; Kapoor, A.; Piavis, J.; Shah, S.; Joppa, L.; Dilkina, B.; et al. BIRDSAI: A dataset for detection and tracking in aerial thermal infrared videos. In Proceedings of the IEEE Winter Conference on Applications of Computer Vision (WACV), Snowmass Village, CO, USA, 1–5 March 2020; pp. 1736–1745. [Google Scholar]
  21. FLIR Thermal Dataset. Available online: https://www.kaggle.com/datasets/deepnewbie/flir-thermal-images-dataset (accessed on 21 September 2024).
  22. Gebhardt, E.; Wolf, M. CAMEL dataset for visual and thermal infrared multiple object detection and tracking. In Proceedings of the 15th IEEE International Conference on Advanced Video and Signal Based Surveillance (AVSS), Auckland, New Zealand, 27–30 November 2018; pp. 1–6. [Google Scholar]
  23. Takumi, K.; Watanabe, K.; Ha, Q.; Tejero-De-Pablos, A.; Ushiku, Y.; Harada, T. Multispectral object detection for autonomous vehicles. In Proceedings of the Thematic Workshops of ACM Multimedia, Mountain View, CA, USA, 23–27 October 2017; pp. 35–43. [Google Scholar]
  24. Bommes, L.; Buerhop-Lutz, C.; Pickel, T.; Hauch, J.; Brabec, C.; Peters, I.M. Georeferencing of photovoltaic modules from aerial infrared videos using structure-from-motion. Prog. Photovolt. 2022, 30, 1122–1135. [Google Scholar] [CrossRef]
  25. Wang, Z.; Liang, Z.; Liu, C. A real-time image processor with combining dynamic contrast ratio enhancement and inverse gamma correction for PDP. Displays 2009, 30, 133–139. [Google Scholar] [CrossRef]
  26. Zuo, C.; Chen, Q.; Sui, X. Range limited bi-histogram equalization for image contrast enhancement. Optik 2013, 124, 425–431. [Google Scholar] [CrossRef]
  27. Lee, K.; Lee, J.; Lee, J.; Hwang, S.; Lee, S. Brightness-based convolutional neural network for thermal image enhancement. IEEE Access 2017, 5, 26867–26879. [Google Scholar] [CrossRef]
  28. Shen, Z.; Wang, H.; Zhang, X.; Li, J.; Liu, T.; Chen, M. IDTransformer: Infrared image denoising method based on convolutional transposed self-attention. Alex. Eng. J. 2024, 110, 310–321. [Google Scholar] [CrossRef]
  29. Agaian, S.; Panetta, K.; Grigoryan, A. A new measure of image enhancement. In Proceedings of the IASTED International Conference on Signal Processing and Communications (SPC), Marbella, Spain, 19–21 September 2000. [Google Scholar]
  30. Trongtirakul, T.; Agaian, S. Unsupervised and optimized thermal image quality enhancement and visual surveillance application. Signal Process. Image Commun. 2022, 105, 116714. [Google Scholar] [CrossRef]
  31. Agaian, S.; Roopaei, M.; Akopian, D. Thermal-image quality measurements. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 1–5. [Google Scholar]
  32. Agaian, S.; Ayunts, H.; Trongtirakul, T.; Hovhannisyan, S. A new method for judging thermal image quality with applications. Signal Process. 2024, 229, 109769. [Google Scholar] [CrossRef]
  33. Ayunts, H.; Grigoryan, A.; Agaian, S. Novel entropy for enhanced thermal imaging and uncertainty quantification. Entropy 2024, 26, 374. [Google Scholar] [CrossRef] [PubMed]
  34. Gonzalez, R.; Woods, R. Digital Image Processing, 2nd ed.; Prentice-Hall: Upper Saddle River, NJ, USA, 2002; Volume 793. [Google Scholar]
  35. Dhariwal, S. Comparative analysis of various image enhancement techniques. Int. J. Electron. Commun. Technol. 2011, 2, 91–95. [Google Scholar]
  36. Mudavath, T.; Niranjan, V. Thermal image enhancement for adverse weather scenarios: A wavelet transform and histogram clipping approach. Signal Image Video Process 2024, in press. [CrossRef]
  37. Grigoryan, A.; Agaian, S. Asymmetric and symmetric gradient operators with application in face recognition in Renaissance portrait art. In Proceedings of the SPIE, Defense + Commercial Sensing, Mobile Multimedia/Image Processing, Security, and Applications, Baltimore, MD, USA, 4–18 May 2019; Volume 10993, p. 12. [Google Scholar]
  38. Agaian, S.; Panetta, K.; Grigoryan, A. Transform-based image enhancement algorithms with performance measure. IEEE Trans. Image Process. 2001, 10, 367–382. [Google Scholar] [CrossRef]
  39. Kastrinaki, V.; Zervakis, M.; Kalaitzakis, K. A survey of video processing techniques for traffic applications. Image Vis. Comput. 2003, 21, 359–381. [Google Scholar] [CrossRef]
  40. González-Cepeda, J.; Ramajo, Á.; Armingol, J. Intelligent video surveillance systems for vehicle identification based on multinet architecture. Information 2022, 13, 325. [Google Scholar] [CrossRef]
  41. Rao, Y.; Lin, W.; Chen, L. Image-based fusion for video enhancement of nighttime surveillance. Opt. Eng. 2010, 49, 120501. [Google Scholar] [CrossRef]
  42. Agaian, S.; Blair, S.; Panetta, K. Transform coefficient histogram-based image enhancement algorithms using contrast entropy. IEEE Trans. Image Process. 2007, 16, 741–758. [Google Scholar] [CrossRef] [PubMed]
  43. Reinhard, E.; Ward, G.; Pattanaik, S.; Debevec, P. High Dynamic Range Imaging: Acquisition, Display, and Image-Based Lighting; Morgan: San Francisco, CA, USA, 2005. [Google Scholar]
  44. Lee, S. An efficient content-based image enhancement in the compressed domain using Retinex theory. IEEE Trans. Circuits Syst. Video Tech. 2007, 17, 199–213. [Google Scholar] [CrossRef]
  45. Balster, E.; Zheng, Y.F.; Ewing, R.L. Combined spatial and temporal domain wavelet shrinkage algorithm for video denoising. IEEE Trans. Circuits Syst. Video Technol. 2006, 16, 220–230. [Google Scholar]
  46. Wan, T.; Tzagkarakis, G.; Tsakalides, P.; Canagarajah, C.N.; Achim, A. Context enhancement through image fusion: A multi-resolution approach based on convolution of Cauchy distributions. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Las Vegas, NV, USA, 31 March–4 April 2008; pp. 1309–1312. [Google Scholar]
  47. Li, J.; Li, S.Z.; Pan, Q.; Yang, T. Illumination and motion-based video enhancement for night surveillance. In Proceedings of the IEEE workshop on Visual Surveillance and Performance Evaluation of Tracking and Surveillance, Beijing, China, 15–16 October 2005; pp. 169–175. [Google Scholar]
  48. Land, E.H.; McCann, J.J. Lightness and Retinex theory. J. Opt. Soc. Am. 1971, 61, 1–11. [Google Scholar] [CrossRef] [PubMed]
  49. Stauffer, C.; Grimson, W.E.L. Learning patterns of activity using real-time tracking. IEEE Trans. Pattern Anal. Mach. Intell. 2000, 22, 747–757. [Google Scholar] [CrossRef]
  50. Liang, J.; Yang, Y.; Li, B.; Duan, P.; Xu, Y.; Shi, B. Coherent event guided low-light video enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–5 October 2023; pp. 10615–10625. [Google Scholar]
  51. Cho, Y.; Lee, H.; Park, D.; Kim, C.Y. Enhancement for temporal resolution of video based on multi-frame feature trajectory and occlusion compensation. In Proceedings of the IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 389–392. [Google Scholar]
  52. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  53. Yuan, C.; Zhao, D.; Agaian, S.S. MUCM-Net: A Mamba powered UCM-Net for skin lesion segmentation. Explor. Med. 2024, 5, 694–708. [Google Scholar] [CrossRef]
  54. Zhang, H.; Zhu, Y.; Wang, D.; Zhang, L.; Chen, T.; Ye, Z. A survey on visual Mamba. arXiv 2024, arXiv:2404.15956v2. [Google Scholar]
  55. Bai, J.; Yin, Y.; He, Q.; Li, Y.; Zhang, X. Retinexmamba: Retinex-based mamba for low-light image enhancement. arXiv 2024, arXiv:2405.03349. [Google Scholar]
  56. Zhang, Z.; Jiang, H.; Singh, H. NeuFlow: Real-time, high-accuracy optical flow estimation on robots using edge devices. arXiv 2024, arXiv:2403.10425. [Google Scholar]
  57. Zhang, Z.; Gupta, A.; Jiang, H.; Singh, H. NeuFlow v2: High-efficiency optical flow estimation on edge devices. arXiv 2024, arXiv:2408.10161. [Google Scholar]
  58. Lucas, B.D.; Kanade, T. An iterative image registration technique with an application to stereo vision. In Proceedings of the 7th International Joint Conference on Artificial Intelligence (IJCAI ‘81), Vancouver, BC, Canada, 24–28 August 1981; Volume 2, pp. 674–679. [Google Scholar]
  59. Wedel, A.; Pock, T.; Zach, C.; Bischof, H.; Cremers, D. An improved algorithm for TV-L1 optical flow. In Proceedings of the Statistical and Geometrical Approaches to Visual Motion Analysis: International Dagstuhl Seminar, Dagstuhl Castle, Germany, 13–18 July 2008; Springer: Berlin/Heidelberg, Germany, 2009; pp. 23–45. [Google Scholar]
  60. Deng, G.; Galetto, F.; Al–nasrawi, M.; Waheed, W. A guided edge-aware smoothing-sharpening filter based on patch interpolation model and generalized gamma distribution. IEEE Open J. Signal Process. 2021, 2, 119–135. [Google Scholar] [CrossRef]
  61. Liu, D.; Wen, B.; Fan, Y.; Loy, C.C.; Huang, T.S. Non-local recurrent network for image restoration. In Proceedings of the Advances in Neural Information Processing Systems (NeurIPS), Montréal, QC, Canada, 3–8 December 2018; Volume 31. [Google Scholar]
  62. Murphy, A.H. Skill scores based on the mean square error and their relationships to the correlation coefficient. Mon. Weather Rev. 1988, 116, 2417–2424. [Google Scholar] [CrossRef]
  63. Butler, D.J.; Wulff, J.; Stanley, G.B.; Black, M.J. A naturalistic open source movie for optical flow evaluation. In Proceedings of the European Conference on Computer Vision (ECCV), Florence, Italy, 7–13 October 2012. [Google Scholar]
  64. Geiger, A.; Lenz, P.; Urtasun, R. Are we ready for autonomous driving? The KITTI vision benchmark suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Providence, RI, USA, 16–21 June 2012; pp. 3354–3361. [Google Scholar]
  65. Kondermann, D.; Nair, R.; Honauer, K.; Krispin, K.; Andrulis, J.; Brock, A.; Gussefeld, B.; Rahimimoghaddam, M.; Hofmann, S.; Brenner, C.; et al. The HCI benchmark suite: Stereo and flow ground truth with uncertainties for urban autonomous driving. In Proceedings of the CVPR Workshops, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 19–28. [Google Scholar]
  66. Su, S.; Delbracio, M.; Wang, J.; Sapiro, G.; Heidrich, W.; Wang, O. Deep video deblurring for hand-held cameras. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 237–246. [Google Scholar]
  67. Son, H.; Lee, J.; Lee, J.; Cho, S.; Lee, S. Recurrent video deblurring with blur-invariant motion estimation and pixel volumes. Assoc. Comput. Mach. 2021, 40, 5. [Google Scholar] [CrossRef]
  68. Wang, C.; Yeh, I.; Liao, H. You only learn one representation: Unified network for multiple tasks. J. Inf. Sci. Eng. 2023, 39, 691–709. [Google Scholar]
  69. Feng, Y.; Huang, J.; Du, S.; Ying, S.; Yong, J.-H.; Li, Y.; Ding, G.; Ji, R.; Gao, Y. Hyper-YOLO: When visual object detection meets hypergraph computation. arXiv 2024, arXiv:2408.04804. [Google Scholar] [CrossRef]
Figure 1. (a,c) Challenging thermal video frames. (b,d) Successful recovery and enhancement by TVEMamba.
Figure 1. (a,c) Challenging thermal video frames. (b,d) Successful recovery and enhancement by TVEMamba.
Information 16 00125 g001
Figure 2. (a) Overall architecture of TVEMamba, (b) Basic denoising state space model and attention-based state space model, and (c) Basic denoising module and optical flow attention module.
Figure 2. (a) Overall architecture of TVEMamba, (b) Basic denoising state space model and attention-based state space model, and (c) Basic denoising module and optical flow attention module.
Information 16 00125 g002
Figure 3. This figure shows the original image from FLIR [40], the corresponding sharpened image, denoised images with different noise levels, and the final merged result.
Figure 3. This figure shows the original image from FLIR [40], the corresponding sharpened image, denoised images with different noise levels, and the final merged result.
Information 16 00125 g003
Figure 4. Visual comparison on (a) BIRDSAI [50], (b) FLIR [40], (c) CAMEL [51], and (d) Autonomous Vehicles [52] datasets.
Figure 4. Visual comparison on (a) BIRDSAI [50], (b) FLIR [40], (c) CAMEL [51], and (d) Autonomous Vehicles [52] datasets.
Information 16 00125 g004
Figure 5. Performance of TVEMamba on BIRDSAI dataset.
Figure 5. Performance of TVEMamba on BIRDSAI dataset.
Information 16 00125 g005
Figure 6. Qualitative results of TVEMamba framework on Solar Panel video frames.
Figure 6. Qualitative results of TVEMamba framework on Solar Panel video frames.
Information 16 00125 g006
Figure 7. Evaluation of the contribution of OFA and BD Blocks in TVEMamba on the FLIR dataset.
Figure 7. Evaluation of the contribution of OFA and BD Blocks in TVEMamba on the FLIR dataset.
Information 16 00125 g007
Figure 8. Details of intermediate results from the proposed TVEMamba, showing the effects of each module.
Figure 8. Details of intermediate results from the proposed TVEMamba, showing the effects of each module.
Information 16 00125 g008
Figure 9. Object detection results on original and enhanced Images using the YOLOR method.
Figure 9. Object detection results on original and enhanced Images using the YOLOR method.
Information 16 00125 g009
Table 1. Comparative analysis of thermal image enhancement methods across key performance metrics (✓: Fully Performs, ±: Partially Performs).
Table 1. Comparative analysis of thermal image enhancement methods across key performance metrics (✓: Fully Performs, ±: Partially Performs).
DCRGCRLBHEIE-CGANBBCNNAverNetShift-NetIDTransformerTVEMamba
Noise reduction ±
Balanced contrast±± ±±
Handles underexposed areas ±±
Handles overexposed areas± ±±±±
Edge preservation ±±±±
Maintains natural brightness ±
Handles complex textures ±±
Artifact-free output
Table 2. Benefits and Limitations of Thermal Video Technology.
Table 2. Benefits and Limitations of Thermal Video Technology.
BenefitsLimitations
Objects can be observed in no light conditions (dark environments).Difficulty distinguishing between objects in proximity or of similar temperatures.
High performance in all weather conditions (rain, fog, snow, smoke).Generally, lower resolution compared to visible light images.
Opportunities for surveillance over large distances and areas, detecting motion over a wide range.Cannot see through glass or water, as these materials reflect infrared radiation, limiting use cases like capturing images of individuals in cars.
Detection of objects even when partially hidden by vegetation.More expensive than visible light cameras.
Promotes early detection of thermal anomalies (e.g., equipment overheating, fire hazards), contributing to preventive safety measures.Cannot identify detected individuals, as infrared radiation does not create detailed enough images.
Table 3. Quantitative comparison of TVEMamba with the state-of-the-art methods.
Table 3. Quantitative comparison of TVEMamba with the state-of-the-art methods.
BBCNNDCRGCIE-CGANRLBHEAverNetShift-NetIDTransformerTVEMamba
BIRDSAI
EME10.06020.26417.74818.3779.7218.7729.92922.942
DMTE0.2970.2970.2960.2970.2990.2600.2970.299
MDIMTE45.06042.62031.62046.00144.99443.78446.82447.132
BDIM0.9740.9860.9880.9860.9700.8670.9670.991
LGTA1.1581.1671.4231.1541.1191.1221.1511.172
BIE0.0850.0980.0760.0880.0860.0840.0990.109
CAMEL
EME14.63324.21424.01023.79617.14017.52014.72825.371
DMTE0.2930.2920.2900.2940.2960.2950.2930.296
MDIMTE39.83341.30932.00440.74740.68041.11741.73142.786
BDIM0.9900.9880.9920.9900.9840.9830.9810.994
LGTA1.2391.2351.3811.0891.1661.1351.1411.548
BIE0.0700.0910.0740.0870.0780.0690.0700.098
FLIR
EME10.74313.42410.56011.1859.8429.6798.34514.152
DMTE0.2950.2960.2950.2940.2980.2970.2970.298
MDIMTE43.80140.62741.02442.48649.00248.06050.01250.146
BDIM0.9720.9770.9650.9710.9630.9610.9580.982
LGTA1.1371.1461.1051.0801.1161.1111.1201.167
BIE0.1830.1800.1920.1960.1990.2020.2070.227
Autonomous Vehicles
EME2.9293.0887.2608.1305.4923.1157.14112.513
DMTE0.2990.2980.2970.2970.2990.2980.3000.310
MDIMTE51.51748.32653.65947.92555.38041.52956.01357.369
BDIM0.9370.9590.9430.9570.9370.9220.9230.963
LGTA1.1801.3931.1891.4111.2531.2051.2231.499
BIE0.0930.1820.0880.1500.0950.0920.0960.097
Table 4. Ablation study on BIRDSAI dataset.
Table 4. Ablation study on BIRDSAI dataset.
EMEDMTEMDIMTEBDIMLGTABIE
w/o SD-Net16.2450.29142.1210.9641.1480.081
w/o MD-Net20.1870.29546.3450.9811.1590.093
w/o BRME-Net21.1450.29846.8990.9891.1680.101
TVEMamba22.9420.29947.1320.9911.1720.109
Table 5. Object detection performance on the BIRDSAI and FLIR datasets. YOLOR1 and Hyper-YOLO1 models are trained on original datasets, and YOLOR2 and Hyper-YOLO2 models are trained on enhanced datasets produced by the TVEMamba framework.
Table 5. Object detection performance on the BIRDSAI and FLIR datasets. YOLOR1 and Hyper-YOLO1 models are trained on original datasets, and YOLOR2 and Hyper-YOLO2 models are trained on enhanced datasets produced by the TVEMamba framework.
DatasetBIRDSAIBIRDSAIBIRDSAIFLIR
Classes2322
ArchitectureYOLOR1YOLOR2YOLOR1YOLOR2Hyper-YOLO1Hyper-YOLO2Hyper-YOLO1Hyper-YOLO2
mAP0.538.144.225.029.738.043.989.889.9
mAP0.5:0.913.216.89.310.912.916.456.656.7
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hovhannisyan, S.; Agaian, S.; Panetta, K.; Grigoryan, A. Thermal Video Enhancement Mamba: A Novel Approach to Thermal Video Enhancement for Real-World Applications. Information 2025, 16, 125. https://doi.org/10.3390/info16020125

AMA Style

Hovhannisyan S, Agaian S, Panetta K, Grigoryan A. Thermal Video Enhancement Mamba: A Novel Approach to Thermal Video Enhancement for Real-World Applications. Information. 2025; 16(2):125. https://doi.org/10.3390/info16020125

Chicago/Turabian Style

Hovhannisyan, Sargis, Sos Agaian, Karen Panetta, and Artyom Grigoryan. 2025. "Thermal Video Enhancement Mamba: A Novel Approach to Thermal Video Enhancement for Real-World Applications" Information 16, no. 2: 125. https://doi.org/10.3390/info16020125

APA Style

Hovhannisyan, S., Agaian, S., Panetta, K., & Grigoryan, A. (2025). Thermal Video Enhancement Mamba: A Novel Approach to Thermal Video Enhancement for Real-World Applications. Information, 16(2), 125. https://doi.org/10.3390/info16020125

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop