Next Article in Journal
Path Planning Algorithm for Manipulators in Complex Scenes Based on Improved RRT*
Previous Article in Journal
Performance Evaluation of Deep Learning Image Classification Modules in the MUN-ABSAI Ice Risk Management Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention †

1
Department of Computer Science, University of Machester, Manchester M13 9PL, UK
2
Faculty of Electronics, Telecommunications and Information Technologies, Polytechnic University Timisoara, 300223 Timisoara, Romania
*
Authors to whom correspondence should be addressed.
This paper is an extended version of our paper published in IEEE International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 7–8 November 2024.
Sensors 2025, 25(2), 327; https://doi.org/10.3390/s25020327
Submission received: 2 December 2024 / Revised: 3 January 2025 / Accepted: 4 January 2025 / Published: 8 January 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Low-light image enhancement (LLIE) techniques improve the performance of image sensors by enhancing visibility and details in poorly lit environments and have significantly benefited from recent research into Transformer models. This work presents a novel Transformer attention mechanism inspired by the Kolmogorov–Arnold representation theorem, incorporating learnable non-linearity and multivariate function decomposition. This innovative mechanism is the foundation of KAN- T , our proposed Transformer network. By enhancing feature flexibility and enabling the model to capture broader contextual information, KAN- T achieves superior performance. Our comprehensive experiments, both quantitative and qualitative, demonstrate that the proposed method achieves state-of-the-art performance in low-light image enhancement, highlighting its effectiveness and wide-ranging applicability. The code will be released upon publication.

1. Introduction

Image sensors often face challenges in low-light conditions, such as noise and reduced contrast, which degrade image quality. Low-light image enhancement (LLIE) is a specialized area of computer vision that focuses on improving image clarity, contrast, and overall quality under such conditions, addressing issues caused by poor lighting and distortions. Such degradation not only results in subjectively unpleasant visual experiences but also impairs the performance of many CV systems. The goal of LLIE is to enhance visibility and contrast while simultaneously addressing the distortions inherent in dark environments.
Low-light conditions describe environments where illumination levels are insufficient for optimal visibility. However, defining precise theoretical thresholds to characterize low-light environments has proven challenging in practical applications. Consequently, there is no universally accepted standard for identifying or quantifying what constitutes low-light conditions [1].
Classical LLIE methods, such as gamma correction and histogram equalization, have limitations in effectively addressing low-light distortions. Traditional CV approaches attempt to improve upon these methods by considering illumination factors; however, they often introduce additional artifacts and imbalances in the restored images [1].
LLIE techniques have significantly progressed with Convolutional Neural Network (CNN) approaches, with solutions generally falling into two main categories. The first approach uses CNNs to directly map low-light images to their normal-light counterparts. While effective, this method often disregards human color perception and lacks theoretical interpretability. The second approach, inspired by Retinex theory [2], employs a more intricate multi-stage training pipeline. This method utilizes different CNNs for specific tasks, such as decomposing the color image, denoising the reflectance, and adjusting illumination. Although more consistent with theoretical models, this approach introduces considerable complexity and requires multiple training stages, posing significant challenges [3].
Although CNNs are highly effective at capturing local features, they often struggle with modeling long-range dependencies and adapting to broader contextual information. The self-attention mechanism introduced in the Transformer architecture [4] has partially addressed these challenges and has been adapted for image processing through Vision Transformers (ViTs) [4]. Recently, ViTs have been utilized in various LLIEs [5,6,7].
Kolmogorov–Arnold Networks (KANs) [8] are a recent advancement in machine learning, providing an innovative alternative to traditional Multi-Layer Perceptrons (MLPs). KANs leverage the decomposition of multivariate functions into univariate functions and linear combinations, enabling learnable activation functions on the edges (connections) between nodes. This design reduces computational complexity while enhancing performance, making KANs a promising solution for various deep learning (DL) applications.
While the specific advantages of KANs over MLPs are still being researched across different domains and scopes, their potential applicability to the LLIE domain is promising. LLIE inherently requires addressing both global context, such as overall brightness and contrast, and local context. Given KANs’ ability to approximate complex, non-linear functions and their expressive architecture, they align well with the intricate requirements of LLIE tasks. As highlighted in [9], KANs demonstrate enhanced capabilities for modeling complex transformations, both pixel-level precision and broader image-level characteristics.
In this work, we introduce a novel self-attention mechanism inspired by the Kolmogorov–Arnold representation theorem [10]. This theorem states that any multivariate continuous function can be expressed as a superposition of continuous functions and addition. In neural networks, KANs utilize this principle to break down multivariate functions into a series of univariate functions and linear combinations. Unlike traditional MLPs, which apply fixed activation functions to nodes and learnable weights to edges, KANs employ learnable activation functions on edges, offering greater flexibility. Drawing from these ideas and building upon our preliminary findings published in [11], our proposed attention mechanism adopts a similar approach to decompose multivariate functions into simpler, more manageable components.
Our primary contributions are outlined below:
  • We develop a robust Transformer self-attention mechanism that significantly outperforms conventional channel-wise multi-headed self-attention methods. This advancement leads to improved model performance and more effective feature representation.
  • We introduce a novel approach that integrates the principles of the Kolmogorov–Arnold representation theorem into attention mechanisms tailored for vision networks. This theoretical foundation enables more sophisticated and efficient feature map decomposition, enhancing the model’s capability to capture complex visual patterns.
  • Our proposed method achieves superior results on various LLIE benchmarks. These advancements are validated through extensive quantitative and qualitative evaluations, demonstrating the effectiveness and reliability of our approach in enhancing images under challenging low-light conditions, as can be seen in Figure 1.
This paper is organized as follows: Section 1 introduces the research and highlights its significance. Section 2, titled “Related Work”, provides an overview of the LLIE domain. In Section 3, “Methods”, we describe the proposed model and its key components. Section 4 presents the results obtained on the LOL dataset [3]. Finally, this paper concludes with an ablation study in Section 5 and a discussion of future research directions in Section 6.

2. Related Work

Early LLIE methods focused on directly enhancing the contrast and brightness of images using techniques such as histogram equalization and gamma correction [1,12,13,14]. While these approaches are straightforward, they often fail to address the complexities introduced by low-light conditions, such as noise and artifacts.
Traditional CV methods based on Retinex theory [2,15,16,17] decompose an image into reflectance (color information) and illumination components. These methods improve image quality by adjusting illumination; however, they often neglect the noise and artifacts prevalent in low-light environments.
The advent of DL has transformed LLIE, with CNNs becoming central to image restoration tasks. LLNet [18] was the first deep-learning-based model for low-light enhancement, while EnGAN [19] introduced a single generator model to directly transform low-light images into normal-light versions. Wei et al. [3] proposed a CNN-based Retinex decomposition framework to enhance the illumination component and reconstruct a well-lit image. Despite their success, these approaches often involve complex, multi-stage training pipelines and are computationally intensive. Additionally, CNNs struggle to capture long-range dependencies, which limits their adaptability in diverse low-light scenarios.
Transformers, originally developed for machine translation [20], have recently gained prominence in image restoration tasks [21,22,23,24]. ViTs [4] excel at modeling long-range dependencies through self-attention mechanisms, surpassing CNNs in various image enhancement tasks. For LLIE, methods like UFormer [25] and Retinexformer [7] have emerged, leveraging Transformers for superior performance. UFormer adapts the U-Net framework [26] by replacing convolutions with Transformer blocks while preserving the hierarchical encoder–decoder structure with skip connections. Retinexformer employs illumination representations to model non-local interactions across regions with varying lighting conditions.
KANs [8] represent a recent innovation in machine learning, offering an alternative to traditional MLPs. By decomposing multivariate functions into univariate functions and linear combinations, KANs introduce learnable activation functions on the edges between nodes. This reduces computational requirements and enhances performance, making KANs a promising solution for various DL applications.
While classical and traditional LLIE methods laid the groundwork for enhancing low-light images, they often fall short in addressing the intricacies of real-world scenarios. DL approaches, especially those using CNNs, have achieved significant advancements but are limited by their inability to capture global dependencies. ViTs have emerged as a powerful alternative, offering robust solutions to LLIE challenges by leveraging self-attention mechanisms. Furthermore, the innovative design of KANs introduces a novel perspective for function decomposition, providing additional opportunities for improvement in LLIE methodologies.

3. Methods

3.1. Overall Framework

Figure 2 illustrates the architecture of our proposed Transformer network, KAN- T , which employs a 3-level encoder-decoder structure. The input image is first passed through a c o n v 1 × 1 layer that performs feature expansion, from H × W × 3 to H × W × C . It is then processed by the encoder part of the network, which comprises Transformer blocks at different resolution levels, namely, H × W × C , H 2 × W 2 × 2 C , and H 4 × W 4 × 4 C . The encoder aims to transform the input image into an abstract internal representation that contains the key features in an image, which is then processed by the bottleneck section. The encoded feature map is then downsampled to H 8 × W 8 × 8 C and passed through the bottleneck of KAN- T , which utilizes four sequential Transformer blocks to enhance the internal feature representation. Then, the internal representation undergoes the decoding process, which consists of a suite of Transformer blocks at various levels, arranged symmetrically with respect to the encoder. The final H × W × C feature map undergoes a convolution to reduce the number of channels, producing the output image at H × W × 3 . KAN- T employs skip connections at corresponding encoder–decoder levels to help with detail preservation and feature enrichment.

3.2. Transformer Block

The Transformer block represents the main building block of KAN- T , and is used for its ability to perform advanced feature processing. As seen in Figure 2, the Transformer block consists of a Kolmogorov–Arnold Multi-headed Self-Attention (KAN-MSA) module, a Feed-Forward Network (FFN), and two Layer Normalization (LN) operations, while also employing residual connections between the two stages of self-attention and feature extraction. Given an input feature map F in R H × W × C , we can formulate the first stage of the Transformer block processing as follows:
F ^ = KAN-MSA ( LN ( F in ) ) + F in , F ^ R H × W × C ,
where F ^ represents the layer-normalized, self-attended, and residually enhanced input feature map. Finally, the intermediate representation F ^ undergoes the feature extraction stage of the Transformer block, where key features are enhanced, irrelevant features are discarded, and new features are discovered, and it can be expressed as
F out = FFN ( LN ( F ^ ) ) + F ^ , F out R H × W × C ,
where F out is the output feature map resulting from the Transformer block processing.

3.3. Kolmogorov–Arnold Network Multi-Headed Self-Attention

The Multi-headed Self-Attention (MSA) module represents the most important component in Transformer architectures. It utilizes multiple attention “heads”, allowing the model to focus on different parts of the input simultaneously. Each head learns to capture distinct features or relationships in the data, enabling effective information representations from various subspaces. In encoder–decoder ViT architectures, it helps the model capture spatial hierarchies and long-range pixel dependencies, making it suitable in vision tasks where high resolution is key.
Given an input feature map F in , it first extracts the Query ( Q ), Key ( K ), and Value ( V ) components, then computes the attention map using Q and K , and finally applies this attention map to V , resulting in the self-attended version of F in . Standard MSAs utilize fully connected (fc) layers to obtain Q , K , and V . As seen in Figure 3, fc layers first flatten the input feature map and compute a weighted sum utilizing a weight matrix W, followed by a fixed non-linear activation and a reshaping operation. Given an input feature map F in R H × W × C , this projection can be mathematically expressed as
Q = σ ( F in W Q + b Q ) , K = σ ( F in W K + b K ) , V = σ ( F in W V + b V ) ,
where σ ( · ) is an activation function; W Q , W K , and W V are weight matrices; and b Q , b K , and b V are bias terms. While fc layers can model complex relationships by processing the entire multivariate input jointly, they may not efficiently capture univariate relationships within individual channels. Additionally, they can be computationally intensive due to the large number of parameters, especially for high-dimensional inputs.
To address these limitations, we introduce a KAN-based MSA mechanism inspired by the Kolmogorov–Arnold representation theorem [10], which states that any multivariate continuous function can be represented as a superposition of continuous univariate functions and addition. Our method also incorporates the aspect of learnable non-linearity, as seen in the original KAN formulation [8]. Given the input feature map F in R H × W × C , we first formulate the multivariate decomposition by performing a channel-wise split as follows:
F in = [ F 1 , F 2 , , F C ] , F i R H × W × 1 , i = { 1 , 2 , , C } ,
thereby reducing the problem of processing F in at once—and therefore computing a multivariate function where each variable is a channel-wise component—to processing multiple single-channeled feature maps, enabling the model to capture more intricate and specific patterns in the data. Then, for each channel i, we process F i through a sequence of three fully connected layers with non-linear activations Φ j i as follows:
h 1 i = Φ i 1 ( W i 1 F i + b i 1 ) , h i 2 = Φ i 2 ( W i 2 h i 1 + b i 2 ) , h i 3 = Φ i 3 ( W i 3 h i 2 + b i 3 ) , h i 3 R H × W × 3
By employing three sequential fcs, we allow the model to activate or deactivate certain neurons as they go through the Φ j i activations, ensuring learnable non-linearity. Finally, the results of the univariate processing are concatenated in a channel-wise manner to obtain F out R H × W × 3 C , which is then split three-way to obtain Q , K , V R H × W × C . These are then reshaped to H W × C and used to produce the self-attended feature map F out as
F out = V × softmax ( K T × Q τ ) , F out R H W × C ,
where τ is a learnable parameter used to balance attention scores, and F out is later reshaped to H × W × C to preserve initial feature map dimensions.

3.4. Feed-Forward Network

The FFN is another key component of the Transformer block as it ensures in-depth feature extraction using the self-attended feature map. It follows a triple-convolution setup with Gaussian Error Linear Unit (GELU) [27] activations, denoted by ψ , and, given an input feature map F in R H × W × C , is formulated as follows:
F out = c o n v 1 × 1 ( ψ c o n v 3 × 3 ( ψ c o n v 1 × 1 ( F in ) ) ) , F out R H × W × C ,
where ψ c o n v 1 × 1 expands the feature map to H × W × 4 C to help with discovering new patterns, ψ c o n v 3 × 3 then performs high-resolution feature extraction by increasing the kernel size, and c o n v 1 × 1 compresses the feature map back to original dimensions H × W × C .

3.5. Loss Function

To achieve precise reconstruction, we employ a composite loss function, denoted as L . Similar to other works like [28,29,30], our hybrid loss integrates multiple components to address various aspects of image quality, including pixel-level accuracy, structural integrity, and perceptual fidelity. The overall loss is formulated as
L = L MAE + α · L MS-SSIM + β · L Perc
where α and β are hyperparameters that balance the contribution of each loss component.
Serving as the primary term in our loss function, the Mean Absolute Error (MAE) Loss L MAE captures the average differences between the predicted image I ^ and the ground truth image I GT . For a pixel at coordinates ( x , y ) , this loss is defined as
L MAE ( x , y ) = 1 N x , y I ^ ( x , y ) I GT ( x , y ) 1
The Multiscale Structural Similarity Index Measure Loss L MS-SSIM component evaluates the structural similarity [31] between the predicted and ground truth images across multiple scales. By assessing structural distortions, especially under challenging conditions like low-light scenarios, L MS-SSIM captures higher-level features that are crucial for maintaining the integrity of image structures. It is mathematically expressed as
L MS-SSIM = 1 m = 1 M 2 η m ζ m + C 1 η m 2 + ζ m 2 + C 1 δ m × 2 ξ m + C 2 κ m + λ m + C 2 ϵ m
Here, C 1 and C 2 are constants for luminance and contrast-structure stability, respectively, and M denotes the number of scales. For each scale m, η m and ζ m represent the means of I ^ and I GT , while κ m and λ m are their variances. The term ξ m stands for the covariance between I ^ and I GT , and δ m and ϵ m are weights assigned to luminance and contrast-structure components.
Finally, the Perceptual Loss L Perc [32] is used to incorporate feature-level supervision by leveraging a pre-trained VGG-19 network [33], denoted by Ψ . This loss measures the discrepancies between high-level feature representations of the predicted and ground truth images, facilitating the learning of meaningful internal representations, and is defined as
L Perc ( x , y ) = 1 N x , y Ψ ( I ^ ( x , y ) ) Ψ ( I GT ( x , y ) ) 1
In this equation, Ψ denotes the VGG-19 feature extractor, and · 1 measures the absolute differences between the feature maps of the predicted and ground truth images.
By integrating these three loss components, our hybrid loss function effectively balances pixel-level accuracy, structural consistency, and perceptual quality, leading to enhanced performance in RGB image reconstruction tasks.

3.6. Implementation Details

We train our model on datasets such as LOL-v1 [3] and LOL-v2 [34], including both Real and Synthetic partitions, and evaluate it on the corresponding benchmarks.
Architecturally, our Transformer employs a framework with three encoder levels and three decoder levels, comprising [ 1 , 2 , 2 ] Transformer blocks at each level, respectively. The model uses [ 2 , 2 , 4 ] attention heads at the corresponding levels, and the FFN expansion rate η is set to 4. The attention dimensions are set as [ 24 , 48 , 96 ] . The loss function parameters are set as α = 0.2 and γ = 0.01 .
Training is conducted on 256 × 256 patches with data augmentation techniques such as random cropping and random flipping, using a batch size of 2. We utilize the AdamW optimizer [35], with the parameters β 1 = 0.9 and β 2 = 0.999 and a weight decay of 1 × 10 4 , over 150,000 iterations. The learning rate starts at 3 × 10 4 and is reduced to 1 × 10 6 via the cosine annealing schedule [36].

4. Results

In this section, we evaluate our proposed method against other state-of-the-art approaches. First, we visually assess the results obtained on the LOL dataset. Second, we measure the quantitative performance of various methods using the Peak Signal-to-Noise Ratio (PSNR) and the Structural Similarity Index Measure (SSIM) [31].
Qualitative Results. Figure 4 presents qualitative comparisons of our model against several state-of-the-art methods. Approaches such as LLFormer [37] and LLFlow [38], while effective in improving brightness, often introduce unwanted lighting distortions, leading to overexposed or unevenly lit areas. Similarly, SNR-Aware [39] methods, although capable of denoising, frequently compromise the preservation of accurate color information, resulting in un-natural color reproduction. In contrast, our proposed KAN- T model demonstrates superior fidelity, closely matching the ground truth images. It achieves visually pleasing results that are both natural and richly detailed, preserving the subtle textures and true-to-life colors of the scene.
This balance is critical for real-world applications in domains such as surveillance, medical imaging, and autonomous driving, where the accurate reproduction of scene details under challenging lighting conditions is essential. KAN- T ’s ability to maintain high visual fidelity, while effectively enhancing low-light images, underscores its superiority in tackling the complex challenges of LLIE. Furthermore, its robustness against unwanted distortions ensures reliability in scenarios where image quality directly impacts performance and decision making, further validating its applicability across diverse fields.
Quantitative Results. Table 1 presents the quantitative performance of various LLIE methods on the LOL-v1, LOL-v2-Real (LOLv2-R), and LOL-v2-Synthetic (LOL-v2-S) datasets. On LOLv2, KAN- T outperforms the previous state-of-the-art model, Retinexformer, by an average of 0.24 dB in PSNR while maintaining competitive SSIM values, showcasing the effectiveness of the proposed KAN-based attention mechanism. Notably, KAN- T secures top ranks across multiple metrics, demonstrating its robustness in handling diverse lighting conditions and complex scenarios in low-light image enhancement.
This performance is further complemented by KAN- T ’s computational efficiency. While achieving superior results, it maintains a parameter count of only 2.80 M, significantly lower than competing methods such as LLFlow (37.68 M), LLFormer (24.55 M), and SNR-Aware (39.13 M). This reduced complexity makes KAN- T more suitable for real-time and resource-constrained applications without compromising quality. The model’s balanced approach to performance and efficiency highlights its potential for practical deployment in fields like surveillance, autonomous driving, and medical imaging, where both accuracy and speed are critical.
Overall, the results illustrate that KAN- T not only excels in quantitative metrics like PSNR and SSIM but also demonstrates scalability and adaptability to different datasets. Its superior performance on LOL-v2 and competitive results on LOL-v1 validate its robustness and make it a promising choice for advancing the state of the art in low-light image enhancement.
Figure 5 presents a comparative analysis of the PSNR performance of LLIE models presented in Table 1 on the LOL-v1, LOL-v2-R, and LOL-v2-S datasets against their respective parameter counts. This balance between high PSNR performance and efficient parameter usage demonstrates the strength and competitiveness of the proposed model.

5. Ablation Study

We conduct an ablation study on the LOL-v1 dataset to demonstrate the effectiveness of our proposed framework KAN- T , and utilize PSNR to measure performance, and number of parameters to evaluate complexity, where applicable.
We begin by evaluating the effectiveness of our proposed QKV extraction mechanism within the MSA framework, as detailed in Table 2a. The baseline model, which employs a standard MSA with fully connected (fc) layers, is the most lightweight, containing 1.29 million parameters. However, it yields a modest PSNR of 25.23 on the LOL-v1 validation set. Introducing our KAN-MSA with a depth of 1—where channel-wise processing is performed via a single fc layer in the KAN—results in a substantial improvement of 0.9 dB PSNR over the baseline. Further experimentation with increased depths of 3 and 5 demonstrates that a depth of 3 provides the highest PSNR gain of 1.43 dB, while a depth of 5 offers a smaller increase of 1.19 dB, indicating diminishing returns.
The effectiveness of our composite reconstruction loss L is substantiated through an ablation study presented in Table 2b. Starting with the foundational MAE loss ( L MAE ), which achieves a PSNR of 25.71, we observe that incorporating the Perceptual Loss ( L Perc ) alone enhances the PSNR to 26.21. This improvement highlights the significance of aligning high-level perceptual features in the reconstruction process. Similarly, integrating the MS-SSIM ( L MS-SSIM ) with L MAE results in a PSNR of 26.07, underscoring the role of structural consistency in achieving high-quality reconstructions. Most notably, the combination of all three loss components— L MAE , L Perc , and L MS-SSIM —yields the highest PSNR of 26.66. This demonstrates the effect of integrating pixel-level accuracy, perceptual fidelity, and structural integrity, thereby validating the necessity of a hybrid loss function for optimal RGB image reconstruction performance.

6. Conclusions and Future Work

In this paper, we introduce an innovative attention mechanism based on Kolmogorov–Arnold Networks (KANs) and seamlessly integrate it into a Transformer architecture, which we designate as KAN- T , specifically designed for low-light image enhancement (LLIE). By leveraging the principles of the Kolmogorov–Arnold representation theorem within vision networks, we have developed a sophisticated feature map decomposition strategy. This approach incorporates learnable non-linearities through the application of multiple non-linear activation functions, enabling the model to capture and process complex visual patterns more effectively.
Leveraging KANs in LLIE tasks seems promising, especially if computational resources and training setups can accommodate their demands. However, future work could empirically validate the advantages of KAN architectures over MLPs in different domains.
Our comprehensive study demonstrates the successful adaptation of Kolmogorov–Arnold Networks to vision-based Transformer architectures, showcasing their superior performance compared with traditional fully connected (fc) layers. The KAN- T model achieves state-of-the-art results across various LLIE benchmarks, underscoring its efficacy in enhancing image quality under low-light conditions. Beyond addressing the inherent limitations of conventional Convolutional Neural Networks (CNNs) and standard Transformer models, our proposed framework establishes a new benchmark in the realm of low-light image enhancement. This advancement is validated through extensive quantitative metrics and qualitative assessments, highlighting the model’s ability to deliver exceptional performance and set a new standard in the field.

Author Contributions

Conceptualization, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); methodology, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); software, A.B. and R.B.; validation, A.B., R.B. and C.O.; formal analysis, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); investigation, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); resources, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); data curation, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); writing—original draft preparation, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); writing—review and editing, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); visualization, A.B., R.B., C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); supervision, C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); project administration, C.O., C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti); funding acquisition, C.A. (Cosmin Ancuti) and C.A. (Codruta Ancuti). All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
CVcomputer vision
CNNConvolutional Neural Network
DLdeep learning
FFNFeed-Forward Network
GELUGaussian Error Linear Unit
KANKolmogorov–Arnold Network
KAN-MSAKolmogorov–Arnold Multi-headed Self-Attention
LLIElow-light image enhancement
LNLayer Normalization
MAEMean Absolute Error
MLPsMulti-Layer Perceptrons
MSAMulti-headed Self-Attention
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index Measure
ViTsVision Transformers

References

  1. Wang, W.; Wu, X.; Yuan, X.; Gao, Z. An experiment-based review of low-light image enhancement methods. IEEE Access 2020, 8, 87884–87917. [Google Scholar] [CrossRef]
  2. Land, E.H. The retinex theory of color vision. Sci. Am. 1977, 237, 108–129. [Google Scholar] [CrossRef] [PubMed]
  3. Wei, C.; Wang, W.; Yang, W.; Liu, J. Deep retinex decomposition for low-light enhancement. In Proceedings of the British Machine Vision Conference (BMVC), Newcastle, UK, 3–6 September 2018. [Google Scholar]
  4. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021. [Google Scholar] [CrossRef]
  5. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  6. Kim, H.; Choi, S.; Kim, C.; Koh, Y. Representative color transform for image enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021. [Google Scholar]
  7. Cai, Y.; Bian, H.; Lin, J.; Wang, H.; Timofte, R.; Zhang, Y. Retinexformer: One-stage Retinex-based Transformer for Low-light Image Enhancement. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 12504–12513. [Google Scholar]
  8. Liu, Z.; Wang, Y.; Vaidya, S.; Ruehle, F.; Halverson, J.; Soljačić, M.; Hou, T.Y.; Tegmark, M. KAN: Kolmogorov-Arnold Networks. arXiv 2024. [Google Scholar] [CrossRef]
  9. Yang, X.; Wang, X. Kolmogorov-arnold transformer. arXiv 2024, arXiv:2409.10594. [Google Scholar]
  10. Schmidt-Hieber, J. The Kolmogorov-Arnold representation theorem revisited. arXiv 2021. [Google Scholar] [CrossRef]
  11. Brateanu, A.; Balmez, R. Kolmogorov-Arnold Networks in Transformer Attention for Low-Light Image Enhancement. In Proceedings of the 2024 International Symposium on Electronics and Telecommunications (ISETC), Timisoara, Romania, 7–8 November 2024; pp. 1–4. [Google Scholar]
  12. Xiao, L.; Li, C.; Wu, Z.; Wang, T. An enhancement method for X-ray image via fuzzy noise removal and homomorphic filtering. Neurocomputing 2016, 195, 56–64. [Google Scholar] [CrossRef]
  13. Kim, S.E.; Jeon, J.J.; Eom, I.K. Image contrast enhancement using entropy scaling in wavelet domain. Signal Process. 2016, 127, 1–11. [Google Scholar] [CrossRef]
  14. Kansal, S.; Purwar, S.; Tripathi, R.K. Image contrast enhancement using unsharp masking and histogram equalization. Multimed. Tools Appl. 2018, 77, 26919–26938. [Google Scholar] [CrossRef]
  15. Park, S.; Moon, B.; Ko, S.; Yu, S.; Paik, J. Low-light image enhancement using variational optimization-based retinex model. IEEE Trans. Consum. Electron. 2017, 63, 178–184. [Google Scholar] [CrossRef]
  16. Gu, Z.; Li, F.; Fang, F.; Zhang, G. A novel retinex-based fractional order variational model for images with severely low light. IEEE Trans. Image Process. 2019, 29, 3239–3253. [Google Scholar] [CrossRef] [PubMed]
  17. Cai, B.; Xu, X.; Guo, K.; Jia, K.; Hu, B.; Tao, D. A jointbintrinsic-extrinsic prior model for retinex. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Venice, Italy, 22–29 October 2017. [Google Scholar]
  18. Lore, K.G.; Akintayo, A.; Sarkar, S. LLNet: A deep autoencoder approach to natural low-light image enhancement. Pattern Recognit. 2017, 61, 650–662. [Google Scholar] [CrossRef]
  19. Jiang, Y.; Gong, X.; Liu, D.; Cheng, Y.; Fang, C.; Shen, X.; Yang, J.; Zhou, P.; Wang, Z. Enlightengan: Deep light enhancement without paired supervision. TIP 2021, 30, 2340–2349. [Google Scholar] [CrossRef] [PubMed]
  20. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the NeurIPS, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  21. Cai, Y.; Lin, J.; Wang, H.; Yuan, X.; Ding, H.; Zhang, Y.; Timofte, R.; Van Gool, L. Degradation-Aware Unfolding Half-Shuffle Transformer for Spectral Compressive Imaging. In Proceedings of the NeurIPS, New Orleans, LA, USA, 28 November 2022. [Google Scholar]
  22. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-Trained Image Processing Transformer. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  23. Jang, S.I.; Pan, T.; Li, Y.; Heidari, P.; Chen, J.; Li, Q.; Gong, K. Spach Transformer: Spatial and channel-wise transformer based on local and global self-attentions for PET image denoising. IEEE Trans. Med. Imaging 2023, 43, 2036–2049. [Google Scholar] [CrossRef]
  24. Liu, D.; Mao, Y.; Zuo, Y.; An, P.; Fang, Y. Light Field Angular Super-Resolution Network Based on Convolutional Transformer and Deep Deblurring. IEEE Trans. Comput. Imaging 2024, 10, 1736–1748. [Google Scholar] [CrossRef]
  25. Wang, Z.; Cun, X.; Bao, J.; Liu, J. Uformer: A General U-Shaped Transformer for Image Restoration. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  26. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the MICCAI, Munich, Germany, 5–9 October 2015. [Google Scholar]
  27. Hendrycks, D.; Gimpel, K. Gaussian error linear units (gelus). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  28. Mohan, A.; Zhang, J.; Cozot, R.; Loscos, C. Consistent multi-and single-view hdr-image reconstruction from single exposures. In Proceedings of the Eurographics Workshop on Intelligent Cinematography and Editing, Reims, France, 28 April 2022. [Google Scholar]
  29. Brateanu, A.; Balmez, R.; Avram, A.; Orhei, C. LYT-Net: Lightweight YUV Transformer-based Network for Low-Light Image Enhancement. arXiv 2024. [Google Scholar] [CrossRef]
  30. Jia, L.; Huang, A.; He, X.; Li, Z.; Liang, J. A residual multi-scale feature extraction network with hybrid loss for low-dose computed tomography image denoising. Signal, Image Video Process. 2024, 18, 1215–1226. [Google Scholar] [CrossRef]
  31. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncell, E.P. Image quality assessment: From error visibility to structural similarity. TIP 2004, 13, 600–612. [Google Scholar] [CrossRef]
  32. Johnson, J.; Alahi, A.; Fei-Fei, L. Perceptual losses for real-time style transfer and super-resolution. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Proceedings, Part II 14. Springer: Berlin/Heidelberg, Germany, 2016; pp. 694–711. [Google Scholar]
  33. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2015. [Google Scholar] [CrossRef]
  34. Yang, W.; Wang, W.; Huang, H.; Wang, S.; Liu, J. Sparse gradient regularized deep retinex network for robust low-light image enhancement. TIP 2021, 30, 2072–2086. [Google Scholar] [CrossRef] [PubMed]
  35. Loshchilov, I.; Hutter, F. Decoupled Weight Decay Regularization. In Proceedings of the International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
  36. Loshchilov, I.; Hutter, F. Sgdr: Stochastic gradient descent with warm restarts. In Proceedings of the ICLR, Toulon, France, 24–26 April 2017. [Google Scholar]
  37. Wang, T.; Zhang, K.; Shen, T.; Luo, W.; Stenger, B.; Lu, T. Ultra-high-definition low-light image enhancement: A benchmark and transformer-based method. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 2654–2662. [Google Scholar]
  38. Wang, Y.; Wan, R.; Yang, W.; Li, H.; Chau, L.P.; Kot, A. Low-light image enhancement with normalizing flow. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 2604–2612. [Google Scholar]
  39. Xu, X.; Wang, R.; Fu, C.W.; Jia, J. SNR-Aware Low-Light Image Enhancement. In Proceedings of the CVPR, New Orleans, LA, USA, 18–24 June 2022. [Google Scholar]
  40. Chen, C.; Chen, Q.; Do, M.N.; Koltun, V. Seeing motion in the dark. In Proceedings of the ICCV, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  41. Wang, R.; Zhang, Q.; Fu, C.W.; Shen, X.; Zheng, W.S.; Jia, J. Underexposed Photo Enhancement Using Deep Illumination Estimation. In Proceedings of the CVPR, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  42. Zhang, Y.; Zhang, J.; Guo, X. Kindling the darkness: A practical low-light image enhancer. In Proceedings of the ACM MM, Nice, France, 21–25 October 2019. [Google Scholar]
  43. Zeng, H.; Cai, J.; Li, L.; Cao, Z.; Zhang, L. Learning image-adaptive 3d lookup tables for high performance photo enhancement in real-time. TPAMI 2020, 44, 2058–2073. [Google Scholar] [CrossRef] [PubMed]
  44. Kosugi, S.; Yamasaki, T. Unpaired Image Enhancement Featuring Reinforcement-Learning-Controlled Image Editing Software. In Proceedings of the AAAI, New York, NY, USA, 7–12 February 2020. [Google Scholar]
  45. Moran, S.; Marza, P.; McDonagh, S.; Parisot, S.; Slabaugh, G. Deeplpf: Deep local parametric filters for image enhancement. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  46. Xu, K.; Yang, X.; Yin, B.; Lau, R.W. Learning to restore low-light images via decomposition-and-enhancement. In Proceedings of the CVPR, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  47. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H.; Shao, L. Learning enriched features for real image restoration and enhancement. In Proceedings of the ECCV, Glasgow, UK, 23–28 August 2020. [Google Scholar]
  48. Liu, R.; Ma, L.; Zhang, J.; Fan, X.; Luo, Z. Retinex-inspired unrolling with cooperative prior architecture search for low-light image enhancement. In Proceedings of the CVPR, Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  49. Yang, W.; Wang, S.; Fang, Y.; Wang, Y.; Liu, J. Band representation-based semi-supervised low-light image enhancement: Bridging the gap between signal fidelity and perceptual quality. TIP 2021, 30, 3461–3473. [Google Scholar] [CrossRef] [PubMed]
  50. Yuhui, W.; Chen, P.; Guoqing, W.; Yang, Y.; Jiwei, W.; Chongyi, L.; Shen, H.T. Learning Semantic-Aware Knowledge Guidance for Low-Light Image Enhancement. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
Figure 1. Visual presentation of our proposed method for LLIE.
Figure 1. Visual presentation of our proposed method for LLIE.
Sensors 25 00327 g001
Figure 2. Overall framework of the proposed transformer. It presents a U-shaped arrangement with skip connections between same-level encoder–decoder blocks to help retain important dependencies.
Figure 2. Overall framework of the proposed transformer. It presents a U-shaped arrangement with skip connections between same-level encoder–decoder blocks to help retain important dependencies.
Sensors 25 00327 g002
Figure 3. Overview of the standard MSA mechanism on the left, and our proposed KAN-based MSA on the right.
Figure 3. Overview of the standard MSA mechanism on the left, and our proposed KAN-based MSA on the right.
Sensors 25 00327 g003
Figure 4. Comparison of qualitative results for different methods on LOL-v1 and LOL-v2-Real datasets. Red and green frames present zoomed-in regions for in-depth analysis.
Figure 4. Comparison of qualitative results for different methods on LOL-v1 and LOL-v2-Real datasets. Red and green frames present zoomed-in regions for in-depth analysis.
Sensors 25 00327 g004
Figure 5. Parameter vs. PSNR performance across LOL-v1, LOL-v2-R, and LOL-v2-S datasets.
Figure 5. Parameter vs. PSNR performance across LOL-v1, LOL-v2-R, and LOL-v2-S datasets.
Sensors 25 00327 g005
Table 1. Results on LOL-v1 and LOL-v2 datasets and parameter counts of different models. Red, blue, and green metrics represent first, second, and third places, respectively.
Table 1. Results on LOL-v1 and LOL-v2 datasets and parameter counts of different models. Red, blue, and green metrics represent first, second, and third places, respectively.
MethodsLOL-v1LOL-v2-RLOL-v2-SParam (M)
PSNRSSIMPSNRSSIMPSNRSSIM
RetinexNet [3] BMVC ’1816.770.46218.370.72317.130.7980.62
SID [40] ICCV ’1914.350.43613.240.44215.040.6107.76
DeepUPE [41] CVPR ’1914.380.44613.270.45215.080.6231.02
KinD [42] MM ’1920.870.79917.540.66916.260.5918.03
3DLUT [43] TPAMI ’2014.350.44517.590.72118.040.8000.59
RF [44] AAAI ’2015.230.45214.050.45815.970.63221.54
DeepLPF [45] CVPR ’2015.280.47314.100.48016.020.5871.77
FIDE [46] CVPR ’2018.270.66516.850.67815.200.6128.62
MIRNet [47] ECCV ’2024.140.84220.360.78221.940.8465.90
IPT [22] CVPR ’2116.270.50419.800.81318.300.811115.31
Sparse [34] TIP ’2117.200.64020.060.81622.050.9052.33
EnGAN [19] TIP ’2117.480.65218.640.67716.570.7348.64
RUAS [48] CVPR ’2118.230.72018.370.72316.550.6520.003
DRBN [49] TIP ’2119.860.83420.130.83023.220.9272.21
UFormer [25] CVPR ’2216.360.77118.820.77119.660.8715.29
Restormer [5] CVPR ’2222.430.82319.940.82721.410.83026.13
SNR-Net [39] CVPR ’2224.610.84221.480.84924.140.9284.01
LLFlow [38] AAAI ’2225.130.87226.200.88824.810.91937.68
Retinexformer [7] ICCV ’2325.160.84522.800.84025.670.9301.61
LLFormer [37] AAAI ’2325.760.82326.200.81928.010.92724.55
LL-SKF [50] CVPR ’2326.800.87928.450.90529.110.95339.91
KAN- T (Ours)26.660.85428.450.88428.770.9392.80
Table 2. Ablation studies on QKV extraction methods and loss functions. Red and blue metrics represent best and second-best results, respectively.
Table 2. Ablation studies on QKV extraction methods and loss functions. Red and blue metrics represent best and second-best results, respectively.
(a) Comparison of QKV Extraction Methods in MSA(b) Impact of Different Loss Functions on Performance
QKVParam. (M)PSNR L NAE L Perc L MS-SSIM PSNR
fc1.2925.23 25.71
KAN d = 1 1.7926.13 26.21
KAN d = 3 2.8026.66 26.07
KAN d = 5 3.8126.4226.66
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Brateanu, A.; Balmez, R.; Orhei, C.; Ancuti, C.; Ancuti, C. Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention. Sensors 2025, 25, 327. https://doi.org/10.3390/s25020327

AMA Style

Brateanu A, Balmez R, Orhei C, Ancuti C, Ancuti C. Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention. Sensors. 2025; 25(2):327. https://doi.org/10.3390/s25020327

Chicago/Turabian Style

Brateanu, Alexandru, Raul Balmez, Ciprian Orhei, Cosmin Ancuti, and Codruta Ancuti. 2025. "Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention" Sensors 25, no. 2: 327. https://doi.org/10.3390/s25020327

APA Style

Brateanu, A., Balmez, R., Orhei, C., Ancuti, C., & Ancuti, C. (2025). Enhancing Low-Light Images with Kolmogorov–Arnold Networks in Transformer Attention. Sensors, 25(2), 327. https://doi.org/10.3390/s25020327

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop