Next Article in Journal
Analysis of High-Dimensional Coordination in Human Movement Using Variance Spectrum Scaling and Intrinsic Dimensionality
Previous Article in Journal
A Survey on Semantic Communications in Internet of Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MambaOSR: Leveraging Spatial-Frequency Mamba for Distortion-Guided Omnidirectional Image Super-Resolution

by
Weilei Wen
*,
Qianqian Zhao
and
Xiuli Shao
*
VCIP, College of Computer Science, Nankai University, Tianjin 300350, China
*
Authors to whom correspondence should be addressed.
Entropy 2025, 27(4), 446; https://doi.org/10.3390/e27040446
Submission received: 3 March 2025 / Revised: 15 April 2025 / Accepted: 18 April 2025 / Published: 20 April 2025
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
Omnidirectional image super-resolution (ODISR) is critical for VR/AR applications, as high-quality 360° visual content significantly enhances immersive experiences. However, existing ODISR methods suffer from limited receptive fields and high computational complexity, which restricts their ability to model long-range dependencies and extract global structural features. Consequently, these limitations hinder the effective reconstruction of high-frequency details. To address these issues, we propose a novel Mamba-based ODISR network, termed MambaOSR, which consists of three key modules working collaboratively for accurate reconstruction. Specifically, we first introduce a spatial-frequency visual state space model (SF-VSSM) to capture global contextual information via dual-domain representation learning, thereby enhancing the preservation of high-frequency details. Subsequently, we design a distortion-guided module (DGM) that leverages distortion map priors to adaptively model geometric distortions, effectively suppressing artifacts resulting from equirectangular projections. Finally, we develop a multi-scale feature fusion module (MFFM) that integrates complementary features across multiple scales, further improving reconstruction quality. Extensive experiments conducted on the SUN360 dataset demonstrate that our proposed MambaOSR achieves a 0.16 dB improvement in WS-PSNR and increases the mutual information by 1.99% compared with state-of-the-art methods, significantly enhancing both visual quality and the information richness of omnidirectional images.

1. Introduction

In recent years, omnidirectional imaging technology has garnered substantial attention from both academic and industrial sectors, driven by the rapid development of immersive applications such as virtual reality (VR) and augmented reality (AR). These applications demand high-quality visual content, which typically requires resolutions of 4K or even 8K to ensure optimal viewing experiences. However, capturing such high-resolution omnidirectional images usually involves sophisticated and costly imaging equipment. Due to the resolution limitations of commonly available devices, ODISR has emerged as an essential research direction in this field. From an information theory perspective, low-resolution (LR) omnidirectional images exhibit diminished information content, resulting from sampling limitations, quantization errors, and sensor noise inherent in the acquisition process. Super-resolution (SR) techniques employ deep learning models to approximate high-frequency details through learned statistical priors, aiming to enhance visual fidelity.
An omnidirectional image exhibits fundamental differences from a traditional 2D image in geometric representation and storage formats. The most common formats, i.e., equirectangular projection (ERP) and cubemap projection (CMP), are designed to represent spherical content on planar surfaces. ERP maps the entire spherical view onto a rectangular grid, often introducing latitudinal stretching artifacts near polar regions, while CMP divides the sphere into six cube faces, reducing distortions but introducing face-boundary discontinuities. These unique properties pose challenges for effective feature extraction.
To address these distinctive characteristics inherent in omnidirectional images, researchers have developed dedicated SR architectures that explicitly model spherical geometry and distortion patterns. LAU-net [1] employs a latitude-based tiling strategy and implements independent processing pipelines to address varying distortion levels. However, such strategies are at risk of introducing boundary artifacts across adjacent tiles. In contrast, OSRT [2] incorporates a transformer-based architecture with a distortion-aware module that performs geometric rectification. Although OSRT demonstrates superior performance over previous ODISR approaches, its transformer-based design incurs quadratic complexity scaling with token size, presenting significant computational bottlenecks when processing high-resolution omnidirectional content.
Existing ODISR methods mainly leverage either CNN-based or transformer-based architectures to formulate their solutions. As shown in Figure 1, the CNN-based model, RCAN [3], is constrained by their local receptive fields, limiting their ability to model self-similarity patterns. This requirement is crucial in SR tasks, where pixel reconstruction relies on contextual information. In ODISR, this dependency intensifies due to geometric distortions demanding strong spatial awareness. Transformers provide global context but face computational challenges due to their quadratic complexity with token length. This motivates a fundamental question: How to design an efficient architecture that jointly addresses long-range dependency modeling in ODISR while enhancing information entropy through effective feature aggregation. Recent advances in state space models (SSMs), with their linear complexity scaling and global context modeling capabilities, present a promising direction to tackle this challenge.
The Mamba architecture [5] integrates SSMs with MLPs through an input-dependent selective scanning mechanism, enabling context-aware feature selection during long-range modeling while preserving task-specific representations. This design enhances the model’s capacity for discriminative representation learning, making it well suited for addressing key ODISR challenges such as the correction of geometric distortions and the preservation of high-frequency details. Therefore, we have adopted Mamba as the backbone architecture. Building upon this foundation, MambaIR [6] and MambaIRV2 [7] have established new benchmarks in image restoration by leveraging Mamba’s efficient long-range modeling capabilities. Despite the success of Mamba in sequence modeling tasks, adapting this architecture directly to high-resolution omnidirectional images remains challenging due to inherent spherical distortions in ERP representations. Addressing these distortions requires explicit modeling and expanded receptive fields to effectively capture the complex relationships within spherical content.
To bridge this gap, we introduce MambaOSR, the first Mamba-based framework for ODISR that integrates spatial-frequency modeling with distortion-aware scanning through a dual-path architecture. Our framework mitigates spherical distortions and utilizes spatial-frequency mixture representations to improve omnidirectional image reconstruction. Specifically, the key component of MambaOSR is the spatial-frequency Mamba block (SFMB), which contains a spatial-frequency visual state space model (SF-VSSM) and a distortion-guided module (DGM). The SF-VSSM integrates a frequency-aware module (FAM) with Mamba to adaptively extract frequency-domain information critical for ODISR. By effectively combining spatial- and frequency-domain features, the module enhances the long-range modeling capability of the Visual State Space Model (VSSM), resulting in more accurate feature reconstruction. Furthermore, to tackle distortions inherent to omnidirectional images, particularly along the latitudinal direction, we introduce the DGM. This module employs learned affine transformation parameters to dynamically adjust image feature mappings, thereby mitigating geometric distortions and improving reconstruction quality. Finally, the proposed multi-scale feature fusion module (MFFM) efficiently aggregates multi-resolution features to preserve fine-grained details in reconstructed images. Our method mitigates spherical distortion artifacts and enhances global context modeling, achieving high-fidelity reconstruction with linear computational complexity.
The main contributions are summarized as follows:
(1)
We introduce state space models into ODISR and propose an efficient network named MambaOSR. By leveraging the strong global modeling capabilities of the Mamba architecture, our method effectively captures long-range dependencies, significantly improving reconstruction quality. Extensive experiments validate the superior performance of MambaOSR compared to existing methods.
(2)
To further enhance global context modeling, we propose an SF-VSSM. Specifically, the SF-VSSM integrates an FAM with the VSSM to adaptively exploit frequency-domain information beneficial to the ODISR task. This integration enhances the model’s ability to capture global structural features and improves the reconstruction accuracy.
(3)
To address the degradation of image quality caused by geometric distortions inherent in omnidirectional imaging, we introduce a DGM. The DGM leverages distortion map priors to adaptively fuse the information of geometric deformation, effectively suppressing distortion artifacts. Additionally, we design an MFFM to integrate features across multiple scales, further strengthening the model’s representation capability and enhancing reconstruction performance.

2. Related Works

In this section, we revisit classical 2D SR and panoramic super-resolution methods, alongside SSM and Fourier transform techniques relevant to this work. Specifically, in the subsection on single-image super-resolution (SISR), we categorize the approaches into four groups: convolutional neural network (CNN)-based SISR methods, generative adversarial network (GAN)-based SISR methods, transformer-based SISR methods, and diffusion-based SISR methods.

2.1. Single-Image Super-Resolution

SISR is a crucial image reconstruction task aimed at reconstructing high-resolution (HR) images from low-resolution (LR) inputs. With the evolution of deep learning techniques, SISR techniques have gone through various approaches, including convolutional neural network CNN-based, GAN-based, transformer-based, and, more recently, diffusion-based models. Each approach offers unique characteristics and has demonstrated significant improvements across different application scenarios.

2.1.1. CNN-Based SISR Methods

CNN-based SISR methods [3,8,9,10] are among the earliest proposed approaches. A notable method is SRCNN [8], which utilizes standard convolutional layers to extract features and upscale LR images. It marks one of the pioneering deep learning models for SISR. Subsequently, RCAN [3] improved the model’s ability to capture fine details by integrating a residual channel attention mechanism. Building on this, HAN [9] introduced a holistic attention mechanism to enhance detail restoration. Although these methods demonstrate commendable performance in reconstructing complex structures and exhibit high computational efficiency, they often struggle to generate fine texture details.

2.1.2. GAN-Based SISR Methods

The application of generative adversarial networks [11] in SISR primarily focuses on generating realistic image textures. SRGAN [12] was the first model to apply GANs for super-resolution, using adversarial training to generate high-frequency details and enhance the visual quality of reconstructed images. Following this, ESRGAN [13] introduced dense skip connections and feature pooling, resulting in sharper and more detailed HR images. GAN-based SISR methods show clear advantages in visual quality, particularly in generating complex textures. However, they may also introduce artifacts or unrealistic details.

2.1.3. Transformer-Based SISR Methods

Recently, transformer architectures have demonstrated strong modeling capabilities in SISR tasks. IPT [14] is one of the first SISR models to incorporate the transformer architecture. It enhances the model’s generalization ability through multi-task learning and large-scale pre-training. SwinIR [4] utilizes the Swin transformer, segmenting images into local windows to improve feature extraction accuracy while effectively reducing computational overhead. HAT [15] employs a hierarchical aggregation mechanism to fuse multi-scale information, significantly enhancing the quality of image detail reconstruction. These methods excel in capturing long-range dependencies and complex textures but often require significant computational resources.

2.1.4. Diffusion-Based SISR Methods

Recently, diffusion models have gained attention in generative tasks, with SR3 [16] being a representative model that reconstructs HR images from LR inputs through a progressive denoising process. SRDiff [17] further optimizes the diffusion process, improving the speed and quality of image generation. Methods like ResShift [18] explore more efficient variants of diffusion models for image generation. Although these approaches have demonstrated significant potential in generating high-quality images, their iterative denoising processes can lead to longer inference times.
The evolution of SISR methods from CNNs to GANs, transformers, and diffusion models showcases continuous improvements in reconstruction quality, computational efficiency, and detail preservation. CNN-based methods are well suited for real-time applications, GANs excel in visual perception, and transformers and diffusion models offer significant advantages in high-quality image reconstruction. Future developments in SISR technology may further integrate the strengths of these diverse approaches to achieve higher quality and efficiency in SISR tasks.

2.2. Omnidirectional Image Super-Resolution (ODISR)

ODISR aims to enhance the resolution of omnidirectional images (ODIs), which are typically captured using omnidirectional cameras. Raw ODIs often appear in fisheye projections and are transformed into 2D images through different stitching techniques. Among these, ERP is the most commonly employed format for storing ODIs. Given its extensive applications in VR, AR, autonomous driving, and remote monitoring, ODISR has increasingly attracted research interest.
Early methods in this domain extended classical SISR techniques to tackle the unique challenges posed by ODIs. For instance, 360-SS [19] introduced an SISR model augmented with a spherical loss function, providing a straightforward approach to solving the ODISR problem. Similarly, Nishiyama et al. [20] incorporated distortion map information into the SISR framework to address deformation distortions in ODISR tasks. Although these methods offered initial solutions, they often fell short of fully leveraging the intrinsic characteristics of ODIs, limiting their effectiveness in addressing distortion-related artifacts.
Recently, researchers have increasingly focused on the challenge of uneven pixel density in ODIs. LAU-Net [1] addresses this issue by hierarchically partitioning ODIs based on latitude differences, enabling the network to process distinct latitude strips at varying levels. Building upon this framework, LAU-Net+ [21] introduces a multi-level pyramid network along with feature enhancement modules to further improve the performance of ODISR models. Despite their advancements, the hierarchical processing strategy, which relies on non-overlapping image patches, leads to noticeable artifacts at the boundaries of latitude strips in reconstructed ERP images. SphereSR [22] introduces a novel feature extraction module based on an icosahedral representation to effectively capture spherical features, utilizing a Spherical Local Implicit Image Function to predict RGB values in spherical coordinates. This approach enables the generation of continuous spherical images and supports flexible HR reconstruction across various projections. To explicitly address deformation distortions in ERP images, OSRT [2] incorporates a distortion-aware transformer with deformable convolutions. Similarly, OPDN [23] adopts a two-stage framework that integrates a position-aware deformable module alongside a frequency fusion module. BPOSR [24] leverages the complementary geometric properties of ERP and CMP projections through a dedicated attention mechanism, significantly improving the performance of ODISR. Cai et al. [25] proposed a spherical pseudo-cylindrical representation with adaptive latitude sampling and viewport-based loss for ODISR, which demonstrated promising performance while maintaining model-agnostic characteristics. Li et al. [26] presented OmniSSR, a zero-shot approach leveraging stable diffusion with tangent projections for ODISR, improving fidelity and realism without training. Yang et al. [27] adapted 2D planar SR models via low-rank adaptation (LoRA) for real-world ODISR, effectively reducing parameters and computational costs. Yang et al. [28] introduced GDGT-OSR with distortion-aware attention and dynamic feature aggregation guided by latitude-variant distortion modulation, achieving superior ODISR performance. Despite the progress made by existing ODISR methods, challenges remain, particularly in addressing ERP distortions and reducing computational complexity.

2.3. State Space Model (SSM)

In recent years, state space models grounded in control theory have achieved significant advances in deep learning. Compared to transformers, these models deliver competitive performance in long-range modeling tasks, with computational complexity scaling linearly with sequence length. Gu et al. [29] have established this linear growth advantage of SSMs when handling long sequence data. The introduction of the Mamba model, which outperforms transformers in natural language processing tasks, has garnered considerable attention. Due to its outstanding performance, the Mamba model has expanded rapidly into the computer vision domain, showing promising progress in tasks such as image restoration [6,7,30], image classification [31], and object detection [32]. However, the application of the Mamba model in foundational vision tasks remains in its early research stages, particularly in the context of ODISR, which remains to be explored.
This paper presents a novel framework for ODISR built upon the Mamba model. The proposed approach enables a more comprehensive global exploration of omnidirectional images by operating within the spatial-frequency dual domain. By integrating frequency domain information into the Mamba model, our method effectively enhances the extraction of high-frequency details, significantly improving the model’s ability to capture global features and structural information.

2.4. Fourier Transform

The Fourier transform has demonstrated an exceptional ability to capture global information, leading to its widespread adoption in many computer vision tasks [33,34,35,36,37] for frequency-domain representation. Frequency-domain characterization methods have driven significant breakthroughs in low-level visual tasks. The DeepRFT framework [38] leverages the global receptive field properties of Fourier transforms to capture both high- and low-frequency components for image deblurring. Building on the fast Fourier transform, the fast Fourier convolution (FFC) framework [39] establishes efficient paradigms for frequency-domain computation. Subsequently, Lama [40] developed an image inpainting model using fast Fourier convolution layers, achieving remarkable performance in reconstructing missing large-scale regions, complex geometric structures, and high-resolution images. Zhou et al. [41] present SFINet++, a spatial-frequency dual-domain network with invertible neural operators and deep Fourier transforms for multi-modal image fusion, outperforming SOTA in deep SR techniques. In the field of remote sensing image SR and ODISR, Xiao et al. [42] propose FMSR, the first Mamba-based framework for remote sensing image SR that integrates frequency-assisted modules and multi-level feature fusion, achieving linear-complexity long-range dependency modeling through learnable adaptors. Additionally, TSFNet [43] introduces a two-stage spatial-frequency joint learning framework, incorporating an amplitude-guided phase adaptive filter module and cross-stage feature fusion mechanism, enabling progressive refinement for large-factor remote sensing image SR. FATO [44] leverages Discrete Cosine Transforms (DCTs), frequency self-attention mechanisms, and frequency loss functions to enhance the preservation of high-frequency details in ODISR tasks.
A Fourier transform maps images from the spatial to the frequency domain, enabling models to capture global frequency characteristics and periodic patterns in images. This property is crucial for addressing geometric distortions commonly found in omnidirectional images. Although frequency-domain information is valuable, it alone cannot fully capture the local features of complex distortions. To overcome this limitation, we propose a frequency-aware module that leverages the Fourier transform for global frequency feature extraction while capturing local spatial details through convolution operations. By integrating spatial- and frequency-domain features, this module can expand the model’s receptive field and enhance the perception of high-frequency information. As illustrated in Figure 1, our method achieves a broader receptive field compared to both CNN-based and transformer-based approaches. Moreover, the computational complexity of our method scales linearly with the input size, whereas transformer-based methods exhibit quadratic growth.

3. Methodology

In this section, we first outline the overall framework of the proposed MambaOSR model, as shown in Figure 2. Subsequently, we introduce the commonly used ERP storage format for omnidirectional images, as shown in Figure 3, and we explain its fundamental principle. For ease of reference, the definitions of symbols and their meanings, which are discussed in Section 3, are listed in Table 1. We then elaborate on the key improvements introduced by Mamba for ODISR, which focus on three main components: the frequency-aware module (FAM) for capturing frequency-specific features, the distortion-guided module (DGM) for enhancing texture reconstruction in distorted regions, and the multi-scale feature fusion module (MFFM) for integrating features across multiple scales. Finally, we introduce the loss functions used in our approach.

3.1. Overview of MambaOSR

As depicted in Figure 2, we propose a Mamba-based network for ODISR that addresses the limited receptive fields hindering the reconstruction of details and the geometric distortions induced by the ERP. The proposed MambaOSR comprises three principal parts: the shallow feature extraction part, the deep feature extraction part, and the high-resolution image reconstruction part.
Given an LR image, shallow features are extracted via a 3 × 3 convolutional layer and enhanced through an MFFM for an adaptive feature representation. The processing procedure can be mathematically formulated as follows:
F 0 = MFFM 1 ( Conv ( I L R ) ) .
where MFFM 1 represents the first instance of the MFFM, and Conv denotes the operation of the 3 × 3 convolutional layer.
Subsequently, F 0 undergoes deep feature extraction through multiple cascaded spatial-frequency Mamba groups (SFMGs), followed by a convolution layer for feature refinement.
F n = Conv ( SFMG n ( SFMG n 1 ( SFMG 1 ( F 0 ) ) ) ) ,
where n denotes the number of SFMGs and F n is the output of the n-th SFMG. Each SFMG i is composed of a cascade of spatial-frequency Mamba blocks (SFMBs):
F i = Conv ( SFMB m ( SFMB m 1 ( SFMB 1 ( F i 1 ) ) ) ) + F i 1 , i = 1 , 2 , , n
where S F M B m denotes the m-th SFGB in each SFMG. Finally, the refined features are further enhanced by another MFFM and upsampled through sub-pixel convolution to reconstruct the super-resolved output I S R :
I S R = Conv ( Upsample ( MFFM 2 ( Conv ( F n ) + F 0 ) ) ) .

3.2. Equirectangular Projection

The ERP implements a linear mapping scheme that transforms spherical coordinates into a 2:1 rectangular domain, discretized into an h × w pixel grid where h and w denote the height and width of the target resolution. Assuming that longitude and latitude are denoted by φ and θ , respectively, with ( φ , θ ) π 2 , π 2 × [ π , π ] , the angular position ( φ , θ ) can be converted to Cartesian coordinates on a standard sphere Q s = ( q s x , q s y , q s z ) through the following relations:
(5) q s x = sin ( φ ) cos ( θ ) , (6) q s y = sin ( θ ) , (7) q s z = cos ( φ ) cos ( θ ) .
As illustrated in Figure 3a, the ERP projection is widely used for omnidirectional image storage due to its computational efficiency, algorithmic simplicity, and minimal distortion in equatorial regions. However, the uniform planar sampling of ERP results in a non-uniform spherical sampling density, which increases near the poles. This variation leads to anisotropic stretching artifacts, with significant geometric distortions becoming more pronounced in polar regions.
The angular coordinates ( θ , φ ) on the unit sphere are transformed into planar coordinates ( x , y ) via the coordinate transformation functions of the ERP. This process is mathematically expressed as
x = h ( θ , φ ) = θ , y = t ( θ , φ ) = φ .
h ( · ) and t ( · ) are coordinate transformation functions mapping spherical coordinates to the projection plane. As shown in Figure 3b, the original ODI exhibits geometric distortion when projected onto the ERP plane, characterized by latitudinal variation and hemispheric symmetry. This distortion motivates the formulation of a distortion map for low-resolution images. Given an LR image I L R R H × W × C i n (where H , W , C i n denote height, width, and channel dimensions, respectively), the corresponding distortion map D R H × W × 1 is formulated as follows [45]:
D ( h , 1 : W ) = cos h + 0.5 H 2 H π ,
where D ( h , 1 : W ) denotes the pixel stretching ratio from the ideal sphere to the 2D ERP image at height h. As shown in Figure 3, the distortion remains minimal in the equatorial region but intensifies toward higher latitudes. These distortions pose significant challenges for ODISR tasks.

3.3. Spatial-Frequency Visual State Space Model (SF-VSSM)

Previous ODISR methods often employ transformers to capture global dependencies via self-attention mechanisms for long-range response modeling. Despite their impressive performance, the high computational complexity of these methods hinders their efficiency in processing omnidirectional images. Inspired by the success of the VSSM in long-term modeling and aggregation with linear complexity, we pioneer its integration into ODISR tasks.
To enhance global information modeling, we introduce SF-VSSM, which contains two parallel modules: VSSM and FAM. The VSSM module, inherited from the Mamba architecture, performs dynamic state space modeling and visual feature extraction, while the proposed FAM enables effective spatial-frequency modeling, which significantly strengthens its ability to extract global information.
As shown in Figure 4, the FAM adopts a dual-branch framework: a convolution branch for spatial feature extraction and a spectral transform branch for frequency-domain feature modeling. The spectral transform branch utilizes the Real Fast Fourier Transform (Real FFT) to represent image features. By exploiting the conjugate symmetry property of the Fourier transform, Real FFT processes only half of the spectrum (positive frequencies), significantly improving the computational efficiency. The spectral transform branch leverages its global receptive field to capture global context, while the convolution branch employs local kernels to model fine-grained details. By integrating their outputs, the dual-branch architecture expands the model’s receptive field and mitigates information loss that typically occurs during localized processing in reconstruction tasks. This fusion enhances the retention of critical image features and improves the structural similarity between the reconstructed image and the original high-resolution image. The specific operational architecture of the spectral transform is detailed in Figure 4. From the perspective of information theory, this integration increases the mutual information between the reconstructed image and its high-resolution counterpart, ensuring a more precise reconstruction.
(a)
The Real FFT2d converts the feature tensor into the complex frequency domain
RealFFT 2 d : R H × W × C C H × W 2 × C ,
and concatenates its real and imaginary parts
ComplexToReal : C H × W 2 × C R H × W 2 × 2 C ;
(b)
Then, a convolutional block with activation and normalization processes the frequency spectrum information
ReLU BN Conv 1 × 1 : R H × W 2 × 2 C R H × W 2 × 2 C ;
(c)
The processed result is inversely transformed to restore spatial structure
RealToComplex : R H × W 2 × 2 C C H × W 2 × C ,
Inverse   Real   FFT 2 d : C H × W 2 × C R H × W × C .
Finally, the outputs of the convolution branch and spectral transform branch are fused through an addition operation.
In the SF-VSSM, the VSSM employs dynamic state space modeling to capture spatial-domain long-term dependencies, while FAM extracts spatial-frequency features through its dual-branch architecture for local detail and global context modeling. Here, LN represents the layer normalization operation. Additionally, we utilize a dynamic scaling factor to adaptively aggregate features based on their contextual significance. The output features y of the SF-VSSM can be formulated as follows:
y = α l · x l 1 + VSSM ( LN ( x l 1 ) ) + FAM ( LN ( x l 1 ) ) .

3.4. Distortion-Guided Module (DGM)

Given the unique deformation characteristics inherent in ERP omnidirectional images, we propose a DGM to address these imaging-specific distortions. As shown in Figure 2, the DGM processes features through three main steps: First, the input feature y undergoes layer normalization to stabilize the feature distribution. Second, a channel attention block (CAB) captures global contextual information. Finally, the Spatial Feature Transform (SFT) layer is used to adaptively fuse the distortion map information with the processed features, enabling precise enhancement of locally significant deformation patterns. The overall process can be formulated as
x l = α l + 1 · y + DGM ( LN ( y ) , c o n d i t i o n ) ,
where α l + 1 is a learnable scaling factor that adaptively integrates feature representations in a data-driven manner. The detailed transformation within the DGM can be further expressed as
DGM ( LN ( y ) , c o n d i t i o n ) = CAB ( LN ( y ) ) · γ + β ,
where the CAB processes normalized features, and the SFT layer modulates these features by applying condition-guided scaling factors γ = Sigmoid ( Conv ( c o n d i t i o n ) ) and additive biases β = Conv ( c o n d i t i o n ) , as shown in Figure 2.

3.5. Multi-Scale Feature Fusion Module (MFFM)

To address the non-uniform spatial resolution distribution in ODIs, we propose an MFFM. By employing multi-scale feature representations, the proposed MFFM effectively mitigates information loss during the SR process while naturally adapting to the inherent spatial variations of ODIs.
As shown in Figure 5, the MFFM employs a parallel multi-branch architecture to capture multi-scale feature representations. The module consists of multiple sparse CNN layers organized in parallel: a 3 × 3 average pooling layer for global context modeling, separable convolutional layers with varying kernel sizes (3 × 3, 5 × 5, 7 × 7) for local feature extraction at different scales, and dilated convolutional layers for expanding the receptive field while maintaining resolution.
To dynamically route features across branches, we integrate a self-attention mechanism as the core of the routing process. This mechanism adaptively weights the contribution of each branch based on input characteristics, enabling context-aware spatial filtering.
Specifically, given the input features, the module first performs average pooling along spatial dimensions (h and w) to compute channel-wise statistics in C dimensions z c R C ,
z c = 1 h × w i = 1 h j = 1 w x i n ( i , j ) ,
where x i n ( i , j ) is the ( i , j ) position of the feature x i n . Then, according to two learnable weight matrices, W 1 R O × T and W 2 R T × C , we generate weights w o u t for each branch:
w o u t = W 2 σ W 1 z c .
In addition, we perform zero-padding on the features to keep the size of the feature maps unchanged. The final output of the multi-scale fusion module can be formally expressed as
x o u t = f 1 × 1 c i = 1 O f b x o u t , w o u t + x i n ,
where f b and O denote the operations of each branch and the total number of branches, respectively. f 1 × 1 c represents a 1 × 1 convolution, and σ ( · ) is the ReLU activation function. [ · ] is the concatenation along the channel dimension. Using the MFFM, multi-scale information fusion can capture rich features in the omnidirectional images, providing more critical information for ODISR tasks.

3.6. Loss Function

The training process is guided by the L 1 loss between the super-resolved images I S R and their corresponding ground-truth counterparts I G T , formulated as
L s r = I S R I G T 1
where · 1 denotes the L 1 norm.

4. Experiments

4.1. Experimental Configuration

In this section, we conduct SR experiments on omnidirectional images using multiple public datasets, including the ODI-SR dataset [1] and the SUN360 [46] omnidirectional dataset. The ODI-SR dataset contains 1200 training images and 100 testing ODIs; the SUN360 dataset contains 100 testing images. Our MambaOSR comprises six SFMGs for deep feature extraction, with each SFMG containing six SFMBs. Empirically, we set the dimension of the internal channel to c = 96 . The proposed model is implemented in PyTorch 2.0.1 with end-to-end training conducted on four NVIDIA A40 GPUs.
According to [2], we clean the training image pairs using the method proposed in OSTR. For the × 4 , × 8 , and × 16 experiments, we create HR/LR image pairs by directly downsampling the ERP images. We train the model using L 1 loss and optimize it with the Adam optimizer for efficient convergence, with an initial learning rate set to 1 × 10 4 . In the × 4 experiments, the input image patch size is set to 64. For the × 8 and × 16 experiments, the LR image patch sizes are 32 and 16, respectively. The model is trained for 400,000 iterations. The learning rate is halved at 250,000 iterations to promote convergence. During the evaluation phase, we assess the model on the ODI-SR and SUN360 test sets, utilizing PSNR, SSIM [47], WS-PSNR [45], WS-SSIM [48], LPIPS [49] and Mutual Information (MI) [50] as evaluation metrics. Among these metrics, MI is employed to quantify the shared information between two images, providing a robust measure of their similarity.
To calculate MI, we first compute the entropy of individual images X and Y. The entropy values H ( X ) and H ( Y ) are defined as follows:
H ( X ) = i p ( x i ) log 2 ( p ( x i ) + ϵ ) ,
H ( Y ) = j p ( y j ) log 2 ( p ( y j ) + ϵ ) ,
where p ( x i ) and p ( y j ) are the probability distributions of the pixel intensity of images X and Y, respectively. To avoid numerical errors in logarithmic calculations, a small constant ϵ = 10 9 is added. Next, joint entropy H ( X , Y ) is computed as
H ( X , Y ) = i , j p ( x i , y j ) log 2 ( p ( x i , y j ) + ϵ ) ,
where p ( x i , y j ) is the joint probability distribution, obtained by normalizing the joint histogram of images X and Y. Finally, Mutual Information M I ( X , Y ) is computed as
M I ( X , Y ) = H ( X ) + H ( Y ) H ( X , Y ) .
This metric quantifies the shared information between images X and Y, making it useful for evaluating the similarity of reconstructed images in ODISR.

4.2. Evaluation Under ERP Downsampling

In the context of ERP downsampling, we compare nine representative SISR methods, including SRCNN [51], VDSR [52], LapSRN [53], MemNet [54], MSRN [55], EDSR [56], RCAN [3], HAT [15], MambaIR [6], and DRN [57], as well as five state-of-the-art (SOTA) ODISR algorithms: 360-SS [19], LAU-Net [1], SphereSR [22], OSRT [2], FATO [44] and BPOSR [24]. Our MambaOSR model is trained on the same ERP downsampling dataset to ensure fairness and consistency throughout the experimental process.
As shown in Table 2, our model demonstrates competitive performance for × 4 , × 8 , and × 16 upscaling factors on ERP downsampling datasets, ODI-SR and SUN360. Using the Mamba architecture as the backbone, our method significantly outperforms the baseline MambaIR across all test datasets. Compared to the state-of-the-art methods, our approach surpasses the Swin transformer-based BPOSR by 0.16 dB in PSNR and 0.0055 in SSIM on the SUN360 dataset for × 4 SR tasks. When evaluated against FATO, a novel frequency-based ODISR approach, our method shows notable improvements at × 4 and × 8 scaling factors, achieving a 0.23 dB gain on ODI-SR and a 0.16 dB gain on SUN360 for × 4 SR tasks, despite being slightly inferior at a × 16 scaling factor. Through quantitative analysis with the Mamba-based SISR method, MambaIR, and the state-of-the-art frequency-domain ODISR method, FATO, our method demonstrates superior overall performance in most scenarios, effectively validating its effectiveness and advancement.
In addition, we conducted experiments evaluating the LPIPS metric, which aligns closely with human perception and thus provides a more accurate measure of perceptual quality. As shown in Table 3, we compared our approach with several recent omnidirectional super-resolution algorithms. Our method demonstrates superior performance across multiple metrics, including PSNR, SSIM, and LPIPS. Specifically, on the LPIPS metric, our approach outperforms the latest state-of-the-art method, BPOSR, by margins ranging from 0.0095 to 0.0351 under various test conditions. This significant improvement highlights the ability of Mamba to extract global information and effectively integrate frequency-domain data, resulting in more perceptually accurate texture restoration.
Table 4 further highlights the superior MI performance of our model compared to other ODISR methods. On the SUN360 test set, our approach achieves MI improvements ranging from 1.99% to 37.95%. On the ODI-SR dataset, the MI gains range from 1.54% to 28.20%. As MI is a critical metric for quantifying structural similarity between super-resolved images and their high-resolution counterparts, these results highlight the robustness of our method. Complementary analyses using PSNR and SSIM confirm that our model achieves closer structural alignment with HR images than other approaches. These metrics provide a comprehensive evaluation of pixel-level accuracy and perceptual quality, highlighting the robustness of our approach.
As demonstrated in Figure 6 and Figure 7, our approach consistently outperforms existing methods, including CNN-based and transformer-based ODISR approaches, in preserving fine texture details across both datasets. Our distortion-aware modeling strategy significantly enhances the uniformity and fidelity of reconstructed line segments by effectively incorporating geometric deformation priors. In comparison, methods such as BPOSR exhibit texture distortion due to their improper handling of such priors.

4.3. Model Efficiency

To comprehensively evaluate the performance of different methods, we compare the number of parameters and computational complexity of each method. The specific results are shown in Figure 8. Our method has a noticeable advantage over OSRT [2] and SwinIR [4] in the number of parameters and computation abilities, and it achieves significantly better performance in the WS-SSIM metric. Compared to the recently proposed method, BPOSR [24], although it has advantages in parameter count and computation, our method outperforms BPOSR by 0.057 on the WS-SSIM metric. Furthermore, the input size of BPOSR is 32 times larger than ours. In particular, in the 8 × SR task, our method follows the design principle of OSTR with an input size of 32 × 32 , while the input size of BPOSR is 128 × 256 . It is worth noting that the input resolution is crucial in the SR task because it directly affects the model’s ability to perceive global information. Larger input sizes usually provide more details and contextual information, thus enhancing the model’s ability to capture global structures and details. Notably, even with this resolution disparity, our method outperforms BPOSR in both quantitative metrics and qualitative comparisons. The superior performance achieved with significantly smaller input patches highlights our model’s advantages concerning efficient information utilization and architectural design.
To further demonstrate the superiority of the Mamba architecture over the transformer architecture, we conducted a comparative analysis between our baseline model, MambaIR, and the widely adopted ODISR baseline model, SwinIR, in terms of their computational efficiencies. As shown in Figure 9a, MambaIR exhibits linear GPU memory consumption with increasing input size, similar to SwinIR, which employs efficient attention mechanisms. However, it is noteworthy that MambaIR maintains a global receptive field comparable to standard full attention, while SwinIR’s receptive field is inherently limited by its window size. Regarding inference time, as illustrated in Figure 9b, experimental results demonstrate MambaIR’s significant efficiency advantages. Particularly, when the input image size increases to 144 × 144 , SwinIR’s inference latency becomes approximately three times higher than that of MambaIR. These findings conclusively prove that the Mamba-based backbone network maintains global modeling capability and significantly outperforms the Swin transformer-based architecture in computational efficiency when processing large-scale input images.

4.4. Ablation Study

To gain a comprehensive analysis of MambaOSR, we conduct ablation studies on the effectiveness of each component in the proposed MambaOSR. To save time, all ablation experiments are based on modifications of the MambaIR-light model. The training dataset consists of 1200 omnidirectional images from ODISR [1], with an upscaling factor of × 8 . The testing dataset includes 100 test images from ODISR [1] and 100 test images from SUN360 [46]. In the following experiments, we use the same framework and ensure that the parameters across different variants are approximately the same to eliminate the influence of parameter variations. We change only one component in each experiment.

4.4.1. Effect of FAM

To investigate the effectiveness of the FAM in ODISR, we conduct ablation experiments by removing it from MambaOSR. We replace the FFC layer with three consecutive 3 × 3 convolutional layers to maintain comparable parameter counts. As shown in Table 5, the FAM-ablated variant (Model-1) exhibits a 0.2 dB reduction in PSNR compared to the complete MambaOSR architecture. Furthermore, Figure 10a compares the intermediate feature maps of MambaOSR and its FAM-ablated variant, revealing that MambaOSR generates more detailed textures in complex regions, particularly in the window areas. In contrast, the variant without frequency-domain processing exhibits noticeable blurring in its intermediate features. These comparative results demonstrate that our FAM enhances high-frequency information extraction through frequency-domain analysis, thereby improving reconstruction quality.

4.4.2. Effect of DGM

To evaluate the effectiveness of the DGM, we conduct an ablation study by replacing the DGM with a channel attention block in the MambaOSR architecture. In the ablated variant, we concatenate the distortion prior features with channel attention outputs along the channel dimension, followed by three 3 × 3 convolutional layers for feature fusion. As shown in Table 5, removing the DGM leads to a 0.36 dB decrease in PSNR compared to the MambaOSR architecture.
Furthermore, as illustrated in Figure 10b, in regions with severe geometric distortions (highlighted by red boxes), the DGM-ablated variant (Model-2) produces visible artifacts, while our MambaOSR successfully preserves fine texture details. These results demonstrate that the distortion-guided modulation enables adaptive feature processing based on local distortion patterns, facilitating more effective reconstruction in regions with varying distortion levels. The quantitative and qualitative improvements validate the effectiveness of our proposed DGM in handling geometric distortions in ODIs.

4.4.3. Effect of MFFM

Through ablation studies where we remove the MFFM from MambaOSR (Model-3), we observe a 0.14 dB reduction in PSNR compared to the MambaOSR architecture. This performance degradation suggests that the MFFM effectively aggregates hierarchical feature representations across different scales. The module’s cross-scale integration capability provides complementary contextual information for the reconstruction process, enabling more effective utilization of multi-resolution features. As indicated in Table 5, quantitative comparisons demonstrate that this architectural component enhances texture preservation in high-frequency regions.

5. Conclusions

This work presents the first application of the Mamba architecture to ODISR tasks. By leveraging the long-range modeling capabilities of the VSSM and integrating spatial-frequency dual-domain insights into ODIs, we propose the MambaOSR framework. To overcome the limitations of conventional approaches in global modeling, we design the SFMB, which incorporates two key components: an FAM for global contextual reconstruction through representation learning, and a DGM for reducing geometric distortion artifacts. Specifically, the FAM enables effective global context modeling, while the DGM adaptively modulates features based on local distortion patterns derived from distortion prior maps, significantly improving reconstruction quality in severely distorted regions. To further enhance feature representation, we introduce an MFFM that facilitates cross-scale semantic integration during shallow feature extraction and high-level reconstruction stages. The extensive experimental results on multiple benchmark datasets demonstrate that our approach achieves state-of-the-art performance while maintaining competitive computational efficiency. It strikes an optimal balance between reconstruction quality and processing speed.

Author Contributions

Conceptualization, methodology, and writing—original draft preparation, W.W.; validation and writing—review and editing, Q.Z. and X.S.; resources, X.S.; supervision, X.S.; funding acquisition, X.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant No. 12275142.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

The original data presented in this study are openly available at the LAU-Net repository (https://github.com/wangh-allen/LAU-Net, accessed on 1 March 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Deng, X.; Wang, H.; Xu, M.; Guo, Y.; Song, Y.; Yang, L. Lau-net: Latitude adaptive upscaling network for omnidirectional image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 9189–9198. [Google Scholar]
  2. Yu, F.; Wang, X.; Cao, M.; Li, G.; Shan, Y.; Dong, C. Osrt: Omnidirectional image super-resolution with distortion-aware transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 13283–13292. [Google Scholar]
  3. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  4. Liang, J.; Cao, J.; Sun, G.; Zhang, K.; Van Gool, L.; Timofte, R. Swinir: Image restoration using swin transformer. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 1833–1844. [Google Scholar]
  5. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  6. Guo, H.; Li, J.; Dai, T.; Ouyang, Z.; Ren, X.; Xia, S.T. Mambair: A simple baseline for image restoration with state-space model. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer: Berlin/Heidelberg, Germany, 2024; pp. 222–241. [Google Scholar]
  7. Guo, H.; Guo, Y.; Zha, Y.; Zhang, Y.; Li, W.; Dai, T.; Xia, S.T.; Li, Y. MambaIRv2: Attentive State Space Restoration. arXiv 2024, arXiv:2411.15269. [Google Scholar]
  8. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 38, 295–307. [Google Scholar] [CrossRef]
  9. Niu, B.; Wen, W.; Ren, W.; Zhang, X.; Yang, L.; Wang, S.; Zhang, K.; Cao, X.; Shen, H. Single image super-resolution via a holistic attention network. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XII 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 191–207. [Google Scholar]
  10. Zang, H.; Zhao, Y.; Niu, C.; Zhang, H.; Zhan, S. Attention network with information distillation for super-resolution. Entropy 2022, 24, 1226. [Google Scholar] [CrossRef]
  11. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
  12. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  13. Wang, X.; Yu, K.; Wu, S.; Gu, J.; Liu, Y.; Dong, C.; Qiao, Y.; Change Loy, C. Esrgan: Enhanced super-resolution generative adversarial networks. In Proceedings of the European Conference on Computer Vision (ECCV) Workshops, Munich, Germany, 8–14 September 2018. [Google Scholar]
  14. Chen, H.; Wang, Y.; Guo, T.; Xu, C.; Deng, Y.; Liu, Z.; Ma, S.; Xu, C.; Xu, C.; Gao, W. Pre-trained image processing transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 12299–12310. [Google Scholar]
  15. Chen, X.; Wang, X.; Zhou, J.; Qiao, Y.; Dong, C. Activating more pixels in image super-resolution transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 22367–22377. [Google Scholar]
  16. Saharia, C.; Ho, J.; Chan, W.; Salimans, T.; Fleet, D.J.; Norouzi, M. Image super-resolution via iterative refinement. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 4713–4726. [Google Scholar] [CrossRef]
  17. Li, H.; Yang, Y.; Chang, M.; Chen, S.; Feng, H.; Xu, Z.; Li, Q.; Chen, Y. Srdiff: Single image super-resolution with diffusion probabilistic models. Neurocomputing 2022, 479, 47–59. [Google Scholar] [CrossRef]
  18. Yue, Z.; Wang, J.; Loy, C.C. Resshift: Efficient diffusion model for image super-resolution by residual shifting. Adv. Neural Inf. Process. Syst. 2023, 36, 13294–13307. [Google Scholar]
  19. Ozcinar, C.; Rana, A.; Smolic, A. Super-resolution of omnidirectional images using adversarial learning. In Proceedings of the 2019 IEEE 21st International Workshop on Multimedia Signal Processing (MMSP), Kuala Lumpur, Malaysia, 27–29 September 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  20. Nishiyama, A.; Ikehata, S.; Aizawa, K. 360 single image super resolution via distortion-aware network and distorted perspective images. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1829–1833. [Google Scholar]
  21. Deng, X.; Wang, H.; Xu, M.; Li, L.; Wang, Z. Omnidirectional image super-resolution via latitude adaptive network. IEEE Trans. Multimed. 2022, 25, 4108–4120. [Google Scholar] [CrossRef]
  22. Yoon, Y.; Chung, I.; Wang, L.; Yoon, K.J. Spheresr: 360deg image super-resolution with arbitrary projection via continuous spherical image representation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5677–5686. [Google Scholar]
  23. Sun, X.; Li, W.; Zhang, Z.; Ma, Q.; Sheng, X.; Cheng, M.; Ma, H.; Zhao, S.; Zhang, J.; Li, J.; et al. OPDN: Omnidirectional position-aware deformable network for omnidirectional image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 1293–1301. [Google Scholar]
  24. Wang, J.; Cui, Y.; Li, Y.; Ren, W.; Cao, X. Omnidirectional Image Super-resolution via Bi-projection Fusion. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 5454–5462. [Google Scholar]
  25. Cai, Q.; Li, M.; Ren, D.; Lyu, J.; Zheng, H.; Dong, J.; Yang, Y.H. Spherical pseudo-cylindrical representation for omnidirectional image super-resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, Vancouver, BC, Canada, 20–27 February 2024; Volume 38, pp. 873–881. [Google Scholar]
  26. Li, R.; Sheng, X.; Li, W.; Zhang, J. OmniSSR: Zero-shot Omnidirectional Image Super-Resolution using Stable Diffusion Model. In Proceedings of the European Conference on Computer Vision, Milan, Italy, 29 September–4 October 2024; Springer: Berlin/Heidelberg, Germany, 2025; pp. 198–216. [Google Scholar]
  27. Yang, C.; Dong, R.; Lam, K.M. Efficient Adaptation for Real-World Omnidirectional Image Super-Resolution. In Proceedings of the 2024 Asia Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Macau, China, 3–6 December 2024; IEEE: Piscataway, NJ, USA, 2024; pp. 1–6. [Google Scholar]
  28. Yang, C.; Dong, R.; Xiao, J.; Zhang, C.; Lam, K.M.; Zhou, F.; Qiu, G. Geometric distortion guided transformer for omnidirectional image super-resolution. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
  29. Gu, A.; Goel, K.; Ré, C. Efficiently modeling long sequences with structured state spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  30. Shi, Y.; Xia, B.; Jin, X.; Wang, X.; Zhao, T.; Xia, X.; Xiao, X.; Yang, W. Vmambair: Visual state space model for image restoration. IEEE Trans. Circuits Syst. Video Technol. 2025. [Google Scholar] [CrossRef]
  31. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Jiao, J.; Liu, Y. Vmamba: Visual state space model. Adv. Neural Inf. Process. Syst. 2025, 37, 103031–103063. [Google Scholar]
  32. Dong, W.; Zhu, H.; Lin, S.; Luo, X.; Shen, Y.; Liu, X.; Zhang, J.; Guo, G.; Zhang, B. Fusion-mamba for cross-modality object detection. arXiv 2024, arXiv:2404.09146. [Google Scholar]
  33. Xu, Q.; Zhang, R.; Zhang, Y.; Wang, Y.; Tian, Q. A fourier-based framework for domain generalization. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14383–14392. [Google Scholar]
  34. Yang, Y.; Soatto, S. Fda: Fourier domain adaptation for semantic segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 4085–4095. [Google Scholar]
  35. Li, C.; Guo, C.L.; Zhou, M.; Liang, Z.; Zhou, S.; Feng, R.; Loy, C.C. Embedding fourier for ultra-high-definition low-light image enhancement. arXiv 2023, arXiv:2302.11831. [Google Scholar]
  36. Lee, J.H.; Heo, M.; Kim, K.R.; Kim, C.S. Single-image depth estimation based on fourier domain analysis. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 330–339. [Google Scholar]
  37. Huang, J.; Liu, Y.; Zhao, F.; Yan, K.; Zhang, J.; Huang, Y.; Zhou, M.; Xiong, Z. Deep fourier-based exposure correction network with spatial-frequency interaction. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 163–180. [Google Scholar]
  38. Mao, X.; Liu, Y.; Liu, F.; Li, Q.; Shen, W.; Wang, Y. Intriguing findings of frequency selection for image deblurring. In Proceedings of the AAAI Conference on Artificial Intelligence, Washington, DC, USA, 7–14 February 2023; Volume 37, pp. 1905–1913. [Google Scholar]
  39. Chi, L.; Jiang, B.; Mu, Y. Fast fourier convolution. Adv. Neural Inf. Process. Syst. 2020, 33, 4479–4488. [Google Scholar]
  40. Suvorov, R.; Logacheva, E.; Mashikhin, A.; Remizova, A.; Ashukha, A.; Silvestrov, A.; Kong, N.; Goka, H.; Park, K.; Lempitsky, V. Resolution-robust large mask inpainting with fourier convolutions. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2022; pp. 2149–2159. [Google Scholar]
  41. Zhou, M.; Huang, J.; Yan, K.; Hong, D.; Jia, X.; Chanussot, J.; Li, C. A general spatial-frequency learning framework for multimodal image fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 1–18. [Google Scholar] [CrossRef]
  42. Xiao, Y.; Yuan, Q.; Jiang, K.; Chen, Y.; Zhang, Q.; Lin, C.W. Frequency-assisted mamba for remote sensing image super-resolution. IEEE Trans. Multimed. 2024, 27, 1783–1796. [Google Scholar] [CrossRef]
  43. Wang, J.; Lu, Y.; Wang, S.; Wang, B.; Wang, X.; Long, T. Two-stage spatial-frequency joint learning for large-factor remote sensing image super-resolution. IEEE Trans. Geosci. Remote Sens. 2024, 62, 1–13. [Google Scholar] [CrossRef]
  44. An, H.; Zhang, X.; Zhao, S.; Zhang, L. FATO: Frequency Attention Transformer for Omnidirectional Image Super-Resolution. In Proceedings of the 6th ACM International Conference on Multimedia in Asia, Auckland, New Zealand, 3–6 December 2024; pp. 1–7. [Google Scholar]
  45. Sun, Y.; Lu, A.; Yu, L. Weighted-to-spherically-uniform quality evaluation for omnidirectional video. IEEE Signal Process. Lett. 2017, 24, 1408–1412. [Google Scholar] [CrossRef]
  46. Xiao, J.; Ehinger, K.A.; Oliva, A.; Torralba, A. Recognizing scene viewpoint using panoramic place representation. In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 2695–2702. [Google Scholar]
  47. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  48. Zhou, Y.; Yu, M.; Ma, H.; Shao, H.; Jiang, G. Weighted-to-spherically-uniform SSIM objective quality evaluation for panoramic video. In Proceedings of the 2018 14th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 12–16 August 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 54–57. [Google Scholar]
  49. Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the CVPR, Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
  50. Latham, P.E.; Roudi, Y. Mutual information. Scholarpedia 2009, 4, 1658. [Google Scholar] [CrossRef]
  51. Dong, C.; Loy, C.C.; He, K.; Tang, X. Learning a deep convolutional network for image super-resolution. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Proceedings, Part IV 13. Springer: Berlin/Heidelberg, Germany, 2014; pp. 184–199. [Google Scholar]
  52. Kim, J.; Lee, J.K.; Lee, K.M. Accurate image super-resolution using very deep convolutional networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1646–1654. [Google Scholar]
  53. Ahn, N.; Kang, B.; Sohn, K.A. Fast, accurate, and lightweight super-resolution with cascading residual network. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 252–268. [Google Scholar]
  54. Tai, Y.; Yang, J.; Liu, X.; Xu, C. Memnet: A persistent memory network for image restoration. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4539–4547. [Google Scholar]
  55. Li, J.; Fang, F.; Mei, K.; Zhang, G. Multi-scale residual network for image super-resolution. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 517–532. [Google Scholar]
  56. Lim, B.; Son, S.; Kim, H.; Nah, S.; Mu Lee, K. Enhanced deep residual networks for single image super-resolution. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Honolulu, HI, USA, 21–26 July 2017; pp. 136–144. [Google Scholar]
  57. Guo, Y.; Chen, J.; Wang, J.; Chen, Q.; Cao, J.; Deng, Z.; Xu, Y.; Tan, M. Closed-loop matters: Dual regression networks for single image super-resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5407–5416. [Google Scholar]
Figure 1. Comparison of Effective Receptive Field (ERF) for (a) the CNN-based method, RCAN [3], (b) the transformer-based model, SwinIR [4], and (c) the proposed Mamba-based network, MambaOSR. A larger ERF is represented by a wider distribution of dark areas. The proposed MambaOSR achieves the largest ERF, demonstrating its superior global context modeling capability.
Figure 1. Comparison of Effective Receptive Field (ERF) for (a) the CNN-based method, RCAN [3], (b) the transformer-based model, SwinIR [4], and (c) the proposed Mamba-based network, MambaOSR. A larger ERF is represented by a wider distribution of dark areas. The proposed MambaOSR achieves the largest ERF, demonstrating its superior global context modeling capability.
Entropy 27 00446 g001
Figure 2. The overall architecture of the proposed MambaOSR. The key component of MambaOSR is the SFMB, consisting of the SF-VSSM and DGM. The SF-VSSM integrates Mamba with an FAM for adaptive frequency-domain feature extraction in ODISR, while the DGM enhances distortion representation to enable accurate texture recovery through distortion-aware learning. Additionally, the MFFM performs multi-scale feature integration to improve texture detail reconstruction through adaptive fusion mechanisms.
Figure 2. The overall architecture of the proposed MambaOSR. The key component of MambaOSR is the SFMB, consisting of the SF-VSSM and DGM. The SF-VSSM integrates Mamba with an FAM for adaptive frequency-domain feature extraction in ODISR, while the DGM enhances distortion representation to enable accurate texture recovery through distortion-aware learning. Additionally, the MFFM performs multi-scale feature integration to improve texture detail reconstruction through adaptive fusion mechanisms.
Entropy 27 00446 g002
Figure 3. Illustration of the equirectangular projection (ERP) for 360° images. (a) Mapping of the spherical image to a 2D rectangular format with longitude ( θ ( π , π ) ) and latitude ( φ ( π / 2 , π / 2 ) ). (b) A distortion map indicating visual distortion exists in the ERP format, highlighting a non-uniform pixel distribution across the projection. The darker regions of the distortion map indicate a higher degree of distortion.
Figure 3. Illustration of the equirectangular projection (ERP) for 360° images. (a) Mapping of the spherical image to a 2D rectangular format with longitude ( θ ( π , π ) ) and latitude ( φ ( π / 2 , π / 2 ) ). (b) A distortion map indicating visual distortion exists in the ERP format, highlighting a non-uniform pixel distribution across the projection. The darker regions of the distortion map indicate a higher degree of distortion.
Entropy 27 00446 g003
Figure 4. Overview of the FAM: The FAM processes features through parallel spatial and frequency branches. The frequency branch applies FFT-based spectral convolution and IFFT reconstruction, while the spatial branch uses conventional convolutions. Cross-domain feature fusion integrates spatial and frequency information to enhance high-frequency texture reconstruction.
Figure 4. Overview of the FAM: The FAM processes features through parallel spatial and frequency branches. The frequency branch applies FFT-based spectral convolution and IFFT reconstruction, while the spatial branch uses conventional convolutions. Cross-domain feature fusion integrates spatial and frequency information to enhance high-frequency texture reconstruction.
Entropy 27 00446 g004
Figure 5. Illustration of the architecture of our proposed MFFM. The MFFM comprises six different convolutional layers with an adaptive weighting mechanism that dynamically fuses multi-scale features in an input-dependent manner.
Figure 5. Illustration of the architecture of our proposed MFFM. The MFFM comprises six different convolutional layers with an adaptive weighting mechanism that dynamically fuses multi-scale features in an input-dependent manner.
Entropy 27 00446 g005
Figure 6. Visual comparisons of our MambaOSR with both SISR and ODISR methods on ODISR datasets on a ×8 scale. Our method demonstrates significant advantages over existing approaches in accurately reconstructing line structures. Zoom in for a better view.
Figure 6. Visual comparisons of our MambaOSR with both SISR and ODISR methods on ODISR datasets on a ×8 scale. Our method demonstrates significant advantages over existing approaches in accurately reconstructing line structures. Zoom in for a better view.
Entropy 27 00446 g006
Figure 7. Visual comparisons of our MambaOSR with both SISR and ODISR methods on SUN360 datasets with a ×8 scaling factor. Our method shows significant improvements in reconstructing line structures compared to other approaches. Zoom in for a better view.
Figure 7. Visual comparisons of our MambaOSR with both SISR and ODISR methods on SUN360 datasets with a ×8 scaling factor. Our method shows significant improvements in reconstructing line structures compared to other approaches. Zoom in for a better view.
Entropy 27 00446 g007
Figure 8. Comparison of WS-SSIM versus the number of parameters across various models on the SUN-SR test set with a × 8 upscaling factor. FLOPs (in G) are calculated based on LR input with a resolution of 256 × 128 .
Figure 8. Comparison of WS-SSIM versus the number of parameters across various models on the SUN-SR test set with a × 8 upscaling factor. FLOPs (in G) are calculated based on LR input with a resolution of 256 × 128 .
Entropy 27 00446 g008
Figure 9. Comparison of (a) GPU memory consumption and (b) inference latency across varying input resolutions. We chose SwinIR (a Swin transformer-based method) and MambaIR (a Mamba-based method) for comparison. We adjust the model to ensure the GPU usage is roughly similar at the beginning, and then we increase the input resolution from 64 × 64 to 144 × 144 .
Figure 9. Comparison of (a) GPU memory consumption and (b) inference latency across varying input resolutions. We chose SwinIR (a Swin transformer-based method) and MambaIR (a Mamba-based method) for comparison. We adjust the model to ensure the GPU usage is roughly similar at the beginning, and then we increase the input resolution from 64 × 64 to 144 × 144 .
Entropy 27 00446 g009
Figure 10. Comparison of feature maps in ablation studies. (a) Ablation study on the FAM: Feature maps with an FAM exhibit an enhanced ability to extract high-frequency information, resulting in sharper edges and finer texture details, while those without an FAM show blurred textures and loss of detail. (b) Ablation study on the DGM: Feature maps with a DGM effectively restore texture details in distorted regions, addressing projection distortions, whereas those without a DGM display limited recovery and less coherent structures.
Figure 10. Comparison of feature maps in ablation studies. (a) Ablation study on the FAM: Feature maps with an FAM exhibit an enhanced ability to extract high-frequency information, resulting in sharper edges and finer texture details, while those without an FAM show blurred textures and loss of detail. (b) Ablation study on the DGM: Feature maps with a DGM effectively restore texture details in distorted regions, addressing projection distortions, whereas those without a DGM display limited recovery and less coherent structures.
Entropy 27 00446 g010
Table 1. Definitions of symbols (and their meanings) used in Section 3.
Table 1. Definitions of symbols (and their meanings) used in Section 3.
SymbolDescription
SFMGSpatial-frequency Mamba group
SFMBSpatial-frequency Mamba block
SF-VSSMSpatial-frequency visual state space model
FAMFrequency-aware module
DGMDistortion-guided module
MFFMMulti-scale feature fusion module
φ , θ Longitude and latitude
Q s = ( q s x , q s y , q s z ) Cartesian coordinates
h ( · ) , t ( · ) Coordinate transformation (sphere to 2D)
D ( h , 1 : W ) Pixel stretching ratio at height h
H , W , C i n Height, width, and channel dimensions
α l , α l + 1 Learnable scaling factors
x i n , x o u t Input and output features of the MFFM
z c Channel-wise statistics
W 1 , W 2 Learnable weight matrices
h , w Feature height and width
w o u t Branch weights in the MFFM
f b , O Branch operation and total branches
σ ( · ) ReLU activation function
Table 2. Quantitative comparisons (WS-PSNR/WS-SSIM) with SISR and ODISR algorithms on ODI-SR and SUN360 datasets, where the best results are highlighted in red. Note that our method follows OSRT by fixing the GT patch size to [ 256 , 256 ] during training, while BPOSR uses [ 1024 , 2048 ] , with an LR patch size defined as [ GT patch size / scale ]. Most of the experimental data results in the table are derived from BPOSR [24]. Methods with a gray background indicate those based on the VSSM or a frequency-domain representation.
Table 2. Quantitative comparisons (WS-PSNR/WS-SSIM) with SISR and ODISR algorithms on ODI-SR and SUN360 datasets, where the best results are highlighted in red. Note that our method follows OSRT by fixing the GT patch size to [ 256 , 256 ] during training, while BPOSR uses [ 1024 , 2048 ] , with an LR patch size defined as [ GT patch size / scale ]. Most of the experimental data results in the table are derived from BPOSR [24]. Methods with a gray background indicate those based on the VSSM or a frequency-domain representation.
DatasetODI-SRSUN360
Scale × 4 × 8 × 16 × 4 × 8 × 16
MethodWS-WS-WS-WS-WS-WS-WS-WS-WS-WS-WS-WS-
PSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIMPSNRSSIM
SISRBicubic24.620.655519.640.590817.120.433224.610.645919.720.540317.560.4638
SRCNN25.020.690420.080.611218.080.450126.300.701219.460.570117.950.4684
VDSR25.920.700921.190.633419.220.590326.360.705721.600.609118.910.5935
LapSRN25.870.694520.720.621418.450.516126.310.700020.050.599818.460.5068
MemNet25.390.696721.730.628420.030.601525.690.699921.080.601519.880.5759
MSRN25.510.700323.340.649621.730.611525.910.705123.190.647721.180.5996
EDSR25.690.695423.970.648322.240.609026.180.701223.790.647221.830.5974
D-DBPN25.500.693224.150.657322.430.605925.920.698723.700.642121.980.5958
RCAN26.230.699524.260.655422.490.617626.610.706523.880.654221.860.5938
DRN26.240.699624.320.657122.520.621226.650.707924.250.660222.110.6092
HAT26.520.749424.420.675922.610.628426.930.785424.260.706322.020.6395
MambaIR26.910.759524.460.673722.590.626327.580.799724.320.699822.060.6404
ODISR360-SS25.980.697321.650.641719.650.543126.380.701521.480.635219.620.5308
LAU-Net26.340.705224.360.660222.520.628426.480.706224.240.670822.050.6058
SphereSR24.370.677722.510.637024.170.682021.950.6342
OSRT26.890.758124.530.678022.690.626127.470.798524.380.707222.130.6388
BPOSR26.950.759824.620.677022.720.628527.590.799724.470.706222.160.6433
FATO26.780.758924.540.678422.730.631427.590.803524.420.712022.180.6449
MambaOSR27.010.761624.620.679222.660.629327.750.804224.490.711922.120.6452
Table 3. Quantitative comparison with state-of-the-art methods on ODI-SR and SUN360 datasets. PSNR/SSIM ↑: the higher, the better; LPIPS ↓: the lower, the better. LPIPS scores can better reflect texture quality; the best and second-best performances are marked in red and blue, respectively.
Table 3. Quantitative comparison with state-of-the-art methods on ODI-SR and SUN360 datasets. PSNR/SSIM ↑: the higher, the better; LPIPS ↓: the lower, the better. LPIPS scores can better reflect texture quality; the best and second-best performances are marked in red and blue, respectively.
MethodScaleODISRSUN360
PSNR↑SSIM↑LPIPS↓PSNR↑SSIM↑LPIPS↓
360-SS [19] × 4 25.5450.72510.387125.4830.71230.4113
BPOSR [24] × 4 27.7740.78120.306428.2890.79660.2754
Ours × 4 27.8750.78390.296928.5120.80210.2561
LAU-Net [1] × 8 25.1360.69530.499024.9570.69670.4949
360-SS [19] × 8 22.7620.65640.554122.4520.63660.6061
BPOSR [24] × 8 25.4530.70780.463125.3390.71330.4596
Ours × 8 25.4500.70970.444225.3590.71830.4245
Table 4. Performance comparison of different methods on mutual information (MI↑, higher is better). The best results are highlighted in red.
Table 4. Performance comparison of different methods on mutual information (MI↑, higher is better). The best results are highlighted in red.
MethodMI↑
MambaOSRBPOSR [24]LAU-Net [1]360-SS [19]
ODI-SR2.64192.60192.48842.0607
SUN3602.55682.50682.39521.8534
Table 5. Ablation studies on MambaOSR components. All models are trained on the × 8 SR task under ERP downsampling and evaluated on ODI-SR and SUN360 datasets. The best results are marked in bold.
Table 5. Ablation studies on MambaOSR components. All models are trained on the × 8 SR task under ERP downsampling and evaluated on ODI-SR and SUN360 datasets. The best results are marked in bold.
ModelFAMDGMMFFMODI-SRSUN360Params.
PSNRSSIMPSNRSSIM(M)
Baseline×××23.950.665023.820.67922.57
Model-1×24.330.663924.150.68592.59
Model-2×24.170.650823.990.67662.62
Model-3×24.390.663724.190.68602.60
MambaOSR24.530.674724.380.70232.60
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wen, W.; Zhao, Q.; Shao, X. MambaOSR: Leveraging Spatial-Frequency Mamba for Distortion-Guided Omnidirectional Image Super-Resolution. Entropy 2025, 27, 446. https://doi.org/10.3390/e27040446

AMA Style

Wen W, Zhao Q, Shao X. MambaOSR: Leveraging Spatial-Frequency Mamba for Distortion-Guided Omnidirectional Image Super-Resolution. Entropy. 2025; 27(4):446. https://doi.org/10.3390/e27040446

Chicago/Turabian Style

Wen, Weilei, Qianqian Zhao, and Xiuli Shao. 2025. "MambaOSR: Leveraging Spatial-Frequency Mamba for Distortion-Guided Omnidirectional Image Super-Resolution" Entropy 27, no. 4: 446. https://doi.org/10.3390/e27040446

APA Style

Wen, W., Zhao, Q., & Shao, X. (2025). MambaOSR: Leveraging Spatial-Frequency Mamba for Distortion-Guided Omnidirectional Image Super-Resolution. Entropy, 27(4), 446. https://doi.org/10.3390/e27040446

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop