Next Article in Journal
An Enhanced, Real-Time, Low-Cost GNSS/INS Integrated Navigation Algorithm and Its Platform Design
Previous Article in Journal
UV-Vis-NIR Broadband Dual-Mode Photodetector Based on Graphene/InP Van Der Waals Heterostructure
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Innovative Small-Target Detection Approach Against Information Attenuation: Fusing Enhanced Programmable Gradient Information and a Novel Mamba Module

1
School of Information Engineering, Inner Mongolia University of Technology, Hohhot 010051, China
2
Department of Mechanical and Electrical Engineering, Inner Mongolia Vocational College of Chemical Engineering, Hohhot 010080, China
*
Author to whom correspondence should be addressed.
Sensors 2025, 25(7), 2117; https://doi.org/10.3390/s25072117
Submission received: 17 January 2025 / Revised: 16 March 2025 / Accepted: 24 March 2025 / Published: 27 March 2025
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Serious information loss often occurs when the input data undergoes the layer-by-layer feature extraction process through deep neural networks. In small-target detection tasks, this problem becomes more serious. The existing theory holds that the information bottleneck leads to the decrease of an algorithm’s recognition rate. In order to improve the precision of small-target recognition, PGI-ViMamba is proposed in this paper to resist the information attenuation of a neural network model. The model mainly uses Multi-Level Attention-Gated Programmable Gradient Information (MLAG PGI) and SPD-Conv-VSS module. Specifically, the backbone uses an improved VSS module as a feature extractor and uses an auxiliary branch similar to YOLOv9. This neural network structure ensures that the neural network retains the feature information of the small target in the process of forward propagation while reducing the model parameters. Another advantage is that the MLAG PGI acts as a reversible branch, providing reliable gradient information. Experiments show the effectiveness of the algorithm. Compared with state-of-the-art (SOTA) models, the proposed method achieves improvements of 1.1% and 2.1% in small-target recognition precision on the VisDrone and DOTA-v1.5 datasets, respectively, with no significant decline in recall rates. Additionally, ablation experiments validate the effectiveness of the MLAG PGI and SPD-Conv-VSS modules.

1. Introduction

Small-target recognition represents a critical challenge within the realm of computer vision, with the objective of identifying objects that are located at a significant distance from the observer within a vast scene. The inherent characteristics of this task, such as the subtlety of the features of the objects to be identified and the presence of an information bottleneck in existing models, contribute to the increased difficulty of detecting small targets. In domains such as aerial target detection, unmanned aerial vehicle (UAV) applications, and industrial inspection, there is an urgent need to enhance the neural network’s capacity for recognizing small targets. In the field of deep learning, the detection of small objects presents a formidable challenge due to the limited pixel representation of these objects within an image, which can impede the model’s ability to effectively capture the requisite features for precise identification. Nevertheless, researchers have devised a variety of strategies to surmount this obstacle. Among the commonly employed techniques are the following:
1.
Image Pyramids and Feature Pyramids [1,2].
An image pyramid generates a series of images of different resolutions by downsampling or upsampling the original image. The images are arranged in order of resolution from highest to lowest, forming a pyramid-like structure. A feature pyramid generates a series of feature maps of different scales through multi-scale processing of images or feature maps. These feature maps are arranged on a scale from large to small, forming a pyramidal structure. The above two techniques are widely used in small-target recognition. Both capture different information about an image or feature through multi-scale representation. An image pyramid realizes multi-scale representation at the pixel level, which is conducive to preserving the fine-grained information of the image. A feature pyramid realizes multi-scale representation at the feature level and is good at extracting the semantic information of an image.
2.
Oversampling [3,4,5].
Oversampling is a technique that increases the proportions of small objects in the image by repeatedly sampling regions of the image containing small objects. This can enhance the representation of small objects in the feature map, thereby improving detection accuracy.
3.
Anchor boxes [6,7,8].
ROI Pooling plays a crucial role in small object detection by converting regions of interest (ROIs) of varying sizes into fixed-size feature maps, enabling subsequent classification or regression tasks. This method was first introduced in Faster R-CNN [8] and has since been widely applied in the field of object detection. The core function of ROI Pooling in small object detection is to map small-sized ROIs to a fixed-size output, ensuring that regardless of the original size of the ROI, a uniform-sized feature representation is obtained. This allows subsequent fully connected layers to accept these features and perform effective classification or regression operations.
Recently, the introduction of Vision Mamba (ViM) [9] and Programmable Gradient Information (PGI) [10] has provided new possibilities for improving the accuracy of small object detection. These two technologies have driven the development of object detection algorithms on different levels.
The main motive of this paper is to reduce the influence of the information bottleneck effect on feature extractors and improve the precision of recognition of small targets. Our main task is to propose the PGI-ViMamba model, which is an innovative fusion of the visual state space (VSS) block [11] and PGI techniques for object detection. The PGI-ViMamba model aims to enhance the accuracy and efficiency of detecting small objects within images.
The main contributions of this paper are as follows:
1. The present work achieves the integration of Vision Mamba and PGI, effectively leveraging their respective strengths in the task of small-target recognition. This results in a comprehensive solution for detecting small objects in images.
2. We have developed the SPD-Conv-VSS module, which enhances the feature extraction capability of small-scale targets by modifying the VSS module of the ViM model. This enhancement facilitates the network in capturing comprehensive and intricate feature information, thereby playing a crucial role in detecting small objects.
3. We specially designed the Multi-Level Attention-Gated PGI to provide reliable gradient information for the backbone. The improvement of this structure effectively solves the problem of information bottlenecking in deep neural networks and ensures a more accurate correlation between the detected object and the input during training. This advance is particularly valuable for algorithms that detect tiny targets.

2. Related Work

2.1. Mamba and Vision Mamba

Mamba [11] is an advanced state space model (SSM) [12] designed for efficient processing of complex data-intensive sequences, and its initial successful application was in the field of NLP. One of the main advantages of Mamba is its ability to solve the computational challenges associated with handling long sequences of conventional transformers. The selective approach of Mamba significantly speeds up inference, has a throughput five times higher than a standard transformer, and shows linear complexity with sequence length in terms of computation. Notably, even when the sequence extends to one million elements, the performance of Mamba continues to improve with increasing actual data.
Inspired by Mamba and ViT [13], ViM is a deep learning model specifically designed for computer vision tasks to improve the efficiency and effectiveness of visual representation learning. It aims to improve computational efficiency while maintaining a global reception field. The innovation of the ViM model lies in its ability to efficiently convert two-dimensional images into a sequence form suitable for processing and capture spatial information through bidirectional sequence modeling. This model addresses the issue of direction sensitivity by introducing the 2D selective scanning module (SS2D) [9], which converts images into ordered patch sequences, achieving linear complexity without sacrificing the global reception field. The ability of ViM to capture rich and detailed feature information is particularly important for small object detection. Specifically, the ViM model first divides a two-dimensional image into a series of small patches, then flattens these patches into a one-dimensional sequence and performs linear projection to obtain a fixed-size vector representation. During this process, ViM also adds positional embedding information to preserve the spatial layout information of the image. The core component of the ViM model is the VSS block, which is specifically designed for vision tasks and integrates the ability of bidirectional sequence modeling. Compared to the original Mamba model, the ViM block is more suitable for processing vision tasks that require spatial awareness. The operation algorithm of the VSS block takes into account the spatial dimension of the image, making it more efficient and accurate in processing visual data. In terms of architectural details, the ViM model contains multiple hyperparameters, such as the number of blocks, which can be adjusted according to specific vision tasks to achieve optimal performance. Previous studies have shown that ViM models provide powerful tools for computer vision tasks through their efficient visual representation learning and processing capabilities. At present, it has been successfully applied in many visual tasks.

2.2. YOLOv9 and Programmable Gradient Information (PGI)

YOLOv9 [10] is an important real-time object detection model following the YOLO series [14,15,16,17,18,19] and represents the latest progress of the YOLO series. YOLOv9 incorporates several innovations, including the concept of Programmable Gradient Information (PGI). PGI consists of three parts: the backbone branch, auxiliary reversible branch, and multi-level auxiliary information. The main branch is usually a PANET [20]-like structure. The core of PGI is to reduce losses in the information layer-by-layer modeling by introducing auxiliary gradient branches to calculate losses and gradients. The auxiliary gradient branches only participate in the calculation of losses in training. The auxiliary reversible branch provides the supervision mechanism and gradient information to ensure the correctness and effectiveness of the main branch information modeling. Multi-level auxiliary information enables all three layers of features to receive all of the information of large and small targets, avoiding the loss of target information in the deep feature pyramid.
By providing complete input information for the target task to compute the objective function, PGI can generate more reliable gradient information to update network weights, thereby helping the model learn better.

3. Method and Principle

3.1. Architecture Overview

The proposed PGI-ViMamba is shown in Figure 1. The design idea of the neural network is to avoid using the bottleneck structure that loses the feature information of small targets and to use SPD-Conv-VSS as the main feature extractor. PGI-ViMamba has two branches, comprising a main branch and an auxiliary branch. The auxiliary branch is mainly composed of the SPD-Conv-VSS module with four auxiliary detection heads. The main branch is mainly composed of the SPD-Conv-VSS block and a patch merging block, and it has a main detection head. Meanwhile, a multi-level gated PGI is specially designed between the main branch and the auxiliary branch, whose main function is to achieve the reverse propagation of gradient information.
In the training process, the main network and the auxiliary branch are trained together, and the auxiliary branch provides reliable gradient information for the main network. In the inference stage, the auxiliary branches is retained, and the inference process is jointly completed by the main network and the auxiliary branches.

3.2. SPD-Conv-VSS Module

In visual tasks, VSS blocks are often embedded in networks. As illustrated in Figure 2, to enhance the network’s capability to recognize small targets, we have improved the VSS block. It consists of two main branches. The first branch is mainly composed of the SPD-Conv [21] module, and the second branch is composed of the VSS module.
As a universal and unified module, SPD-Conv can improve the learning efficiency of feature representation without losing too much fine-grained information. The basic principle of SPD-Conv is to convert the spatial information of images into depth information, so that convolutional neural networks (CNNS) can learn image features more effectively. Specifically, the core of SPD-Conv is the Space-to-Depth operation (SDP), which compacts the original spatial dimension while remapping the values of multiple pixels to the depth dimension (i.e., the channel dimension). For example, the dimension of the feature graph is H × W × C, which is divided into several smaller regions. The spatial dimension of each region is reduced to H/2 × W/2, and the number of channels is increased to 4 × C. The principle of SPD-Conv is as shown in the following formula.
f i = X [ 0 : W : s c a l e , 0 : H : s c a l e , : ] , i = 0 , , s c a l e
o u t p u t = s t a c k f 0 , f 1 , , f s c a l e
where X represents the feature map and stack () indicates stack operations along the channel dimension. The term s c a l e indicates the interval at which data is fetched from the feature map. The specific process is to split the feature map by pixel intervals s c a l e and stack the results of these splits by channel direction. With this approach, SPD-Conv is able to retain richer information in the feature extraction stage, thereby improving the model’s recognition performance for small objects and low-resolution images.
The second branch consists of the VSS block, which mainly includes SS2D. The SS2D module consists of three main parts: the scan expanding operation, the Selective Scan Space State Sequential (S6) block operation, and the scan merging operation.
As shown on the left side of Figure 2, the scan expanding module in the diagram represents the scan expanding operation. This operation flattens the input 2D feature map into a 1D vector in four distinct directions: top-left to bottom-right, bottom-left to top-right, bottom-right to top-left, and top-right to bottom-left. The operation ensures that each element in the feature map integrates information from all other positions across different directions, thereby forming a global reception field. The output of the scan expanding block consists of image patch sequences scanned along the four directions, and these four sequences are fed into four separate S6 blocks as their respective inputs. The Scan merging module in Figure 2 is responsible for reshaping and merging the four sequences processed by S6, then reconstructing them into a two-dimensional feature map as the output. Now, let us describe the principle of S6 in detail.
The S6 block in the figure is that the four one-dimensional vector sequences obtained from the previous step are independently fed into the S6 block for feature extraction. The S6 block effectively reduces quadratic complexity to linearity by selectively scanning the spatial state sequence model, interacting each element with information that has been scanned before. This step ensures that information in all directions is thoroughly scanned to capture different features.
State Space Models (SSMs) [22,23] are typically used to describe time-invariant systems. The SS2D module models image processing tasks as a discretized linear time-invariant (LTI) system operating on patch sequences. A continuous linear time-invariant system can be represented by a linear ordinary differential equation (ODE):
h t = A h t + B u ( t )
y t = C h t + D u ( t )
where u ( t ) R is the map of the input signal, y ( t ) R is the output response,   h ( t ) is the hidden state, and h t represents the derivative of h ( t ) . A R N × N , B R N × 1 , C R 1 × N , D R 1 are the weighting parameters.
For deep learning-based image processing tasks, the required state transitions are discrete rather than continuous. We therefore introduce the state discretization methodology. Let the image patch sequence received by the S6 block from the scan expanding module be denoted as x . The t -th image patch can be represented as x t . The method of discretizing the state space model using a deep neural network is as follows:
B = L i n e a r x , C = L i n e a r x
  = L i n e a r ( L i n e a r ( x ) )
The above formulas indicate that the image patch sequence x is projected into the parameters B R N × 1 , C R 1 × N , and R N × N   through some linear neural network L i n e a r ( ) . Since the linear neural network incorporates learnable parameters, the matrices B, C, and in the equations above can be regarded as learnable matrices. The following equations describe the computational process for discretizing the state space model.
A ¯ = e   A
B ¯ = ( A ¯ I ) A 1 B   B
C ¯ = C , D ¯ = D
The equations above achieve the discretization of the matrices A, B, C, and D. The specific calculation process of the S6 module can be expressed by the following formulas:
h t = A ¯ h t 1 + B ¯ x t
y t = C ¯ h t + D ¯ x t
y = [ y 1 , y 2 , , y n ]
Equation (10) represents the discretized state transition equation, corresponding to the continuous state transition equation in Equation (3), where h t 1 denotes the hidden state at step t − 1-th and h t represents the current hidden state. Here, x t indicates the t-th image patch in the sequence x. Equation (11) corresponds to the continuous output equation in Equation (4), where y t is the model’s prediction for the t-th image patch. Finally, y in Equation (12) denotes the aggregated output formed by concatenating all of the predictions to integrate information across all of the image patches.
The scan merging module in the figure indicates that the scan merging operation will add and merge the sequences from all four directions, restoring the output image to the same size as the input. This step completes the integration of the global information and outputs the 2D feature map containing rich feature information.
The two branches are fused by a concat operation and dimensionally transformed by a 1 × 1 convolution at the output. The SS2D module exerts the most significant influence on the parameters in the VSS block. It serves as the core technology in ViM for achieving linear scanning of image blocks and extracting image features. The SPD-Conv-VSS module integrates SS2D and SPD-Conv, serving as a feature extraction technique highly sensitive to small targets.

3.3. Multi-Level Attention-Gated PGI (MLAG PGI)

As shown in Figure 3, we compare the PGI with the proposed MLAG PGI. The PGI proposed by YOLOv9 mainly consists of three parts: the main branch, the auxiliary reversible branch, and the multi-level auxiliary information. The main branch is the main part of neural network reasoning. The secondary reversible branch is used to generate reliable gradient information and provide a reverse transmission path for the primary branch. A multi-level auxiliary information control backbone is used to learn programmable multi-level semantic information. The MLAG PGI proposed in this paper makes three improvements on the basis of PGI. The first improvement is to reduce the number of detection heads on the main branch. The second improvement is to increase the number of detection heads on the auxiliary reversible branch. An important improvement is the addition of a multi-level attention-gated mechanism in the main branch and the auxiliary reversible branch, the detailed design of which can be seen in Figure 4.
Inspired by PGI, we designed an auxiliary network in parallel with the backbone network. The auxiliary network is connected to the backbone network through a backward step-down connection, which provides reliable gradient information for the main backbone network. The original PGI was modified to remove multiple levels of auxiliary information. The multi-scale feature fusion is removed from the backbone network. We believe that this improvement is conducive to improving the accuracy of small-target recognition. The principle of Multi-Level-Attention-Gated Programmable Gradient Information (MLAG PGI) can be seen in the following formulas.
V i = W V F i
K i = W K F i
Q i = W Q F i
F i = A G L a y e r F i 1 , F i = W τ · σ s o f t m a x Q i K i T d V i + F i 1 , i 1
F i   = M L A G x , F 1 , , F i 1     = A G   L a y e r ( F 0 , F 1 ) , i = 1 A G   L a y e r F i 1 , F i = A G   L a y e r A G   L a y e r F i 2 , F i 1 , F i , i 2
Formulas (13)–(15) are the calculation formula of V i , K i , and Q i in the attention mechanism. Formula (16) is the calculation formula of the AG Layer; its inputs are F i 1 and F i . F i 1 is the output of the CBLinear module or the output of the superior AG Layer, where σ ( ) represents the sigmoid() function and d represents the feature dimension. F i is the output of the MLAG of the previous level.   F i indicates the feature map of the MLAG PGI output of the previous level. It should be noted that the calculation of F i uses recursive calls, as seen by Equation (17).   W V ,   W K ,   W Q ,   W τ represent four learnable matrixes.

4. Experiments

4.1. Datasets

The VisDrone [24] dataset, which is a small-target dataset of aerial drone scenes, was released in 2018 and further expanded in 2019 by Tianjin University and other institutions. 10,209 images were provided for the small-target detection task, of which 6471 images were used for training, 343,204 were for marked boxes, 548 were for validation, and 3190 were for testing. The image data are characterized by large image resolutions, many small targets, and an average image size of 1002 × 1520.
Figure 5 reveals a notable class imbalance in the dataset, which may bias the model’s generalization capability. Sampling analysis indicates that the dataset encapsulates challenging real-world scenarios prevalent in small-target recognition, including severely occluded objects (e.g., pedestrians obscured by vehicles), low-light targets (e.g., unilluminated nighttime pedestrians and vehicles), and high-density overlapping instances (e.g., crowded urban areas). These conditions closely mirror practical deployment challenges yet remain underexplored in current literature.
The DOTA-v1.5 [25] is a high-quality image data set for ground target detection acquired by satellite. The dataset contains 403,318 target instances in 16 categories. Each image ranges in size from 800 × 800 pixels to 20,000 × 20,000 pixels and contains objects that exhibit a vast variety of scales, orientations, and shapes. It is worth mentioning that this data set is richer in the collection and labeling of small-size objects, and even 10-pixel objects are meticulously labeled. As shown in Figure 6, this is a comparison of the number of targets in different categories in this dataset.
By comparing the different characteristics of the two data sets used in this experiment, we can find that the VisDrone data set is mainly a city image in a complex light environment, in which the objects to be identified are greatly affected by light and seriously blocked. The DOTA-v1.5 dataset is characterized by a large gap between the size of the image and the size of the recognition target, a relatively simple light condition and a complex image background. By summarizing the characteristics of the above two data sets, we can find that all the situations in the target recognition task are included, so the experimental results are of practical significance.

4.2. Experimental Details

In the experiments, we used Ubuntu 22.04 as the operating system and Python 3.10.13, PyTorch 2.1.1, and Cuda 11.8 as the software environment. The experiment used an NVIDIA GeForce RTX 3090 24 G graphics card (The graphics card was manufactured by ASUSTeK Computer Inc. (ASUS) (Taipei, Taiwan) and procured within China through the JD.com (Jingdong) e-commerce platform) as hardware. The implementation code of the neural network was modified based on Ultralytics version 8.0.105. The hyperparameters used in the training, testing, and validation of the project were consistent. The training epochs were set to 800, and the image input into the network was rescaled to 640 × 640. All experiments used the same data augmentation operations to more fairly determine the performance of the model, including horizontal and vertical flips and random rotation operations.

4.3. Comparison Results

4.3.1. Comparison with State-of-the-Art Methods

Table 1 presents the results of comparative experiments based on the VisDrone dataset and the DOTA-v1.5 dataset.
Through observation, we find that the total parameter count of PGI-ViMamba (27.8 M) is moderately low. This is because our model removes the SPPF and PANET-like structures commonly used in the YOLO series and replaces them with an additional auxiliary detection head and the MLAG PGI structure. In terms of the computational metric FLOPs, the computational cost of our proposed method (209.0 G) is at a moderately high level compared to other models. The increase in computational cost is primarily due to the inclusion of the SS2D module in the PGI-ViMamba network, as its four-directional scanning mechanism contributes to additional computations. Additionally, the use of attention mechanisms in our approach further increases the computational load. In terms of the Frames Per Second (FPS) metric, there is not much difference between PGI-ViMamba and the lightweight YOLOv8s. PGI-ViMamba reaches 153 FPS, a decrease of only 35 FPS, and it is at a moderate level. This is an acceptable result.
By observing the P, R, and mAP50 metrics, it can be seen that the results of any method on the DOTA-v1.5 dataset are higher than those on the VisDrone dataset. The reason for this phenomenon lies in the fact that although DOTA-v1.5 has more detection categories and a larger variation in target sizes, its lighting conditions are relatively uniform, and the detected objects are not occluded. Precision (P) measures the proportion of correct predictions made by the model. A higher precision indicates fewer false positives (FP) in the model’s predictions. Specifically, on different datasets, the precision of PGI-ViMamba is 1.1% and 2.1% higher than that of the second-best model, YOLOv11. Recall (R) measures the model’s ability to detect all targets. A higher recall indicates fewer missed detections by the model, corresponding to a reduction in false negatives (FN). On the VisDrone dataset, the recall of PGI-ViMamba is 1.2% lower than that of the best-performing model, YOLOv11x, while on the DOTA-v1.5 dataset, it is 0.4% lower than that of YOLOv11x.
Based on the analysis of the above results, it can be observed that PGI-ViMamba excels at detecting targets with prominent features, which is why its precision is higher than other methods. However, when the target features are less apparent or the noise is excessive, the recall rate of PGI-ViMamba tends to be lower compared to other models. This comprehensive analysis suggests that PGI-ViMamba adopts a cautious approach in target detection, prioritizing the avoidance of false positives over the risk of missing detections. In other words, it tends to be more conservative, preferring to miss some targets rather than incorrectly detect non-targets.
The mAP50 is the mean average precision calculated at an IoU (Intersection over Union) threshold of 0.5. It comprehensively evaluates the detection performance of the model across different categories, with an IoU threshold of 0.5 indicating that the overlap between the predicted bounding box and the ground truth box is at least 50%. A higher mAP50 signifies better overall detection performance by the model. On both datasets, PGI-ViMamba outperforms the second-best model by 1.0% and 0.5%, respectively.
To provide a more intuitive comparison of model performance, we present Figure 7 based on experimental results from the VisDrone dataset. In the figure, we compare the accuracy, recall, and number of learnable parameters across different models. As observed from the figure, PGI-ViMamba achieves higher accuracy with a moderate number of parameters while maintaining recall rates comparable to other models.

4.3.2. Comparison of Precision and Recall Rate of Different Models

As shown in Figure 8, we collected the training process data of nine models on the VisDrone dataset for comparison, which is conducive to evaluating the model capability of PGI-ViMamba. Using the experimental data, we drew the accuracy and recall curves of different models in 800 training sessions. By observing the change of the curve, we can find that the accuracy and recall rate of different models tended to be stable when the training rounds were between 700 and 800 epochs.
First, analyzing accuracy dynamics, we observe that PGI-ViMamba’s accuracy increases rapidly from epochs 100 to 300, demonstrating a significantly faster improvement rate compared to other models. This indicates PGI-ViMamba’s superior ability to learn the features of small-scale targets early in training. Second, while PGI-ViMamba’s recall rate exhibits a slower ascent than YOLOv10x and YOLOv11x, its final recall performance remains comparable to these models.
From these analyses, PGI-ViMamba emerges as a high-precision, prudent small-target detection algorithm. Given the inherent information decay in extremely small target images, these characteristics align with the demands of small-target detection tasks, enhancing the reliability of deep neural networks in practical applications.

4.3.3. Comparison of Precision and Recall of Different Objects

To further compare the performance of different models, we used experimental results from the more difficult VisDrone dataset to make Table 2 and the radar diagram shown in Figure 9. In the table we list the accuracy and recall rates of different types of targets, as well as the average.
In Figure 9, the left panel displays radar charts illustrating the accuracy rates for ten distinct target recognition categories, while the right panel shows radar charts for their corresponding recall rates.
Observations reveal that PGI-ViMamba outperforms other models in accuracy across all categories, as evidenced by its radar curve fully encompassing those of competing models.
The experimental results of fine-grained feature categories with rich features and high recognition difficulty, such as motorcycles, pedestrians, vans, tricycles, and awning tricycles, demonstrate the significant advantages of using the PGI-ViMamba model. Its performance exceeds that of the second-best model by 0.8%, 1.3%, 1.2%, 2.0%, and 0.8% respectively. For occlusion-prone categories like People and Bicycle, PGI-ViMamba still achieves accuracy improvements of 0.2% and 0.6%. Coarse-grained feature categories with abundant training samples and prominent features (e.g., car, truck, bus) exhibit comparable performance between PGI-ViMamba and other models, with marginal gains of 0.1%, 0.5%, and 0.1% over the second-best model.
In recall comparisons, PGI-ViMamba does not dominate most categories but demonstrates superiority in Car, Van, and Truck—categories with dominant coarse-grained features—outperforming the second-best model by 1.1%, 0.1%, and 0.4%, respectively.
Collectively, these results indicate that PGI-ViMamba prioritizes minimizing false positives over maximizing recall when category features are ambiguous or corrupted, aligning with its design objective of ensuring reliable detection under challenging conditions.

5. Ablation Study

In this paper, we conducted two ablation experiments, one of which was to reduce the depth of MLAG PGI to assess the impact on model accuracy. On the other hand, a classical feature extraction network was used to replace our proposed SPD-Conv-VSS module to evaluate the influence of different backbone on the performance of our proposed model.

5.1. Comparison of Different Levels of MLAG PGI

The first aspect of the ablation experiment is to reduce the reversible depth of MLAG PGI. The network structure under different reversible depths is shown in Figure 3. From Table 3, we observe that gradually reducing the reversible depth of MLAG PGI will lead to changes in the accuracy and recall rates. It was found that the accuracy and recall rate decreased with the decrease of the reversible depth. It can be seen from the experimental results that the larger MLAG PGI reversible depth is beneficial to the feature extraction of small targets.
The second aspect of the ablation experiment was to verify whether a larger MLAG PGI reversible depth is more conducive to target recognition. When the reversible depth of MLAG PGI was increased to five levels, both the accuracy and recall rate were found to decrease. This shows that a greater reversibility depth of MLAG PGI is not better, so it is necessary to design a suitable MLAG PGI reversibility depth using a neural network model. The reason for this phenomenon is that a too-large MLAG PGI reversible depth may cause the loss of feedback gradient information.

5.2. Comparison of Different Information Extraction Backbone Networks

We conducted ablation experiments from two perspectives, with the results summarized in Table 4. First, by maintaining the feature extraction module unchanged, we compared the model’s performance with and without the MLAG PGI module. Significant degradation across all of the evaluation metrics was observed. For instance, when PGI-ViMamba employed the SPD-Conv-VSS feature extraction module without MLAG PGI, Precision (P) decreased by 28.9%, Recall (R) dropped by 17.8%, mAP50 declined by 16.4%, and mAP50:95 fell by 13.7%. This severe performance deterioration stems from the dual functionality of MLAG PGI: it acts not only as a gradient propagation pathway to stabilize backward signal flow but also as a multi-scale feature fusion channel to enhance hierarchical representation learning, both of which are critical to the model’s effectiveness.
Second, we evaluated the impact of the SPD-Conv-VSS module. When SPD-Conv-VSS was replaced in models equipped with MLAG PGI, the performance suffered at least a 4.7% reduction in P, 10.3% in R, 4% in mAP50, and 2.5% in mAP50:95. Even without MLAG PGI, replacing SPD-Conv-VSS still led to declines of 3.8% in P, 11.2% in R, 1% in mAP50, and 2% in mAP50:95. These comparative results unequivocally demonstrate that the SPD-Conv-VSS module significantly contributes to improving detection metrics across diverse experimental configurations.
By systematically analyzing these ablation outcomes, we conclude that both MLA G PGI and SPD-Conv-VSS are indispensable components of the proposed architecture, each addressing distinct challenges in small-target detection while synergistically enhancing overall robustness and precision.

6. Conclusions

This study focuses on the information bottleneck problem in deep neural networks. The information bottleneck problem refers to the gradual loss of input information during data propagation through multiple network layers in the training process, which prevents models from fully utilizing effective information from raw data. This phenomenon significantly degrades the performance of the model, especially in small-target recognition tasks, where the impact is more pronounced.
To mitigate information loss caused by the information bottleneck, we designed a novel network architecture termed PGI-ViMamba. The core design philosophy of this model deviates from the conventional “information extraction → information fusion → information spatial mapping” pipeline, instead adopting a parallel multi-branch architecture, which includes two forward information extraction branches and one reverse branch. The two forward branches consist of a primary branch and an auxiliary branch. The primary branch is equipped with a single detection head, while the auxiliary branch incorporates multiple detection heads. Both branches utilize our proposed SPD-Conv-VSS module as the core information extraction component. This module integrates the SS2D module and SPD-Conv through parallel substructures, and its effectiveness has been validated via ablation experiments. Another critical innovation is the introduction of a reverse branch—a specially designed attention-gated pathway called MLAG PGI (Multi-Level Attention-Gated Progressive Gradient Integration). This reversible branch not only generates reliable gradient information but also enables multi-scale feature fusion, ensuring that deep-layer features retain critical image information. To validate MLAG PGI’s effectiveness, two ablation studies were conducted. First, hierarchical impact analysis comparing MLAG PGI at different levels revealed that the 4-level configuration delivers optimal small-target recognition performance. Second, removing MLAG PGI caused significant performance degradation across all tested feature extractors.
The proposed PGI-ViMamba, with 27.8M parameters and 209.0 GFLOPs, achieves 1.1% and 2.1% precision improvements on the VisDrone and DOTA-v1.5 datasets, respectively, while maintaining recall rates comparable to other models. Comparative detection analyses further indicate its reduced tendency for false positives. Our work demonstrates that the Mamba model, through its unique feature extraction mechanism, and the enhanced PGI framework, when synergistically integrated, significantly enhance small-target recognition performance while effectively mitigating the inherent information degradation challenges in deep neural networks.

Author Contributions

Supervision, Y.J. and Y.L.; conceptualization, Y.J. and Y.L.; formal analysis, Y.J., Y.L. and N.W.; methodology, Y.L. and N.L.; software, Y.L. and M.L.; writing-original draft, Y.L.; funding acquisition, B.S. and Q.R. All authors have read and agreed to the published version of the manuscript.

Funding

This study is supported by the National Natural Science Foundation of China (62206138, 62466044), Inner Mongolia Natural Science Foundation (2024MS06009, 2024QN06021), Research program of science and technology at Universities of Inner Mongolia Autonomous Region (NJZZ23081), Science Research Foundation of Inner Mongolia University of Technology (BS2021079).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available due to privacy restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Lin, T.-Y.; Dollár, P.; Girshick, R.B.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  2. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In Proceedings of the European Conference on Computer Vision (ECCV), Amsterdam, The Netherlands, 11–14 October 2016; pp. 21–37. [Google Scholar]
  3. He, H.; Bai, Y.; Garcia, E.A.; Li, S. ADASYN: Adaptive Synthetic Sampling Approach for Imbalanced Learning. In Proceedings of the IEEE International Joint Conference on Neural Networks (IJCNN), Hong Kong, China, 1–8 June 2008; pp. 1322–1328. [Google Scholar]
  4. Han, H.; Wang, W.-Y.; Mao, B.-H. Borderline-SMOTE: A New Over-Sampling Method in Imbalanced Data Sets Learning. In Proceedings of the International Conference on Intelligent Computing (ICIC), Hefei, China, 23–26 August 2005; pp. 878–887. [Google Scholar]
  5. Chawla, N.V.; Japkowicz, N.; Kotcz, A. SMOTE for Learning from Imbalanced Data: Progress and Challenges. J. Artif. Intell. Res. 2004, 21, 1–16. [Google Scholar]
  6. Yang, T.; Zhang, X.; Li, Z.; Zhang, W.; Sun, J. MetaAnchor: Learning to Detect Objects with Customized Anchors. arXiv 2018, arXiv:1807.00980. [Google Scholar]
  7. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  8. Ren, S.; He, K.; Girshick, R.B.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  9. Zhu, L.; Liao, B.; Zhang, Q.; Wang, X.; Liu, W.; Wang, X. Vision Mamba: Efficient Visual Representation Learning with Bidirectional State Space Model. arXiv 2024, arXiv:2401.09417. [Google Scholar]
  10. Wang, C.; Yeh, I.; Liao, H. YOLOv9: Learning What You Want to Learn Using Programmable Gradient Information. arXiv 2024, arXiv:2402.13616. [Google Scholar]
  11. Gu, A.; Dao, T. Mamba: Linear-time sequence modeling with selective state spaces. arXiv 2023, arXiv:2312.00752. [Google Scholar]
  12. Gu, A.; Johnson, I.; Goel, K.; Saab, K.; Dao, T.; Rudra, A.; Ré, C. Combining Recurrent, Convolutional, and Continuous-time Models with Linear State-Space Layers. arXiv 2021, arXiv:2110.13985. [Google Scholar]
  13. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  14. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  16. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  17. Bochkovskiy, A.; Wang, C.; Liao, H.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  18. Wang, C.; Bochkovskiy, A.; Liao, H.M. Scaled-YOLOv4: Scaling Cross Stage Partial Network. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13024–13033. [Google Scholar]
  19. Ge, Z.; Liu, S.T.; Wang, F.; Li, Z.M.; Sun, J. Yolox: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  20. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  21. Li, Y.; Chen, X.; Wang, Z.; Liu, Y. SPD-Conv: A New Convolutional Layer for Improving CNN Performance on Low-Resolution Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 12345–12354. [Google Scholar]
  22. Gu, A.; Goel, K.; Ré, C. Efficiently Modeling Long Sequences with Structured State Spaces. arXiv 2021, arXiv:2111.00396. [Google Scholar]
  23. Liu, Y.; Tian, Y.; Zhao, Y.; Yu, H.; Xie, L.; Wang, Y.; Ye, Q.; Liu, Y. VMamba: Visual State Space Model. arXiv 2024, arXiv:2401.10166. [Google Scholar]
  24. Zhu, P.; Wen, L.; Du, D.; Bian, X.; Fan, H.; Hu, Q.; Ling, H. Detection and Tracking Meet Drones Challenge. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7380–7399. [Google Scholar] [PubMed]
  25. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A large-scale dataset for object detection in aerial images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
Figure 1. Architecture of PGI-ViMamba.
Figure 1. Architecture of PGI-ViMamba.
Sensors 25 02117 g001
Figure 2. SPD-Conv-VSS module mainly adds branches containing SPD-Conv and SS2D.
Figure 2. SPD-Conv-VSS module mainly adds branches containing SPD-Conv and SS2D.
Sensors 25 02117 g002
Figure 3. PGI vs. MLAG PGI.
Figure 3. PGI vs. MLAG PGI.
Sensors 25 02117 g003
Figure 4. Multi-Level-Attention-Gated Programmable Gradient Information (MLAG PGI).
Figure 4. Multi-Level-Attention-Gated Programmable Gradient Information (MLAG PGI).
Sensors 25 02117 g004
Figure 5. VisDrone Dataset.
Figure 5. VisDrone Dataset.
Sensors 25 02117 g005
Figure 6. DATA-v1.5 Dataset.
Figure 6. DATA-v1.5 Dataset.
Sensors 25 02117 g006
Figure 7. Comparison of the precision, recall, and parameter numbers of different models.
Figure 7. Comparison of the precision, recall, and parameter numbers of different models.
Sensors 25 02117 g007
Figure 8. Comparison of precision and recall rate of different models.
Figure 8. Comparison of precision and recall rate of different models.
Sensors 25 02117 g008
Figure 9. Comparison of precision and recall of different object.
Figure 9. Comparison of precision and recall of different object.
Sensors 25 02117 g009
Table 1. Result of comparison experiments on the VisDrone dataset and DOTA-v1.5 dataset.
Table 1. Result of comparison experiments on the VisDrone dataset and DOTA-v1.5 dataset.
ModelParam. (M)FLOPs (G)P (%)R (%)mAP50 (%)FPS
YOLOv8s11.228.641.3/60.1 *37.8/58.322.1/25.6188
YOLOv8-C2f-VSS14.935.943.0/63.836.8/57.835.9/38.9164
YOLOv9-c25.3102.141.3/61.349.7/63.441.3/42.5150
YOLOv10x29.5160.452.8/70.150.8/66.740.8/43.684
YOLOv11x56.9194.953.1/72.050.2/65.441.6/44.892
GELAN-c25.5102.845.3/63.250.0/66.238.3/41.6160
PP-YOLO-Plus54.6115.849.5/68.249.6/53.342.1/43.8150
RT DERT-X67.0234.145.6/61.836.5/50.342.5/43.6148
PGI-ViMamba27.8209.054.2/74.149.6/66.343.5/45.3153
* Result on the VisDrone dataset/Result on the DOTA-v1.5 dataset. In the table, the bold indicates the experimental results of the best-performing model, while the underline denotes the results of the second-best model.
Table 2. Comparison results of different categories of VisDrone datasets in different models.
Table 2. Comparison results of different categories of VisDrone datasets in different models.
CategoryYOLOv8YOLOv8-C2f-VSSYOLOv9-cYOLOv10xYOLOv11xGELAN-cPP-YOLO-PlusRT DERT-XPGI-ViMamba
Bicycle26.3/26.427.3/26.326.6/30.930.6/33.831.5/30.527.2/28.230.6/31.327.9/24.932.1/30.1
People23.5/25.525.2/23.223.1/33.938.9/38.539.7/38.730.1/37.135.1/32.930.8/27.839.9/36.2
Pedestrian33.9/33.934.1/28.133.1/33.943.8/43.742.4/40.436.2/36.240.2/42.336.3/30.345.1/33.1
Motor35.2/35.435.2/30.234.2/35.535.7/35.636.9/36.632.1/37.135.7/34.732.5/30.537.7/36.2
Bus62.3/52.367.3/51.361.3/79.386.8/81.287.6/77.681.1/88.186.9/86.181.9/59.287.7/83.1
Awning-tricycle31.1/31.935.1/33.131.3/40.340.9/40.840.2/39.234.1/38.139.1/37.134.1/30.141.7/35.1
Tricycle40.2/32.241.2/34.238.8/41.840.8/40.741.1/40.135.2/38.238.8/38.135.2/32.243.2/36.2
Car77.4/62.778.2/64.477.4/85.491.4/83.490.9/81.380.4/85.485.6/85.480.4/65.491.5/86.5
Van30.1/30.933.2/28.232.2/55.955.6/55.154.8/54.848.1/54.151.6/54.648.6/32.156.8/55.2
Truck53.1/46.753.1/49.152.9/60.463.5/55.165.9/62.948.5/57.551.5/53.548.5/32.566.4/64.3
Avg *41.3/37.843.0/36.841.3/49.752.8/50.853.1/50.245.3/50.049.5/49.645.6/36.554.2/49.6
Avg * represents the average of all categories of the identified target. The numbers in front of “/” are Precision, and the numbers after “/” represent Recall rates. In the table, the bold indicates the experimental results of the best-performing model, while the underline denotes the results of the second-best model.
Table 3. The results of the comparison of different levels of MLAG PGI.
Table 3. The results of the comparison of different levels of MLAG PGI.
n-Level-PGIP (%)R (%)mAP50 (%)mAP50:95 (%)
0-level-PGI 120.331.820.114.1
1-level-PGI21.937.828.923.2
2-level-PGI29.340.935.325.5
3-level-PGI45.341.738.328.6
4-level-PGI54.249.643.532.3
5-level-PGI41.541.834.120.3
1 0-level-PGI indicates that programmable gradient information is not added. In the table, the bold indicates the experimental results of the best-performing model.
Table 4. The results of the comparison of different BackBones.
Table 4. The results of the comparison of different BackBones.
ModelBackBoneMLAG 1P (%)R (%)mAP50 (%)mAP50:95 (%)
PGI-ViMambaEfficientNetX20.310.810.110.1
PGI-ViMambaEfficientNet38.539.331.919.2
PGI-ViMambaResNet-50X20.317.919.35.5
PGI-ViMambaResNet-5035.338.133.318.6
PGI-ViMambaVGG-16X21.520.626.117.3
PGI-ViMambaVGG-1649.533.939.530.5
PGI-ViMambaSPD-Conv-VSSX25.331.827.119.3
PGI-ViMambaSPD-Conv-VSS54.249.643.533.0
1 MLAG specifies whether to use MLAG PGI. MLAG PGI is 4-Level-PGI by default. In the table, the bold indicates the experimental results of the best-performing model, while the underline denotes the results of the second-best model. In the table, “X” indicates that MLAG is not used in the experiment, and “√” indicates that MLAG is used in the experiment.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, Y.; Ji, Y.; Ren, Q.; Shi, B.; Liu, N.; Lu, M.; Wu, N. An Innovative Small-Target Detection Approach Against Information Attenuation: Fusing Enhanced Programmable Gradient Information and a Novel Mamba Module. Sensors 2025, 25, 2117. https://doi.org/10.3390/s25072117

AMA Style

Liu Y, Ji Y, Ren Q, Shi B, Liu N, Lu M, Wu N. An Innovative Small-Target Detection Approach Against Information Attenuation: Fusing Enhanced Programmable Gradient Information and a Novel Mamba Module. Sensors. 2025; 25(7):2117. https://doi.org/10.3390/s25072117

Chicago/Turabian Style

Liu, Yang, Yatu Ji, Qingdaoerji Ren, Bao Shi, Na Liu, Min Lu, and Nier Wu. 2025. "An Innovative Small-Target Detection Approach Against Information Attenuation: Fusing Enhanced Programmable Gradient Information and a Novel Mamba Module" Sensors 25, no. 7: 2117. https://doi.org/10.3390/s25072117

APA Style

Liu, Y., Ji, Y., Ren, Q., Shi, B., Liu, N., Lu, M., & Wu, N. (2025). An Innovative Small-Target Detection Approach Against Information Attenuation: Fusing Enhanced Programmable Gradient Information and a Novel Mamba Module. Sensors, 25(7), 2117. https://doi.org/10.3390/s25072117

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop