Next Article in Journal
A Quantitative Comparison of 31P Magnetic Resonance Spectroscopy RF Coil Sensitivity and SNR between 7T and 10.5T Human MRI Scanners Using a Loop-Dipole 31P-1H Probe
Previous Article in Journal
A Novel Medium Access Policy Based on Reinforcement Learning in Energy-Harvesting Underwater Sensor Networks
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Spectrum Sensing Method Based on STFT-RADN in Cognitive Radio Networks

by
Anyi Wang
*,
Tao Zhu
and
Qifeng Meng
School of Communication and Information Engineering, Xi’an University of Science and Technology, Xi’an 710054, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5792; https://doi.org/10.3390/s24175792
Submission received: 7 August 2024 / Revised: 26 August 2024 / Accepted: 5 September 2024 / Published: 6 September 2024
(This article belongs to the Special Issue Sensors for Enabling Wireless Spectrum Access)

Abstract

:
To address the common issues in traditional convolutional neural network (CNN)-based spectrum sensing algorithms in cognitive radio networks (CRNs), including inadequate signal feature representation, inefficient utilization of feature map information, and limited feature extraction capabilities due to shallow network structures, this paper proposes a spectrum sensing algorithm based on a short-time Fourier transform (STFT) and residual attention dense network (RADN). Specifically, the RADN model improves the basic residual block and introduces the convolutional block attention module (CBAM), combining residual connections and dense connections to form a powerful deep feature extraction structure known as residual in dense (RID). This significantly enhances the network’s feature extraction capabilities. By performing STFT on the received signals and normalizing them, the signals are converted into time–frequency spectrograms as network inputs, better capturing signal features. The RADN is trained to extract abstract features from the time–frequency images, and the trained RADN serves as the final classifier for spectrum sensing. Experimental results demonstrate that the STFT-RADN spectrum sensing method significantly improves performance under low signal-to-noise ratio (SNR) conditions compared to traditional deep-learning-based methods. This method not only adapts to various modulation schemes but also exhibits high detection probability and strong robustness.

1. Introduction

With the widespread adoption of 5G communication and the development of 6G technology, the era of the Internet of Everything and intelligent connectivity is approaching [1]. However, the rapid increase in wireless devices and the static management of the radio spectrum have led to a shortage within the available spectrum [2]. CRN is an effective solution to this spectrum resource scarcity [3]. The concept of CRN was introduced by Mitola in 1999. CRN is an intelligent wireless communication system that allows secondary users (SUs) to utilize the licensed spectrum without interfering with primary users (PUs). Spectrum sensing is a key technology in CRN, enabling SUs to determine in real time whether a frequency band is occupied by a PU, thereby efficiently utilizing the available spectrum [4].
Traditional spectrum sensing methods encompass energy detection [5], cyclostationary feature detection [6], matched filter detection [7], and eigenvalue-based detection [8]. These methods share the common characteristic of computing detection statistics and comparing them with a predetermined threshold to make a final decision. The accuracy of the detection statistics and thresholds directly affects the detection results. Traditional methods often perform poorly under low SNR conditions.
To overcome the challenge of determining detection thresholds in traditional spectrum sensing, researchers have increasingly proposed deep-learning-based methods. Deep learning excels at classifying large datasets, and spectrum sensing can be viewed as a binary classification problem to determine the presence or absence of a PU signal in complex wireless data. Compared to traditional machine learning methods, deep-learning-based spectrum sensing provides higher accuracy and robustness in complex wireless environments and high-interference conditions, enhancing its effectiveness [9]. CNNs are effective for extracting target features and outperform other machine learning algorithms in extracting two-dimensional image features [10]. Pan et al. [11] proposed a spectrum sensing method based on cyclostationary features and a CNN, which significantly improved detection probability compared to traditional methods. However, this method still has limitations, such as not being applicable to all modulation signals, insufficient feature extraction capabilities of the CNN, and gradient vanishing issues with deeper networks. Liu [12] and Han [13] used sample covariance matrices and cyclostationary and energy features as detection statistics and network inputs. These methods are applicable to various modulation signals and enhance detection probability by preprocessing time-domain signals, but traditional CNNs struggle to effectively extract features under low SNR conditions. Zheng et al. [14] proposed a ResNet-based spectrum sensing method that enhanced detection performance over traditional CNN. Both Cai [15] and Gai [16] employed STFT for signal preprocessing, leading to significant performance improvements. However, their approaches are still hindered by the limitations of a simple CNN architecture, which is inadequate for effective feature extraction. Wang et al. [17] proposed the RDN-CBAM method, significantly improving spectrum sensing performance under low SNR conditions by employing dense connections and integrating the CBAM attention mechanism. However, it still falls short in adequately representing signal features.
To overcome these limitations, we propose a more effective spectrum sensing method. By applying the STFT for the time–frequency analysis of received signals, we transform one-dimensional signals into two-dimensional time–frequency spectrograms, capturing the signal’s characteristics more comprehensively. Thus, we convert the spectrum sensing problem into an image binary classification problem. To enhance the feature extraction capability of traditional CNN and address gradient vanishing issues in deeper networks, we improve the basic residual block and introduce CBAM, combining dense and residual connections to propose RADN. This significantly enhances the network’s feature extraction capability. Combined with STFT, this forms the STFT-RADN spectrum sensing method.

2. System Model

In CRN, spectrum sensing is essentially a signal detection problem with two possible outcomes: the presence or absence of the primary user. This is similar to binary detection; thus, spectrum sensing can be transformed into a binary classification task. The binary sensing model is described as follows:
H 0 : y ( n ) = g ( n ) H 1 : y ( n ) = s ( n ) + g ( n )
where H 0 indicates the absence of the PU, H 1 indicates the presence of the PU, n is the time series of the discrete signal, y ( n ) is the received signal, s ( n ) is the signal transmitted by the PU through a Rayleigh fading channel, and g ( n ) is the Gaussian noise with a mean of 0 and variance of σ 2 .
The performance of spectrum sensing is typically evaluated using two metrics: detection probability ( P d ) and false alarm probability ( P f ). P d and P f can be expressed as follows:
P d = P { H 1 | H 1 } P f = P { H 1 | H 0 }

3. STFT-RADN Spectrum Sensing Algorithm

This paper adopts a deep-learning-based approach to address the spectrum sensing problem, utilizing RADN combined with STFT. The proposed spectrum sensing method consists of three main parts: data processing, model training, and spectrum sensing, as illustrated in Figure 1.

3.1. Short-Time Fourier Transform

To fully utilize the features of two-dimensional CNN, one-dimensional time series signals must be converted into two-dimensional matrix forms. Most existing deep-learning-based spectrum sensing methods generate two-dimensional grayscale images by segmenting and concatenating one-dimensional received signals. Using grayscale images as input aligns with the characteristic that most neural networks are more suitable for handling high-dimensional data, but this approach lacks the frequency-domain information of the signals.
By employing the STFT in signal processing, we can convert simple time-domain signals into feature spectra containing both time-domain and frequency-domain information [18]. This method generates two-dimensional grayscale images from received one-dimensional signals through time–frequency analysis, effectively reflecting the frequency-domain characteristics and providing strong robustness and noise resistance.
To efficiently extract signal features, this study employs STFT to convert time-domain signals into time–frequency representations. This representation captures the time–frequency characteristics of the signals, facilitating the distinction between signals and noise.
STFT breaks the signal samples into chunks using a window function, and each chunk is then discrete Fourier transformed (DFT) [19]. The mathematical expression for STFT is as follows:
S T F T ( y ( n ) ) [ t , ω ] = n = y ( n ) w ( n t ) e j 2 π ω n / N
where y ( n ) is the input discrete-time signal, and N denotes the number of points in the DFT. The indices t and ω correspond to the time and frequency indices, respectively, and w n t represents the window function. Common window functions include rectangular, Hanning, Hamming, Gaussian, and Kaiser. Among these, the Kaiser window stands out due to its superior frequency characteristics, lower spectral leakage, and adjustable attenuation factor.
Through STFT conversion, the resulting time–frequency matrix can be represented as Y ( t , ω ) . The matrix is then normalized to produce grayscale time–frequency images, as illustrated in Figure 2. The resulting grayscale images can be represented as follows:
S ( t , ω ) = Y ( t , ω ) min ( Y ) max ( Y ) min ( Y )

3.2. Convolutional Block Attention Module

The CBAM, proposed by Woo S et al. [20], is a representative model of hybrid attention mechanisms. It combines channel and spatial attention modules to adaptively optimize features by applying attention mechanisms to the feature maps of CNN, thereby enhancing the representation of important features. This effectively improves the network’s performance and generalization capabilities.
CBAM consists of two submodules: the channel attention module and spatial attention module. These modules operate sequentially, enhancing feature representation by focusing on different feature channels and spatial locations, respectively. The overall structure of CBAM is shown in Figure 3, and its implementation process is described by Equations (5) and (6).
F = M c ( F ) F
F = M s ( F ) F
In the formulae, F represents the input feature map, M c denotes the channel attention weights, M s denotes the spatial attention weights, represents element-wise multiplication, F signifies the feature map after applying the channel attention weights, and F is the final output of CBAM.
The goal of the channel attention module is to enhance important features by weighting different channels. The structure of the channel attention module is shown in Figure 4. First, the input feature map F undergoes global max pooling and global average pooling to aggregate spatial information. Then, the results of both global max pooling and global average pooling are separately fed into a shared multilayer perceptron (Shared MLP) to obtain two feature maps. Finally, the outputs of the MLP are combined using element-wise addition, and the resulting feature vector is passed through a sigmoid activation function to produce the channel attention weight matrix M c .
The formula for the channel attention mechanism can be formulated as follows:
M c ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L F ( M a x P o o l ( F ) ) ) = σ ( W 1 ( W 0 ( F a v g c ) ) + W 1 ( W 0 ( F max C ) ) )  
where F a v g c and F max c represent the feature maps output by the average pooling and max pooling operations, respectively, and σ denotes the sigmoid function. W 0 and W 1 are the weights of the shared multilayer perceptron. M c ( F ) represents the output of the channel attention mechanism.
The objective of the spatial attention module is to emphasize important spatial features by focusing on different locations within the feature map. The structure of the spatial attention module is illustrated in Figure 5. First, the input feature map F undergoes global max pooling and global average pooling along the channel dimension, resulting in two separate feature maps. These feature maps are then concatenated along the channel axis. Finally, a convolution operation is applied to the concatenated result, followed by a sigmoid activation function to obtain the spatial attention weight matrix M s .
The formula for the spatial attention mechanism can be formulated as follows:
M s ( F ) = σ ( f 7 × 7 ( [ A v g P o o l ( F ) ; M a x P o o l ( F ) ] ) = σ ( f 7 × 7 F a v g s ; F max s )
where f 7 × 7 represents the convolution operation with a kernel size of 7 × 7 , and M s ( F ) represents the output of the spatial attention module.

3.3. Res-Inception Attention Block (RAB)

The residual block concept was first introduced by He et al. [21] to address gradient vanishing and exploding problems arising from increasing network depth. The core idea is to introduce shortcut connections for direct input propagation across layers, thereby improving deep network training and mitigating gradient vanishing or exploding issues. Additionally, residual blocks can effectively reduce overfitting in deep networks. The basic structure of a residual block is shown in Figure 6a.
However, the basic residual block has limitations in feature extraction capabilities. The inception module can simultaneously extract features at different scales, capturing richer image information [22]. The CBAM module can focus on important features in both channel and spatial dimensions, enhancing feature extraction. Therefore, this paper proposes the RAB, which incorporates the inception module and the CBAM module into the residual block, forming a more powerful feature extraction unit.
Furthermore, to improve feature extraction efficiency, this paper introduces batch normalization (BN), which accelerates the network’s convergence process, making the training more robust. The RAB is illustrated in Figure 6b.
The mathematical expression of the res-inception block can be formulated as follows:
F x = f ( x ) + x
where x represents the input, f ( x ) denotes the residual mapping of the feature map, and F ( x ) is the output of the RAB. This design allows the RAB to extract multiscale features more effectively and capture key features within the image more efficiently. Additionally, by leveraging the advantages of the residual structure, this design significantly enhances the model’s performance while maintaining reasonable complexity.

3.4. Dense Group (DG)

To further enhance the feature extraction capabilities of the RAB, this paper proposes a new fundamental module inspired by Zhang et al. [23]. This module forms a DG by stacking RABs and using a dense connection approach, as shown in Figure 7.
Stacking RABs allows the network to progressively extract richer and more abstract feature representations. This enables each block to learn new features and allows the network to capture more complex patterns and information.
Moreover, the dense connections link the output of each block to subsequent blocks, facilitating more efficient feature propagation and combination. This ensures maximal information flow between layers, promoting feature reuse. Additionally, adding a convolutional layer at the end of the DG enhances feature extraction, alignment, deeper feature learning, and eliminates information bottlenecks. This operation enhances the model’s performance, making it more efficient and accurate in handling complex tasks.
The mathematical expression for the DG is as follows:
Y = G ( [ X , F 1 ( X ) , F 2 ( X ) , F 3 ( X ) ] )
where X represents the input feature map, F i ( X ) denotes the output feature map of the i -th residual attention block, and [ ] represents the concatenation of multiple inputs. G ( ) indicates the mapping of the feature maps, and Y denotes the output of the DG block.

3.5. Residual Attention Dense Network

The proposed RADN consists of three parts: a shallow feature extraction module, an RID module, and an output module. First, the shallow feature extraction module captures basic features and reduces computational burden. Next, the RID module, consisting of densely connected residual blocks, enhances feature reuse and extraction capabilities, performs deep feature extraction, and improves gradient flow. Finally, in the output module, global average pooling aggregates information from the entire feature map, preventing overfitting. The features are then fed into a fully connected layer and classified using a Softmax classifier. The structure of RADN is shown in Figure 8.
In the STFT-RADN spectrum sensing algorithm, we selected n pairs of data ( S 1 , T 1 ) , , ( S n , T n ) as the training set and m pairs of data ( S n + 1 , T n + 1 ) , , ( S n + m , T n + m ) as the test set. Here, S represents the grayscale images obtained by applying the STFT and normalization to the received signals, and T represents the ground-truth classification label of the received signal. The input–output mapping of the whole network is as follows:
f w , b ( S n ) = T ^ n
where w and b are the trained weights and biases in the network, respectively, and T ^ represents the output mapping of S n after passing through the network. Due to the superiority of the cross-entropy loss function in binary image classification tasks, we have chosen it as our loss function. The cross-entropy loss function can be expressed as follows:
L o s s = i = 1 2 T n ( i ) log ( T ^ n ( i ) )
Algorithm 1 provides a detailed explanation of the spectrum sensing algorithm based on STFT-RADN.
Algorithm 1: Spectrum Sensing Algorithm of STFT-RADN
Input:Training and Testing Datasets
Output:Detection probability P d and false-alarm probability P f
Data Collection:Collect the signal samples X and preprocess the dataset into spectrograms S using STFT
Initialization:Gaussian initialized weights θ and maximum iteration number I t e r M a x
Training:Input the training set samples ( S n + 1 , T n + 1 ) , , ( S n + m , T n + m )
While i I t e r M a x = 200 :
a. Update θ by performing backpropagation with an Adam optimizer on the loss function
b. Increment i by 1
Save the model with lowest loss
Until maximum epochs ( I t e r M a x = 500) reached
Online Testing:Apply the trained RDAN model to the test dataset ( S n + 1 , T n + 1 ) , , ( S n + m , T n + m ) and output the classification results
Calculate the P d and the P f

4. Experimental Results and Discussion

4.1. Experimental Environment and Data

To validate the performance of the proposed STFT-RADN spectrum sensing algorithm, this section presents the relevant simulation experiments and analysis of the results. The simulations were conducted on a system equipped with an Intel(R) Core(TM) i7-13650HX CPU and a GeForce RTX 4060 GPU.
We used MATLAB 2023a to generate five types of primary user signals: BPSK, QPSK, 8PSK, 64QAM, and 16QAM. The SNR range was set from −20 dB to 10 dB, with intervals of 2 dB. For each modulation type and SNR level, 1000 samples were generated, with a sampling rate of 200 kHz.
Before performing STFT time–frequency analysis to obtain the spectrograms, additive white Gaussian noise (AWGN) was added to the signals, and spectrograms without primary user signals were generated based on a Gaussian distribution. The experiment also considered Rician multipath fading to simulate the random attenuation of wireless signals in a multipath propagation environment.
The obtained STFT time–frequency dataset was divided into training, validation, and testing sets in a ratio of 7:2:1.

4.2. Effect of Sampling Points

In this experiment, the detection probability of the STFT-RADN spectrum sensing model with different sampling points N is compared with the change in SNR. The experimental results are shown in Figure 9.
The experimental results show that detection probability increases with the number of sampling points. More sampling points result in a more significant improvement in detection probability under low SNR conditions (e.g., −20 dB to −10 dB). This indicates that more sampling points provide richer information, aiding in better signal detection under low SNR conditions.
However, more sampling points also mean more input data, leading to increased computational complexity and longer detection times. When the number of sampling points reaches 1024, the growth trend slows down. This is because too many sampling points can cause inconsistencies in the information before and after sampling, making it difficult for the network to learn more effective and relevant information.
Balancing computational complexity and detection probability, we chose 1024 as the optimal number of sampling points.

4.3. Effect of Window Function

In this experiment, we compared the detection probability of the STFT-RADN model using various window functions. We conducted experiments at various SNR levels using several window functions, including the rectangular, Hanning, Hamming, Gaussian, and Kaiser windows with different attenuation factors ( β ). The experimental results are presented in Figure 10. The results indicate that the Kaiser window with β = 0.85 provides the best overall performance, particularly under low SNR conditions, followed by the Hamming window. The Gaussian and rectangular windows demonstrate relatively poor performance. Consequently, this study adopts the Kaiser window with β = 0.85 .

4.4. Effectiveness of STFT

In this experiment, we compared the detection probability of datasets processed using traditional methods (cutting and splicing) [24] with that of datasets processed using STFT in the RADN model. The experimental results are shown in Figure 11.
The experimental results show that the STFT processing method significantly improves detection probability at low SNR values compared to traditional methods, with increases of 8.6%, 13.5%, 9.3%, 5.2%, and 3.5% at −20 dB, −14 dB, −10 dB, −4 dB, and 0 dB, respectively. This improvement occurs because STFT converts time-domain signals into the time–frequency domain. Even at low SNRs, the time–frequency representation provided by STFT can reveal the underlying signal structure, improving detection accuracy. This conversion helps capture both the temporal and spectral characteristics of the signal, enhancing detection and classification performance.
This demonstrates that using STFT to process datasets in the RADN model effectively captures signal information at low SNRs, thereby improving the overall performance of spectrum sensing.

4.5. Comparison of Detection Probability

In this experiment, we used the same dataset to compare the proposed RADN model with other state-of-the-art models. We evaluated the performance of our method against CNN [25], AlexNet [26], ResNet-18 [21], RDN (1), and RDN (2). Here, RDN (1) refers to the RADN model without the CBAM module, and RDN (2) replaces the res-inception block in RDN (1) with a basic residual block. The P d at different SNR levels was evaluated with P f = 0.01 . The experimental results are shown in Figure 12.
The figure shows that the RADN, RDN (1), and RDN (2) models perform significantly better than traditional CNN, AlexNet, and ResNet models, demonstrating the superior feature extraction capabilities of the RID structure.
Compared to ResNet, the introduction of dense connections in RDN (2) improves the overall detection probability by approximately 8% in the range from −20 dB to 0 dB. Additionally, replacing the residual block in RDN (1) with the res-inception block results in a further 6% increase in detection probability between −20 dB and −10 dB. This improvement indicates that the res-inception block effectively enhances feature extraction capabilities under low SNR conditions.
The best-performing model overall is our proposed RADN network, which incorporates the CBAM attention mechanism into RDN (1). Compared to RDN (1), RADN improves the detection probability P d by approximately 5% between −20 dB and −10 dB. This enhancement is attributed to the CBAM module’s attention mechanism, which helps to focus on important features and thus improves performance under low SNR conditions. Therefore, the RADN model demonstrates a significant advantage in spectrum sensing tasks, maintaining a high detection probability under low SNR conditions and markedly improving overall spectrum sensing performance.

4.6. Comparison of ROC

To further evaluate the detection performance of RADN, we recorded the P d and P f from a series of extensive spectrum sensing experiments. By comparing the proposed RADN with RDN (1), RDN (2), ResNet-18, and AlexNet, we generated receiver operating characteristic (ROC) curves. All experimental results were conducted at an SNR of −14 dB, and the results are shown in Figure 13.
The experimental results indicate that with a constant SNR, P d increases with P f for all models. However, our proposed RADN network consistently exhibits a higher P d at the same P f compared to other models. For instance, when P f is 0.01, the P d of RADN is 5.8%, 12.2%, 23.4%, and 38.3% higher than that of RDN (1), RDN (2), ResNet-18, and AlexNet, respectively. Therefore, the experiments demonstrate that our proposed RADN model exhibits superior performance in spectrum sensing tasks.

4.7. Impact of Different Modulation Schemes on Detection Performance

In this experiment, we evaluated the performance of different modulation schemes under RADN, including BPSK, QPSK, 8PSK, 64QAM, and 16QAM. The experimental results are shown in Figure 14. The figure shows that RADN maintains a consistent P d across various SNR levels for different modulation schemes (BPSK, QPSK, 8PSK, 64QAM, 16QAM). This consistency indicates that RADN can handle different modulation schemes without significant performance degradation. It demonstrates that our proposed method is insensitive to the modulation order, making it suitable for various modulated signals.

4.8. Comparison of Efficiency

To validate the sensing efficiency of the STFT-RADN algorithm, we compared it with CNN-based, STFT-ResNet-18, and RADN algorithms using the same dataset. Table 1 presents the offline training time and online sensing time for each algorithm.
Experimental results indicate that while the STFT-RADN algorithm has slightly higher training and sensing times than traditional CNN and RADN methods, it reduces these times compared to STFT-ResNet-18. Based on the experimental results mentioned above, the STFT-RADN method slightly increases detection time compared to RADN without STFT preprocessing, but significantly improves P d under low SNR conditions. Although the STFT-RADN method incurs higher training and sensing times than traditional CNN, it significantly enhances detection probability, with the additional detection time remaining within an acceptable range. Furthermore, RADN outperforms STFT-ResNet-18 in both effectiveness and detection probability. Therefore, the STFT-RADN method demonstrates strong sensing efficiency, making it suitable for real-time detection.

4.9. Discussion

The experimental results clearly demonstrate that the proposed STFT-RADN spectrum sensing method outperforms traditional CNN-based methods. It effectively overcomes limitations such as inadequate feature extraction and poor performance in low SNR conditions. However, the increased complexity of RADN and the use of STFT come with trade-offs, particularly in terms of higher computational demands. This method requires greater processing power and memory than traditional approaches. In practical applications, especially in resource-constrained environments, striking a balance between computational efficiency and detection performance is crucial for real-time spectrum sensing. Future research could focus on optimizing the RADN architecture or developing lightweight neural networks that maintain high performance while reducing computational requirements.

5. Conclusions

This paper proposes a spectrum sensing method based on STFT and RADN. Unlike existing deep-learning-based spectrum sensing methods that directly extract features from one-dimensional signals, the STFT-RADN spectrum sensing method utilizes STFT to transform received signals into time–frequency spectrograms. This approach better captures signal characteristics, thereby enhancing detection performance. The proposed RADN model introduces the CBAM attention mechanism and improves the basic residual block. By combining the advantages of dense connections and residual connections, it forms a powerful deep feature extraction structure known as RID. This significantly enhances the network’s feature extraction capabilities and improves spectrum sensing performance. Experimental results demonstrate that the STFT-RADN spectrum sensing method significantly improves performance under low SNR conditions compared to traditional deep-learning-based methods. This method achieves a high detection probability and strong robustness and adapts to various modulation schemes, indicating broad application prospects. Future work will explore the application of RADN in multiuser cooperative spectrum sensing methods.

Author Contributions

Conceptualization, A.W., T.Z. and Q.M.; data curation, T.Z.; methodology, A.W., T.Z. and Q.M.; writing—review and editing, A.W., T.Z. and Q.M.; funding acquisition, A.W. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (62471384).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, Q.; Su, B.; Wang, C.; Qian, L.P.; Wu, Y.; Yang, X. ConvLSTM-Based Spectrum Sensing at Very Low SNR. IEEE Wirel. Commun. Lett. 2023, 12, 967–971. [Google Scholar] [CrossRef]
  2. Arjoune, Y.; Kaabouch, N. A comprehensive survey on spectrum sensing in cognitive radio networks: Recent advances, new challenges, and future research directions. Sensors 2019, 19, 126. [Google Scholar] [CrossRef] [PubMed]
  3. Khaled, H.; Ahmad, I.; Habibi, D.; Phung, Q.V. A secure and energy-aware approach for cognitive radio communications. IEEE Open J. Commun. Soc. 2020, 1, 900–915. [Google Scholar] [CrossRef]
  4. Solanki, S.; Dehalwar, V.; Choudhary, J. Deep learning for spectrum sensing in cognitive radio. Symmetry 2021, 13, 147. [Google Scholar] [CrossRef]
  5. Costa, L.D.S.; Guimaraes, D.A.; Uchoa-Filho, B.F. On the signal-to-noise ratio wall of energy detection in spectrum sensing. IEEE Access 2022, 10, 16499–16511. [Google Scholar] [CrossRef]
  6. Shen, J.C.; Alsusa, E. Joint cycle frequencies and lags utilization in cyclostationary feature spectrum sensing. IEEE Trans. Signal Process. 2013, 61, 5337–5346. [Google Scholar] [CrossRef]
  7. Ma, L.; Li, Y.; Demir, A. Matched filtering assisted energy detection for sensing weak primary user signals. In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 3149–3152. [Google Scholar]
  8. Zhao, W.; Li, H.; Jin, M.; Liu, Y.; Yoo, S.-J. Eigenvalues-based universal spectrum sensing algorithm in cognitive radio networks. IEEE Syst. J. 2020, 15, 3391–3402. [Google Scholar] [CrossRef]
  9. Syed, S.N.; Lazaridis, P.I.; Khan, F.A.; Ahmed, Q.Z.; Hafeez, M.; Ivanov, A.; Poulkov, V.; Zaharis, Z.D. Deep neural networks for spectrum sensing: A review. IEEE Access 2023, 11, 89591–89615. [Google Scholar] [CrossRef]
  10. Chae, K.; Park, J.; Kim, Y. Rethinking autocorrelation for deep spectrum sensing in cognitive radio networks. IEEE Internet Things J. 2023, 10, 31–41. [Google Scholar] [CrossRef]
  11. Pan, G.; Li, J.; Lin, F. A cognitive radio spectrum sensing method for an OFDM signal based on deep learning and cycle spectrum. Int. J. Digit. Multimed. Broadcast. 2020, 2020, 5069021. [Google Scholar] [CrossRef]
  12. Liu, C.; Wang, J.; Liua, X.; Liang, Y.-C. Deep CM-CNN for spectrum sensing in cognitive radio. IEEE J. Sel. Areas Commun. 2019, 37, 2306–2321. [Google Scholar] [CrossRef]
  13. Han, D.; Sobabe, G.C.; Zhang, C.; Bai, X.; Wang, Z.; Liu, S.; Guo, B. Spectrum sensing for cognitive radio based on convolution neural network. In Proceedings of the 2017 10th International Congress on Image and Signal Processing, Biomedical Engineering and Informatics (CISP-BMEI), Shanghai, China, 14–16 October 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 1–6. [Google Scholar]
  14. Zheng, S.; Chen, S.; Qi, P.; Zhou, H.; Yang, X. Spectrum sensing based on deep learning classification for cognitive radios. China Commun. 2020, 17, 138–148. [Google Scholar] [CrossRef]
  15. Cai, L.; Cao, K.; Wu, Y.; Zhou, Y. Spectrum sensing based on spectrogram-aware CNN for cognitive radio network. IEEE Wirel. Commun. Lett. 2022, 11, 2135–2139. [Google Scholar] [CrossRef]
  16. Gai, J.; Zhang, L.; Wei, Z. Spectrum Sensing Based on STFT-ImpResNet for Cognitive Radio. Electronics 2022, 11, 2437. [Google Scholar] [CrossRef]
  17. Wang, A.; Meng, Q.; Wang, M. Spectrum Sensing Method Based on Residual Dense Network and Attention. Sensors 2023, 23, 7791. [Google Scholar] [CrossRef] [PubMed]
  18. Thomazella, R.; Lopes, W.N.; Aguiar, P.R.; Alexandre, F.A.; Fiocchi, A.A.; Bianchi, E.C. Digital signal processing for self-vibration monitoring in grinding: A new approach based on the time-frequency analysis of vibration signals. Measurement 2019, 145, 71–83. [Google Scholar] [CrossRef]
  19. Yao, S.; Piao, A.; Jiang, W.; Zhao, Y.; Shao, H.; Liu, S.; Liu, D.; Li, J.; Wang, T.; Hu, S.; et al. Stfnets: Learning sensing signals from the time-frequency perspective with short-time fourier neural networks. In Proceedings of the World Wide Web Conference, San Francisco, CA, USA, 13–17 May 2019; pp. 2192–2202. [Google Scholar]
  20. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  21. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  22. Szegedy, C.; Liu, W.; Jia, Y.; Sermanet, P.; Reed, S.; Anguelov, D.; Erhan, D.; Vanhoucke, V.; Rabinovich, A. Going deeper with convolutions. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 1–9. [Google Scholar]
  23. Zhang, Y.; Li, K.; Li, K.; Wang, L.; Zhong, B.; Fu, Y. Image super-resolution using very deep residual channel attention networks. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 286–301. [Google Scholar]
  24. Yue, C.; Xi, Z.; Wenbao, A.I. Spectrum sensing algorithm based on residual neural network. Mod. Electron. Tech 2022, 45, 1–5. [Google Scholar]
  25. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
  26. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. In Proceedings of the Advances in Neural Information Processing Systems 25, Lake Tahoe, NV, USA, 3–5 December 2012. [Google Scholar]
Figure 1. Spectrum sensing framework.
Figure 1. Spectrum sensing framework.
Sensors 24 05792 g001
Figure 2. (a) Time–frequency image of H 1 ; (b) time–frequency image of H 0 .
Figure 2. (a) Time–frequency image of H 1 ; (b) time–frequency image of H 0 .
Sensors 24 05792 g002
Figure 3. Convolutional block attention module.
Figure 3. Convolutional block attention module.
Sensors 24 05792 g003
Figure 4. Channel attention module.
Figure 4. Channel attention module.
Sensors 24 05792 g004
Figure 5. Spatial attention module.
Figure 5. Spatial attention module.
Sensors 24 05792 g005
Figure 6. (a) Residual block; (b) res-inception attention block.
Figure 6. (a) Residual block; (b) res-inception attention block.
Sensors 24 05792 g006
Figure 7. The structure of the dense group.
Figure 7. The structure of the dense group.
Sensors 24 05792 g007
Figure 8. The structure of the RADN.
Figure 8. The structure of the RADN.
Sensors 24 05792 g008
Figure 9. Effect of sampling points.
Figure 9. Effect of sampling points.
Sensors 24 05792 g009
Figure 10. Effect of window function.
Figure 10. Effect of window function.
Sensors 24 05792 g010
Figure 11. Effectiveness of STFT.
Figure 11. Effectiveness of STFT.
Sensors 24 05792 g011
Figure 12. Comparison of detection probability.
Figure 12. Comparison of detection probability.
Sensors 24 05792 g012
Figure 13. ROC curves of different algorithms.
Figure 13. ROC curves of different algorithms.
Sensors 24 05792 g013
Figure 14. Performance of different modulation schemes in RADN.
Figure 14. Performance of different modulation schemes in RADN.
Sensors 24 05792 g014
Table 1. Comparison of efficiency.
Table 1. Comparison of efficiency.
AlgorithmTraining Time/sSensing Time/s
CNN26.443.14
STFT-ResNet-1838.415.34
STFT-RADN32.583.74
RADN31.843.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, A.; Zhu, T.; Meng, Q. Spectrum Sensing Method Based on STFT-RADN in Cognitive Radio Networks. Sensors 2024, 24, 5792. https://doi.org/10.3390/s24175792

AMA Style

Wang A, Zhu T, Meng Q. Spectrum Sensing Method Based on STFT-RADN in Cognitive Radio Networks. Sensors. 2024; 24(17):5792. https://doi.org/10.3390/s24175792

Chicago/Turabian Style

Wang, Anyi, Tao Zhu, and Qifeng Meng. 2024. "Spectrum Sensing Method Based on STFT-RADN in Cognitive Radio Networks" Sensors 24, no. 17: 5792. https://doi.org/10.3390/s24175792

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop