Next Article in Journal
A Numerical Study on Entropy Generation in Two-Dimensional Rayleigh-Bénard Convection at Different Prandtl Number
Previous Article in Journal
Disproportionate Allocation of Indirect Costs at Individual-Farm Level Using Maximum Entropy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Automatic Modulation Classification Technique for Fading Channels via Deep Neural Network

Department of Electrical and Biomedical Engineering, Hanyang University, Seoul 04763, Korea
*
Author to whom correspondence should be addressed.
Entropy 2017, 19(9), 454; https://doi.org/10.3390/e19090454
Submission received: 31 July 2017 / Revised: 25 August 2017 / Accepted: 28 August 2017 / Published: 30 August 2017
(This article belongs to the Section Information Theory, Probability and Statistics)

Abstract

:
In this paper, we propose a deep neural network (DNN)-based automatic modulation classification (AMC) for digital communications. While conventional AMC techniques perform well for additive white Gaussian noise (AWGN) channels, classification accuracy degrades for fading channels where the amplitude and phase of channel gain change in time. The key contributions of this paper are in two phases. First, we analyze the effectiveness of a variety of statistical features for AMC task in fading channels. We reveal that the features that are shown to be effective for fading channels are different from those known to be good for AWGN channels. Second, we introduce a new enhanced AMC technique based on DNN method. We use the extensive and diverse set of statistical features found in our study for the DNN-based classifier. The fully connected feedforward network with four hidden layers are trained to classify the modulation class for several fading scenarios. Numerical evaluation shows that the proposed technique offers significant performance gain over the existing AMC methods in fading channels.

1. Introduction

In digital communications, a task of classifying a modulation class from the received signal is referred to as automatic modulation classification (AMC) [1,2,3,4]. While conventional communication systems use training signal or control channels to provide the information on the modulation type to the receiver, there are some scenarios where such information is not available and thus the modulation type should be blindly estimated from the received data. Such scenarios often occur in signal intelligence in military applications, signal sensing for cognitive radio systems, and inter-cell interference cancellation for cellular communications. Since the receiver does not have the knowledge on which data symbols have been sent from the transmitter, it is challenging to classify the modulation type solely based on the received data samples.
The previous AMC techniques can be roughly divided into two categories. The first approach is to build a probabilistic model for the received signal and classify a modulation class based on some optimality criterion such as maximum likelihood function. Though these methods offer optimal detection performance for the given model, its classification accuracy degrades in the presence of model mismatch. In addition, the model-based approach needs the knowledge on various model parameters, which requires substantial computational complexity. The alternative approach is the machine learning-based approach, where the machine is trained to classify the modulation type based on the training data in off-line and then the trained machine is deployed to apply to the real data. Assuming that the test data is sampled from the same distribution as that of the training data, the machine learning-based methods yield satisfactory classification performance without the exact knowledge of the system model. Since machine learning-based method is simple to design and robust to model mismatch, we are mainly concerned with this approach in this paper.
In general, two key steps should be performed for the machine learning-based AMC: (1) feature extraction step and (2) classification step. Widely used features for AMC include the variance of signal magnitude, frequency, and phase, wavelet coefficients and high order statistics such as cumulants [5,6,7,8]. After extracting the features, the classifier is applied to decide which class they belong to. So far, various kinds of classifiers have been used, including decision tree [5,7,9], support vector machine (SVM) [8,10,11], and artificial neural network (ANN) [12,13]. While these works have mostly focused on AMC task under additive white Gaussian noise (AWGN) channels, there exist only a few works presenting the AMC system designed for fading channels [14].
In this paper, we introduce an enhanced AMC technique based on deep neural network (DNN). First, we analyze the effectiveness of the extensive list of the statistical features for fading channels according to information theoretic criterion. We reveal that the features found to be powerful for fading channels can be quite different from those widely used for AWGN channels. We provide the extensive and diverse set of the features found by the forward greedy feature search based on mutual information. Second, the statistical features found by our analysis are used to classify the modulation type via DNN method. Recently, DNN has shown excellent performance in various classification tasks including image classification, speech recognition, natural language processing, and so on [15]. The deep architecture used in DNN provides great capacity to learn the complex structure of data from high dimensional input data. It has been shown that (1) a deep model requires a smaller number of parameters (or hidden units) to approximate the target function and (2) it can learn hierarchical features from low to high level through layers [15]. Such benefits from the deep model allow the DNN to learn complex function efficiently. In our work, the modulation classification task under fading channels is quite complex and the dimension of feature space is relatively high. In order to achieve good classification performance under adverse fading channel environment, our AMC method leverages the capability of DNN. The feedforward neural network with four hidden layers is trained to classify the modulation type for various fading scenarios. The extensive numerical evaluation show that the proposed scheme offers significantly better classification performance than the existing AMC methods in fading channels.
The rest of this paper is organized as follows. In Section 2, we describe the system model for the proposed AMC system. In Section 3, we present our analysis on the statistical features that are used for the proposed method. In Section 4, we present the detailed design of the proposed DNN-based AMC system. In Section 5, the simulation results are provided and in Section 6, the paper is concluded.

2. System Model

In this section, we describe the system model for the proposed AMC method. The transmitter converts the binary information (or coded) bits { b k } into the modulated symbols { a k } and transmits them through the transmit antenna over a carrier frequency f c r . The receiver acquires the transmitted signal using the receive antenna and converts it to the baseband samples. The AMC system processes the baseband samples to find the modulation format used in the transmitter. In digital communication systems, the baseband signal x ( t ) at the transmitter at time t can be expressed
x ( t ) = k = a k s ( t k T s ) ,
where T s is the symbol period, a k is the modulated symbol, and s ( t ) is the waveform for the transmit pulse shaping filter. According to the modulation class used, a k takes a value from different set of symbol values. For example, for quadrature phase shift keying (QPSK) modulation, a k takes a value from 1 2 + j 1 2 , 1 2 j 1 2 , 1 2 + j 1 2 , 1 2 j 1 2 [16]. Such baseband signal is converted to radio frequency (RF) band signal and transmitted through the time-varying channel whose impulse response is given by h ( t , τ ) , where t and τ are time and delay variables, respectively. At the receiver, the received signal is converted back to the baseband signal r ( t ) . The baseband signal r ( t ) is written by
r ( t ) = h ( t , τ ) x ( τ ) d τ + w ( t )
where the additive noise w ( t ) is modeled by white Gaussian process with zero mean [17]. From (1) and (2), we have
r ( t ) = k = a k s ( τ k T s ) h ( t , τ ) d τ + w ( t )
= k = a k g ( t , k T s ) d τ + w ( t ) ,
where g ( t , τ ) = s ( u τ ) h ( t , u ) d u . The received signal r ( t ) sampled at the symbol rate T s is given by
r ( n T s ) = k = a k g ( n T s , k T s ) + w ( n T s )
= a n g ( n T s , n T s ) + k n a k g ( n T s , k T s ) + w ( n T s ) .
Assuming that the delay spread of channel is much less than the symbol period T s , the second term in (6) can be negligibly small compared to the first term. In rich scattering environments where a number of multi-path channel components exist, the received sample sees the superposition of multiple copies of the transmitted signal, each traversing a different path. This causes constructive or destructive combining of the copies, consequently leading to fluctuation of the signal amplitude g ( n T s , n T s ) in time. This type of channel is often called flat fading channel [17]. The received signal sample r [ n ] ( = r ( n T s ) ) under flat fading channel can be simplified to
r [ n ] = g [ n ] a n + w [ n ] ,
where g [ n ] = g ( n T s , n T s ) and w [ n ] = w ( n T s ) . Various models can be used to describe the statistical characteristics of the channel gain g [ n ] . For rich scattering environment, Rayleigh fading channel model is widely adopted where the channel gain g [ n ] is assumed to be a zero-mean complex Gaussian and the temporal correlation of the channel gain g [ n ] is given by
E [ g n g n + k * ] = J 0 ( 2 π f d T s k )
where J 0 ( . ) is the zeroth-order Bessel function of the first kind, i.e., J 0 ( x ) = 1 π 0 π e x j cos θ d θ . When a strong ling of sight (LOS) component exists, Rician channel model is often used [17]. In order to perform AMC using the received data samples, we typically use the sampling rate 1 / T higher than the symbol rate 1 / T s . From now on, we assume that r [ n ] is sampled by the rate 1 / T , i.e., r [ n ] = r ( n T ) and we call O f = T s / T the oversampling factor.

3. Feature Extraction for AMC

In the feature extraction step for AMC, we convert the raw received data samples into the simpler form that can capture the distinctive pattern of the modulation classes. Suppose that we have a sequence of N samples of the received data r [ 1 ] , , r [ N ] . We convert them into the set of the K features f 1 , , f K . These features form a feature vector f ¯ = [ f 1 , , f K ] T in K dimensional vector space. In order to extract good features, the feature vectors belonging to different classes should be maximally separated from each other in feature space. In fading channels, the channel state fluctuates over time so that it is a common practice to use the statistics obtained from the sample average over long duration as features. Using such statistical features, we can have additional benefit of reducing the effect of background noise due to sample averaging. In our work, we provide the extensive set of statistical features that can provide rich and diverse information on the modulation class. They include the features that have been used for the existing AMC methods but some of them have not been used before. Table 1 lists 28 statistical features including cumulant, kurtosis, skewness, and peak to average ratio.
Some of variables used in the table are defined as
  • r [ i ] = r I [ i ] + j r Q [ i ]
  • ψ i = phase ( r [ i ] ) and f [ i ] = frequency ( r [ i ] )
  • m a = 1 N i = 1 N r i
  • σ d p = 1 c r n i > r t ψ N L 2 i 1 c r i > r t ψ N L i 2
  • σ a p = 1 c r n i > r t ψ N L 2 i 1 c r i > r t ψ N L i 2
  • σ a a = 1 N i = 1 N r c n 2 i 1 N i = 1 N r c n i 2
  • σ v = 1 N i = 1 N r v i 2 1 N i = 1 N r v i 2
  • σ a f = 1 N i = 1 N f N 2 i 1 N i = 1 N f N i 2
  • r c n i = r i m a 1 ,
  • r v i = r i σ r 2 1 , where σ r 2 = 1 N i = 1 N r i m a 2
  • ψ N L i = ψ i 1 N i = 1 N ψ i
  • f N i = f i 1 N i = 1 N f i T s
  • c u m x 1 , , x n = π π 1 ! 1 π 1 B π E i π x i
  • M p + q , p = E r i p r i * q
Joint cumulant of n variables x 1 , , x n is defined as c u m x 1 , , x n = π π 1 ! 1 π 1 B π E i π x i , where π runs through the list of all partitions of 1 , , n and B runs through the list of all blocks of the partition π . Thus, C n m is defined as C n m = c u m x , , x , x * , , x * , where x is used n m times and the conjugated variable x * is used m times. For example, the fourth-order cumulant is given by C 42 = c u m ( x , x , x * , x * ) and the sixth-order cumulant is given by C 62 = c u m ( x , x , x , x , x * , x * ) [18].
Now, we analyze the effectiveness of the statistical features listed in Table 1 in fading channels. In order to evaluate the quality of features, we use the information theoretic criterion called mutual information. Let the modulation class be represented by a discrete random variable c. Then, the mutual information between the ith feature f i and the modulation class c is expressed as
I ( f i , c ) = P ( f i , c ) log p ( f i , c ) P ( f i ) P ( c ) d f i d c ,
where P ( f i , c ) is the joint probability density for f i and c. Higher mutual information implies that the feature f i contains more information relevant to the modulation class. Since it is difficult to have accurate knowledge on the joint distribution P ( f i , c ) , we calculate (9) numerically using the Parzen window method. Specifically, we generate the received data samples using the model described in Section 2, create a number of K-dimensional feature vectors, and calculate (9) for each element of the feature vector. Table 2 provides the features sorted in descending order of mutual information.
We consider five modulation classes including binary phase shift keying (BPSK), QPSK, 8 phase shift keying (PSK), 16 quadrature amplitude modulation (QAM), 64 QAM. The random variable c is uniformly distributed between 1 and 5. We use 20,000 received data samples to obtain the feature vector and generate 1000 feature vectors to calculate the mutual information. To evaluate mutual information for different channel conditions, we consider the following three scenarios;
  • Scenario 1. AWGN channel, signal to noise power ratio SNR = 5 dB
  • Scenario 2. Rayleigh flat fading channelm SNR = 5 dB, Doppler frequency = 50 Hz
  • Scenario 3. Rayleigh flat fading channel, SNR = 5 dB, Doppler frequency = 100 Hz
  • Scenario 4. Rayleigh flat fading channel, both SNR and Doppler frequency are unknown. SNR is uniformly distributed between [ 5 , 15] dB and the Doppler frequency lies between [50, 100] Hz.
Note that in Scenario 4, the SNR and Doppler frequency are assumed to be uncertain so that channel parameters, SNR and Doppler frequency are randomly chosen within their given range when we generate the feature vectors. In Table 2, we observe that the result of ranking is quite different for four scenarios. For example, f 7 and f 17 (i.e., C 61 and σ v ) are fairly good features for AWGN channels (in Scenario 1) but not for fading channels (in Scenarios 2–4). This implies that the features effective for AWGN channels can lose relevance to the modulation class for the fading channels.
Next, we consider a problem of selecting the best L ( < K ) features according to information theoretic criterion. In order to find the best group of the features, we can evaluate the mutual information between the set of the features S L and the modulation class c
I ( S L ; c ) = P ( S L , c ) log P ( S L , c ) P ( S L ) P ( c ) d S L d c
Unfortunately, feature selection based on (10) requires evaluation of mutual information for all K L possible sets of L features. Furthermore, multi-dimensional integration in (10) needs high computational complexity. In order to alleviate this issue, we adopt a sub-optimal greedy feature search which selects locally best individual feature one by one. More specifically, we start with empty set ( S 0 = ) and add the strongest feature into the set at each iteration until L features are found. Since the set of features selected grow at every iteration, this is called forward search strategy. There exist more sophisticated feature selection strategies such as forward-backward search [19] but exploration of them is not in the scope of this paper. The criterion for the greedy feature selection is given by [20]
L m ( f j ) = I f j ; c 1 m 1 f i S m 1 I f j ; f i ,
where m is the iteration index and S m is the set of features collected at the mth iteration. At each iteration, the metric in (11) is evaluated for the features not selected yet and the strongest feature is added to the set S m . While the first term in the right hand side of (11) indicates the relevance of each feature to the modulation class, the second term penalizes the features with high dependency with the features previously found. Note that the second term is introduced to avoid selecting redundant features. Table 3 provides the list of features in the order of how the features are picked. If we desire to select the best L features, we only have to take the first L features from the list. While at the first iteration, the selected feature is equal to the top ranked feature obtained according to the metric (9), the features selected afterwards are different due to the term accounting for the redundancy between the features. We observe from Table 3 that the forward greedy search results in different set of features for the AWGN and fading channels. The last remaining step is to determine the number of the features L. Actually, it depends on which classifier we use and one appropriate way to choose L that leads to the smallest validation error.

4. DNN-Based AMC Method

Figure 1 depicts the structure of the proposed DNN-AMC method (DNN-AMC). The received base-band signal samples are obtained by the matched filter followed by the 1 / T rate-sampler. Then, the feature extractor is applied to convert the base-band signal samples into the L-dimensional feature vector f ¯ = [ f 1 , , f L ] T . After normalization of the feature vector, the DNN is applied to produce the final classification result in the form of probability that the given modulation class is used. The proposed system goes through two steps: (1) training stage and (2) test stage. In training stage, we generate the M feature vectors f ¯ ( 1 ) , , f ¯ ( M ) from the received data samples and optimize the weights of the DNN using them. In test stage, the trained DNN system is applied to the received data samples collected independently from the training data set for evaluation.

4.1. DNN Structure

First, we scale the received signal samples such that they have zero mean and unit variance and calculate the feature vector [ f 1 , , f L ] T from them. Then, we normalize the feature vector as
f i = ( f i μ i ) σ i .
Note that μ i and σ i are obtained from
μ i = 1 M k = 1 M f i ( k )
σ i = 1 M k = 1 M | f i ( k ) μ i | 2
where f i ( 1 ) , , f i ( M ) are the ith element of the M feature vectors in the training set. Note that in the test phase, we use the fixed value of μ i and σ i that have computed in the training step. The normalization step is nothing but a linear transformation of the features, which makes the feature extraction invariant to any scaling of signal samples. Thus, this normalization step makes the system more robust in real scenarios. The normalized feature vector [ f 1 , , f L ] is fed into the DNN which consists of the fully connected feedforward network with Q = 4 hidden layers. Figure 2 depicts the structure of DNN. The rectified linear unit (ReLU) function is used as a nonlinear function [15]. The number of hidden nodes is set to 500, 200, 40, 5 from the input to the output layer, which is determined based on intensive empirical simulations. At the output layer, softmax function is used to produce the probability score associated with each modulation class. That is, the probability score y k associated with the kth modulation class is given by [21]
y k = exp ( a k ) j exp ( a j ) ,
where a j is the activation corresponding to the jth output node. The final classification result is given by choosing the modulation class with the highest probability score.

4.2. Training

As mentioned above, the goal of the training step is to optimize the weights of the DNN using the training data. In the training data set, each feature vector is associated with the class label, which is represented by one hot encoding [21]. We use an appropriate loss function to optimize the weights of the DNN for the given task. For multi-class classification task, negative log-likelihood function is widely used [21]
L = i = 1 M log P c ( i ) | [ f 1 ( i ) , , f L ( i ) ] T , W , b
where W and b are the weight matrix and bias vector and c ( i ) and [ f 1 ( i ) , , f L ( i ) ] T are the class label and the feature vector for the ith training example, respectively. The stochastic gradient descent method with the mini-batch size B is used to minimize the loss function with respect to the weights of DNN. Standard back propagation algorithm is used to obtain the gradient of loss function with respect to the weights at each layer [21]. Note that 20% of training data is not used for the training but for validation. We start with the initial learning rate ϵ = 0 . 001 , reduce the learning rate by half when the validation error stops to improve, and then, terminate training if validation error converges to fixed value. Note that the training procedure mentioned above is conducted for four scenarios mentioned in Section 3. Note that we generate the training set for Scenario 4 using the randomly sampled channel parameters in order to consider the scenario where the channel parameters are not known. Table 4 summarizes the training setup used in our experiments.

4.3. Test

Once the weights and biases of the DNN are determined in the training step, we apply the trained DNN-AMC to the test data and evaluate the accuracy of the proposed method. We generate 10,000 test feature vectors and count the number of classification errors for performance evaluation.

5. Experiments

In this section, we evaluate the performance of the proposed algorithm through computer simulations.

5.1. Simulation Setup

In simulations, we consider the classification task that selects one of the five modulation classes, BPSK, QPSK, 8PSK, 16QAM, and 64QAM. We use a raised cosine filter with the roll-off factor 0.2. The symbol rate is set to 10,000 sym/s, and the oversampling rate is set to 10. A feature vector is obtained by processing 20,000 samples (corresponding to 2000 symbols over 0.2 s). M = 24,000 feature vectors are used to train the DNN and 6000 feature vectors are used for validation. We use 10,000 test examples (2000 examples per class) to evaluate the classification accuracy. The classification accuracy for the modulation class c is defined by how many feature vectors are correctly classified among all feature vectors labeled by c.

5.2. Simulation Results

We propose two versions of the proposed DNN-AMC method. The first version called DNN-AMC1 uses all 28 features for classification while the second version, DNN-AMC2 ( L ) uses the L features selected based on the list in Table 3. We compare the proposed algorithm with the following existing state of the art AMC methods;
  • Artificial neural network algorithm 1 (ANN1) [12]: 10 features ( C 20 , C 21 , C 40 , C 41 , C 42 , C 60 , C 61 , C 62 , C 63 , C 80 ) are used with the feedforward neural network with single hidden layer.
  • Artificial neural network algorithm 2 (ANN2) [13]: 6 features ( β , v 20 , σ a a , σ a p , σ d p , the mean of signal magnitude) are used with the feedforward neural network with single hidden layer.
  • Support vector machine algorithm (SVM) [10]: 4 feature vectors ( C 40 , C 41 , C 42 , C 63 ) are used with SVM.
  • Hierarchical classification scheme (HCS) [9]: 3 feature vectors ( C 20 , C 40 , C 42 ) are used with decision tree classifier.
Note that ANN1, ANN2, SVM, and HCS are trained with the same training data set used for the proposed method. The threshold for the HCS scheme is empirically found for fading channels.
We evaluate the performance of the proposed method. For Scenario 1, all AMC schemes considered achieve 100% classification accuracy for Gaussian channel. Table 5 and Table 6 show the classification accuracy for Scenarios 2 and 3, respectively. For both scenarios, the proposed methods, DNN-AMC1 and DNN-AMC2 achieve the significant performance gain over the other AMC methods. We observe that in fading channels, the performance of the proposed AMC method is slightly degraded for higher doppler frequency but significant performance gain of the proposed method is retained. We also observe that the DNN-AMC2 achieves the performance comparable to the DNN-AMC1 which uses the whole set of features considered. This shows that the DNN appropriately controls the contribution from the features depending on their reliability. Table 7 shows the classification accuracy when the SNR and doppler frequency are randomly distributed. Due to the uncertainty of channel parameters, the classification accuracy of the AMC algorithms is worse in Scenario 4 as compared to other scenarios. Nevertheless, our method outperforms the other AMC schemes significantly as shown in the Table 7.

6. Conclusions

In this paper, we proposed the DNN-based AMC method which can achieve robust classification performance for fading channels. First, we analyzed a variety of statistical features used to perform AMC based on the information theoretic criterion. From this study, we find the key features that are strong for fading channels. The DNN is applied to classify the modulation class based on the statistical features found in our study. Using the feed-forward neural network with four hidden layers, the proposed method achieved good classification performance even when a number of statistical features are used.

Acknowledgments

This work was supported by the research fund of Signal Intelligence Research Center supervised by Defense Acquisition Program Administration and Agency for Defense Development of Korea.

Author Contributions

Jung Hwan Lee, Byeoungdo Kim and Jaekyum Kim conducted the research under supervision of Jun Won Choi and Dongwoen Yoon.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Zhu, Z.; Nandi, A.K. Automatic Modulation Classification: Principles, Algorithms and Applications; John Wiley & Sons: River Street Hoboken, NJ, USA, 2014. [Google Scholar]
  2. Dobre, O.A.; Abdi, A.; Bar-Ness, Y.; Su, W. Blind modulation classification: A concept whose time has come. In Proceedings of the IEEE/Sarnoff Symposium on Advances in Wired and Wireless Communication, Princeton, NJ, USA, 18–19 April 2005; pp. 223–228. [Google Scholar]
  3. Hazza, A.M.; Alshebeili, S.; Fahad, A. An overview of feature-based methods for digital modulation classification. In Proceedings of the 1st International Conference on Communications, Signal Processing, and Their Applications, Sharjah, UAE, 12–14 February 2013; pp. 1–6. [Google Scholar]
  4. Azzouz, E.; Nandi, A.K. Automatic Modulation Recognition of Communication Signals; Springer Science & Business Media: New York, NJ, USA, 2013. [Google Scholar]
  5. Dobre, O.A.; Abdi, A.; Bar-Ness, Y.; Su, W. The classification of joint analog and digital modulations. In Proceedings of the IEEE Military Communications Conference, Atlantic City, NJ, USA, 17–21 October 2005; pp. 3010–3015. [Google Scholar]
  6. An, N.; Li, B.; Huang, M. Modulation classification of higher order MQAM signals using mixed-order moments and Fisher criterion. In Proceedings of the 2nd International Conference on Computer and Automation Engineering, Singapore, 26–28 February 2010; pp. 150–153. [Google Scholar]
  7. Huang, F.; Zhong, Z.; Xu, Y.; Ren, G. Modulation recognition of symbol shaped digital signals. In Proceedings of the International Conference on Communications, Circuits and Systems, Xiamen, China, 25–27 May 2008; pp. 328–332. [Google Scholar]
  8. Xin, Z.; Ying, W.; Bin, Y. Signal classification method based on support vector machine and high-order cumulants. Wirel. Sens. Netw. 2010, 2, 48. [Google Scholar]
  9. Swami, A.; Sadler, B.M. Hierarchical digital modulation classification using cumulants. IEEE Trans. Commun. 2000, 48, 416–429. [Google Scholar] [CrossRef]
  10. Li, P.; Zhang, Z.; Wang, X.; Xu, N.; Xu, Y. Modulation recognition of communication signals based on high order cumulants and support vector machine. J. China Univ. Posts Telecommun. 2012, 19, 61–65. [Google Scholar] [CrossRef]
  11. Han, G.; Li, J.; Lu, D. Study of modulation recognition based on HOCs and SVM. In Proceedings of the IEEE 59th Vehicular Technology Conference, Milan, Italy, 17–19 May 2004; pp. 898–902. [Google Scholar]
  12. Popoola, J.J.; van Olst, R. Automatic recognition of analog modulated signals using artificial neural networks. Comput. Technol. Appl. 2013, 2, 29–35. [Google Scholar]
  13. Popoola, J.J.; van Olst, R. Effect of training algorithms on performance of a developed automatic modulation classification using artificial neural network. In Proceedings of the IEEE AFRICON, Pointe-Aux-Piments, Mauritius, 9–12 September 2013; pp. 1–6. [Google Scholar]
  14. Xi, S.; Wu, H.C. Robust automatic modulation classification using cumulant features in the presence of fading channels. In Proceedings of the IEEE Wireless Communications and Network Conference, Las Vegas, NV, USA, 3–6 April 2006; pp. 2094–2099. [Google Scholar]
  15. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  16. Goldsmith, A. Wireless Communications; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  17. Tse, D.; Viswanath, P. Fundamentals of Wireless Communications; Cambridge University Press: Cambridge, UK, 2005. [Google Scholar]
  18. Orlic, V.D.; Dukic, M.L. Automatic modulation classification: Sixth-order cumulant features as a solution for real-world challenges. In Proceedings of the 20th Telecommunications Forum, Belgrade, Serbia, 20–22 November 2012; pp. 392–399. [Google Scholar]
  19. Guyon, I.; Elisseeff, A. An introduction to variable and feature selection. J. Mach. Learn. Res. 2003, 3, 1157–1182. [Google Scholar]
  20. Peng, H.; Long, F.; Ding, C. Feature selection based on mutual information criteria of max-dependency, max-relevance, and min-redundancy. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27, 1226–1238. [Google Scholar] [CrossRef] [PubMed]
  21. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: New York, NY, USA, 2009. [Google Scholar]
Figure 1. The structure of the proposed AMC system.
Figure 1. The structure of the proposed AMC system.
Entropy 19 00454 g001
Figure 2. The structure of deep neural network (DNN) architecture.
Figure 2. The structure of deep neural network (DNN) architecture.
Entropy 19 00454 g002
Table 1. List of statistical features used for the proposed automatic modulation classification (AMC) method.
Table 1. List of statistical features used for the proposed automatic modulation classification (AMC) method.
f 1 C 20 f 2 C 21 f 3 C 40
f 4 C 41 f 5 C 42 f 6 C 60
f 7 C 61 f 8 C 62 f 9 C 63
f 10 C 80 f 11 σ a p f 12 σ a a
f 13 σ d p f 14 σ a f f 15 v 20 = M 4 , 2 M 2 , 1 2 = i = 1 N r i 4 i = 1 N r i 2
f 16 β = i = 1 N r Q 2 i i = 1 N r I 2 i f 17 σ v f 18 C 81
f 19 C 82 f 20 C 83 f 21 C 84
f 22 C 10 , 0 f 23 C 10 , 1 f 24 C 10 , 2
f 25 K = i = 1 N r i m a 4 i = 1 N r i m a 2 2 (Kurtosis) f 26 S = i = 1 N r i m a 3 i = 1 N r i m a 2 3 2 (Skewness) f 27 P A = m a x r : 1 N i = 1 N r i (Peak to average ratio)
f 28 P R = m a x r : 2 1 N i = 1 N r i 2 (Peak to RMS ratio)
Table 2. List of features sorted in the order of the relevance to the modulation class.
Table 2. List of features sorted in the order of the relevance to the modulation class.
ScenarioFeatures
1 f 25 , f 12 , f 15 , f 7 , f 5 , f 3 , f 17 , f 14 , f 13 , f 9 , f 11 , f 27 , f 16 , f 10 , f 19 , f 28 , f 21 , f 2 , f 26 , f 6 , f 23 , f 18 , f 20 , f 8 , f 1 , f 24 , f 4 , f 22
2 f 15 , f 25 , f 1 , f 16 , f 12 , f 3 , f 27 , f 11 , f 4 , f 28 , f 8 , f 24 , f 7 , f 14 , f 20 , f 18 , f 6 , f 19 , f 5 , f 9 , f 22 , f 13 , f 17 , f 26 , f 23 , f 10 , f 21 , f 2
3 f 15 , f 12 , f 5 , f 25 , f 1 , f 27 , f 16 , f 28 , f 3 , f 24 , f 11 , f 4 , f 8 , f 7 , f 20 , f 14 , f 19 , f 18 , f 6 , f 22 , f 9 , f 17 , f 21 , f 10 , f 23 , f 26 , f 13 , f 2
4 f 15 , f 25 , f 12 , f 1 , f 16 , f 27 , f 28 , f 3 , f 4 , f 5 , f 8 , f 11 , f 24 , f 7 , f 14 , f 6 , f 18 , f 20 , f 19 , f 9 , f 22 , f 17 , f 21 , f 13 , f 26 , f 10 , f 23 , f 2
Table 3. The list of features sorted in the order of that they are selected based on the forward greedy search.
Table 3. The list of features sorted in the order of that they are selected based on the forward greedy search.
ScenarioFeatures
1 f 25 , f 7 , f 12 , f 14 , f 15 , f 5 , f 13 , f 17 , f 3 , f 11 , f 9 , f 10 , f 15 , f 19 , f 16 , f 21 , f 28 , f 26 , f 6 , f 18 , f 23 , f 20 , f 8 , f 1 , f 4 , f 24 , f 22 , f 2
2 f 15 , f 25 , f 1 , f 11 , f 4 , f 3 , f 16 , f 24 , f 27 , f 14 , f 8 , f 12 , f 7 , f 6 , f 9 , f 22 , f 20 , f 13 , f 26 , f 18 , f 5 , f 28 , f 19 , f 23 , f 17 , f 21 , f 10 , f 2
3 f 15 , f 1 , f 25 , f 4 , f 11 , f 5 , f 24 , f 12 , f 3 , f 14 , f 16 , f 27 , f 8 , f 7 , f 9 , f 20 , f 6 , f 19 , f 18 , f 26 , f 28 , f 22 , f 13 , f 28 , f 21 , f 10 , f 23 , f 2
4 f 15 , f 16 , f 25 , f 1 , f 5 , f 4 , f 11 , f 28 , f 3 , f 14 , f 9 , f 7 , f 26 , f 12 , f 13 , f 24 , f 8 , f 27 , f 6 , f 21 , f 19 , f 22 , f 17 , f 20 , f 18 , f 23 , f 10 , f 2
Table 4. Training parameters.
Table 4. Training parameters.
The Number of Data Samples Converted to a Feature Vector ( P ) 20,000
The Number of Feature Vectors in Training Set ( M ) 30,000
The Number of Feature Vectors in Validation Set ( M ) 6000
Initial Learning Rate ϵ 0 . 001
Mini-Batch Size ( B ) 50
Table 5. Classification accuracy of several AMC schemes for Scenario 2.
Table 5. Classification accuracy of several AMC schemes for Scenario 2.
BPSKQPSK8PSK16QAM64QAM
DNN-AMC199.9893.7597.5872.1674.64
DNN-AMC2 (14)99.9890.1795.2263.6169.29
ANN197.8886.1691.7557.5961.93
ANN299.6996.1440.0057.3466.09
SVM81.962.7074.2025.4028.60
HCS19.2240.1065.0731.3828.27
Table 6. Classification accuracy of several AMC schemes for Scenario 3.
Table 6. Classification accuracy of several AMC schemes for Scenario 3.
BPSKQPSK8PSK16QAM64QAM
DNN-AMC199.9790.3795.6669.2370.85
DNN-AMC2 (14)99.9886.7992.2267.1467.7
ANN196.6584.1090.5556.3261.71
ANN299.9063.6233.2756.169.14
SVM76.365.974.426.439.1
HCS23.3741.0077.7125.5549.6
Table 7. Classification accuracy of several AMC schemes for Scenario 4.
Table 7. Classification accuracy of several AMC schemes for Scenario 4.
BPSKQPSK8PSK16QAM64QAM
DNN-AMC196.6178.6285.4260.2266.35
DNN-AMC2 (14)95.5876.8783.9452.9664.83
ANN192.3972.2484.4042.9758.82
ANN284.6244.0250.8145.5765.90
SVM67.2653.758.2626.8633.92
HCS18.0441.2850.8135.0929.24

Share and Cite

MDPI and ACS Style

Lee, J.H.; Kim, J.; Kim, B.; Yoon, D.; Choi, J.W. Robust Automatic Modulation Classification Technique for Fading Channels via Deep Neural Network. Entropy 2017, 19, 454. https://doi.org/10.3390/e19090454

AMA Style

Lee JH, Kim J, Kim B, Yoon D, Choi JW. Robust Automatic Modulation Classification Technique for Fading Channels via Deep Neural Network. Entropy. 2017; 19(9):454. https://doi.org/10.3390/e19090454

Chicago/Turabian Style

Lee, Jung Hwan, Jaekyum Kim, Byeoungdo Kim, Dongweon Yoon, and Jun Won Choi. 2017. "Robust Automatic Modulation Classification Technique for Fading Channels via Deep Neural Network" Entropy 19, no. 9: 454. https://doi.org/10.3390/e19090454

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop