Next Article in Journal
Lossless Data Compression for Time-Series Sensor Data Based on Dynamic Bit Packing
Next Article in Special Issue
An Optimization Method for Non-IID Federated Learning Based on Deep Reinforcement Learning
Previous Article in Journal
A Lightweight Recurrent Grouping Attention Network for Video Super-Resolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Specific Emitter Identification System Design for Crossing Signal Modes in the Air Traffic Control Radar Beacon System and Wireless Devices

1
School of Computer Science, Sichuan University, Chengdu 610065, China
2
Sichuan Jiuzhou Electric Group Co., Ltd., Mianyang 621000, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(20), 8576; https://doi.org/10.3390/s23208576
Submission received: 25 September 2023 / Revised: 16 October 2023 / Accepted: 17 October 2023 / Published: 19 October 2023
(This article belongs to the Special Issue AI-Based Security and Privacy for IoT Applications)

Abstract

:
To improve communication stability, more wireless devices transmit multi-modal signals while operating. The term ‘modal’ refers to signal waveforms or signal types. This poses challenges to traditional specific emitter identification (SEI) systems, e.g., unknown modal signals require extra open-set mode identification; different modes require different radio frequency fingerprint (RFF) extractors and SEI classifiers; and it is hard to collect and label all signals. To address these issues, we propose an enhanced SEI system consisting of a universal RFF extractor, denoted as multiple synchrosqueezed wavelet transformation of energy unified (MSWTEu), and a new generative adversarial network for feature transferring (FTGAN). MSWTEu extracts uniform RFF features for different modal signals, FTGAN transfers different modal features to a recognized distribution in an unsupervised manner, and a novel training strategy is proposed to achieve emitter identification across multi-modal signals using a single clustering method. To evaluate the system, we built a hybrid dataset, which consists of multi-modal signals transmitted by various emitters, and built a complete civil air traffic control radar beacon system (ATCRBS) dataset for airplanes. The experiments show that our enhanced SEI system can resolve the SEI problems associated with crossing signal modes. It directly achieves 86% accuracy in cross-modal emitter identification using an unsupervised classifier, and simultaneously obtains 99% accuracy in open-set recognition of signal mode.

1. Introduction

With the great advancements in science and technology, wireless devices have been applied in various fields, such as radar systems, the Internet of Things (IoT), and more. As a result, the security of wireless communication has attracted considerable attention [1]. The address resolution of the transmitted protocol is a traditional method used to identify wireless devices; however, it has practical limitations, as it can be easily intercepted and tampered with. To resolve these problems, reference [2] first proposed the concept of a radio frequency fingerprint (RFF) for transient signals, which is unique to each wireless transmitter. The various manufacturing biases in different hardware components of wireless devices result in each device having a unique RFF, like human fingerprints. Moreover, reference [3] explored a method to extract RFF from steady signals, which are easier to collect. This further advanced the development of RFF technology.
The enhanced RFF technique greatly improves the security of identification within the wireless domain. Furthermore, the introduction of the RF machine learning system by the Defense Advanced Research Projects Agency (DARPA) has propelled RFF research to pivot toward deep learning (DL) [4]. With the improvement of RFF and DL [5,6], SEI has also made great progress in distinguishing specific emitters based on physical layer characteristics. Typically, popular SEI systems consist of an RFF feature extractor and a DL classifier. References [7,8,9,10] extracted robust RFF features to improve the performance of SEI in complex environments. Moreover, references [11,12] proposed complex-valued convolutional neural networks (CVCNNs) to enhance the adaptability of the DL model when dealing with complex-valued signals in SEI, and compressed CVCNNs to save calculation resources.
Although there have been many breakthroughs in the field of SEI, previous works have mainly focused on optimizing a typical signal for higher robustness. In practical applications, working devices tend to use more than one modal signal to address the vulnerability or interference of a single modal signal in challenging environments. Multimodal signals usually operate in different frequency bands or have different modulation methods, which can effectively solve this problem. For example, in fast flight mode, high-frequency signals are extremely sensitive to Doppler shifts, while low-frequency signals are less affected by speed changes. Moreover, aircraft broadcast multi-modal signals, such as automatic dependent surveillance–broadcast (ads-b) signals, air traffic control (atc) signals [13], and radar signals, enabling accurate detection and guidance by air traffic controllers on the ground. Reference [14] explored the RFFs of aircraft based on radar signals. Moreover, references [15,16,17] explored the RFFs of airplanes, targeting the ads-b s modal signal. The traditional ads-b signal represents the ads-b s mode, which is also used in this paper. We use rm1/2 to represent the ads-b a/c mode in this paper, as seen in Table 1. Moreover, civil ATCRBS contains the ads-b a/c/s mode, which might be emitted by different parts of the same aircraft. Therefore, it is crucial for a robust SEI system to recognize wireless devices across different modal signals.
Typically, existing SEI systems [5,6,10,19,20] consist of RFF feature extractors and emitter classifiers, both of which are based on a single modal signal. However, when dealing with multi-modal signals, it is necessary to first identify the signal mode as a signal arrives. Moreover, different modal signals require different RFF extractors and SEI classifiers, resulting in increased complexity and reduced robustness when dealing with unknown modal signals. To tackle the challenge of identifying wireless devices across different modal signals in a more cost-effective manner, one possible solution is to develop a universal RFF extractor that unifies the structures and semantics for multi-modalities. Additionally, it is crucial to introduce DL methods to improve the adaptability of classifiers for multi-modal signals [21,22]. A previous study [21] proposed pre-training a SEI DL classifier using ads-b signals and subsequently fine-tuning it to recognize emitters with Wi-Fi signals. However, this approach requires continuous fine-tuning whenever new modal signals are introduced, which also leads to building multiple models for different modal signals.
To resolve these issues, the MSWTEu is proposed to extract general and robust RFF features, unifying the structures and semantics of different modal signals. However, solely relying on the general RFF extraction method may not be sufficient, as some modal signals exhibit significant differences that MSWTEu cannot fully unify. To address this limitation, we propose a novel FTGAN to enhance the adaptation of the entire SEI system to multi-modal signals. Since it is impractical to build signal models for all existing modal signals, we focus on analyzing and obtaining information from one representative modal signal, which we refer to as the “known modal signal“. The remainder of the modal signals are deemed as “unknown modal signals”. The FTGAN tries to transfer the MSWTEu distributions of unknown modal signals to match the distribution of the known modal signal. With the above methods, only one SEI classifier is required to be built and trained for the known modal signal. But when confronted with other modal signals, the SEI classifier can also recognize their emitters and achieve improved comprehensive accuracy recognition. Furthermore, with slight modifications to the trained SEI classifier, open-set signal mode recognition can be achieved. An overview of traditional SEI systems, as well as our system addressing multi-modal signals, is depicted in Figure 1.
The main contributions of this paper are as follows:
1.
We explore an efficient and general method for extracting prominent RFF features from multi-modal signals;
2.
We explore the MSWTEu algorithm for unifying the structure and semantics of different modal signals;
3.
We explore an FTGAN to further unify RFF features, thus enhancing the efficiency and performance of the SEI system while accommodating multi-modal signals. With this approach, the SEI classifier is trained only with known modal signal features, but it can also recognize emitters using unknown modal signals without additional training;
4.
We leverage MSWTEu and FTGAN; the trained SEI classifier can achieve open-set modulation recognition with minimal fine-tuning;
5.
We construct a hybrid dataset for SEI experiments crossing signal modes, consisting of a real dataset, a simulated dataset, and a public dataset. The public dataset has been widely used in recent works and consists of Wi-Fi signals. The simulated dataset comprises two modal communication reply signals, whose waveforms are carefully designed and generated by different waveform generators to simulate emitters. The real dataset is comprised of ads-b signals collected over a span of 12 months; it has undergone comprehensive preprocessing. Each modal signal is sent by three different emitters. The details of the dataset are depicted in Table 1. The simulated dataset and real dataset consist of ads-b a/c/s modal signals, which can be built as a complete dataset of civil ATCRBS (CATCRBS). With the datasets, our optimized SEI system can obtain the identification of airplanes, disregarding variations in signal modes and, thus, replacing the identification-decoding module within CATCRBS.
Therefore, our proposed system can solve the SEI problems posed by multi-modal signals at the same time: (1) MSWTEu allows different modal signals to only utilize one RFF extractor, decreasing the number of RFF extractors; (2) FTGAN ensures that different modal signals require only one unsupervised classifier, solving the hard labeling work and reducing the number of classifiers; (3) the proposed training strategy for MSWTEu and FTGAN makes it unnecessary to add additional processes to recognize signal modes, increasing the efficiency of the whole system; (4) a little fine-tuning of the proposed SEI system enables open-set signal mode recognition. The rest of the paper is organized as follows. Section 2 describes the process of the feature extractor MSWTEu. In Section 3, we present the details of FTGAN and the training strategy. The dataset details and experimental results are shown in Section 4. Lastly, Section 5 summarizes this paper.

2. RFF Feature Extracting and Unifying MSWTEu

This work is mainly inspired by our previous work [13], which proposed a robust RFF extractor denoted as improving SWT by energy regularization (ISWTE), enhancing the performance of SEI at low SNRs. In a previous work [13], we fully discussed the advantages of wavelet transformation in improving RFF features and conducted experiments to compare ISWTE, SWT, spectrograms, and traditional signal processing methods. We concluded that SWT outperforms other RFF extractors in terms of accuracy and efficiency. Moreover, by optimizing SWT with energy regularization, ISWTE can enhance the performance of SEI under low SNRs. However, this kind of enhancement is limited to one particular modal signal. Thus, our aim is to expand our previous work, and further extract general and robust RFF features from multi-modal signals.
This work is roughly divided into three stages, as shown in Figure 2.
In the first stage, we propose a reasonable downsampling strategy to unify the lengths of multi-modal signals. This strategy helps to address the signal variation issue and ensures the effectiveness of subsequent processing.
The second stage focuses on improving the versatility and efficiency of the MSWTEu method based on the ISWTE method. While ISWTE decomposes signals into a fixed, limited number of layers, and adopts fixed parameters, which are manually adjusted to achieve the best performance, MSWTEu allows for optional layers and optional parameters for each layer. We propose the MSWTEu, which introduces an optional parameter for each layer, and design an adaptive selection algorithm to select the best SWT decomposition layers and the best parameter per layer. Here, we use the known modal dataset and pick a few unknown modal signals to join the selecting algorithm. The accuracy of a clustering algorithm is adopted to assess the performance. And in turn, the layers and parameters are adaptively determined by the best performance. By doing so, the MSWTEu method combines the advantages of multiple SWT layers and the automatic selection of the best parameters, eliminating the need for manual adjustments. This approach ensures that different modal signals have a unified structure and shape, and also achieves robust RFF feature extraction.
In the third stage, we address the energy explosion issue commonly encountered in SWT transformation and optimize MSWTEu for greater efficiency. The energy explosion is mainly concentrated in the latter decomposition layer from SWT, resulting in a substantial increase in the weight of these latter layers in the RFF feature. It could cause the information of previous layers to be missing. Therefore, the ISWTE is optimized by energy regularization, specifically, by computing the energy of each layer and reversely weakening them by an energy scale. In this way, ISWTE has achieved good performance in noisy domains based on ads-b signals. However, the time complexity of energy regularization by ISWTE is not ideal since the procedure of reverse weakening is complex. The MSWTEu aims to simplify the energy strategy and be more adaptive to multi-modal signals.
In conclusion, the MSWTEu overcomes the constraints of relying on one modal signal and manual parameter tuning during RFF feature extraction. It improves the efficiency of energy computation and resolves the problem of energy explosion.
The MSWTEu calculation process is shown in Figure 2 and the details are below:
(1) Wavelet transformation (WT): In practice, signals are usually downsampled upon reception and transformed to the discrete domain. Therefore, the signal is represented as x ( n ) , not x ( t ) . The normal wavelet transformation (WT) is calculated as below:
W a , b ( x ( n ) ) = 1 a n = 0 N x ( n ) φ ( n b a ) ,
where N is the length of x ( n ) , φ is the mother wavelet function, a scales the wavelet function, and b shifts the time domain. The WT can transform signals from the time domain to the time–frequency domain, by slicing the frequency domain based on the continuous short-time Fourier transform (STFT). Reference [3] proposed that the signal in the frequency domain performs better in SEI than in the time domain. Our previous experiments show that the signals in the time domain are vulnerable to noises but robust in Doppler shifts, while the signals in the frequency domain behave in the opposite way. The WT combined the advantages of the time domain and frequency domain, which is the main reason we adopted it as the base transformation of the RFF extractor. Normally, the WT can decompose signals into multiple layers, with each containing a high-frequency part (the detail coefficient) and a low-frequency part (the approximation coefficient).
(2) Synchrosqueezed wavelet transformation (SWT): Reference [23] proposed SWT, which is inspired by empirical mode decomposition (EMD) and confirms its effectiveness through mathematical methods. Unlike the traditional discrete wavelet transform (DWT), EMD does not require much prior knowledge and is more stable and adaptive in signal decomposition and feature extraction. Moreover, DWT reduces the length by half with each decomposition to the subsequent layer, whereas SWT maintains the same length at each layer, making it challenging to evaluate the length and unify the shapes of multi-modal signal features. Therefore, we adopt SWT as the main transformation for signals. SWT can be calculated as below:
W s ( a , b ) = A / 4 π · a 1 / 2 · φ ( a ω ) e n b ω ,
W a , b ( x ( n ) ) = n = 0 N x ( n ) W s ,
S W T 1 x n = W a , b ( x n ) L n , W a , b ( x n ) H n n N , S W T 2 x n = W a , b ( S W T 1 g ) L n , W a , b ( S W T 1 g ) H n n N , ,
where S W T 1 g = W a , b ( x n ) L n , S W T 2 g = W a , b ( S W T 1 g ) L n , W s ( a , b ) is given and proved by reference [23], L(n) is a low-pass filter, and H(n) is a high pass filter. With L(n) and H(n), SWT can obtain detailed and approximate information for each layer. In Figure 2, S W T 1 h = W a , b ( x ( n ) ) H ( n ) , S W T 2 h = W a , b ( S W T 1 g ) H ( n ) .
(3) Multiple synchrosqueezed wavelet transformations (MSWT): Utilizing SWT based on SFST, signals are sliced from long-term time domain features into multiple short-frequency domain features, transforming from the time domain to the time–frequency domain. Therefore, SWT can obtain more obvious RFF features and extract multiple-layer features from one signal. Each layer feature includes an approximation feature and a detail feature. However, during experiments, the RFF features of different modal signals exhibit varying performance in each layer, which poses a challenge in selecting the appropriate and general decomposition layer for multi-modal signals. One work [24] proposed the method of downsampling one dataset into several subsets to replace one single dataset, augmenting the datasets [25] and enhancing the pre-existing knowledge. However, this approach has limited impact on the improvement of the SEI system and requires significant computational resources for training the classifier. Based on the idea of multiple data enhancement, MSWT is proposed to integrate the details and approximations information from multiple SWT layers, effectively enhancing the RFF feature extraction and disregarding the disparate performances of different modal signals across various SWT layers. With it, feature extraction is simple and easy, but with strong and general features.
M S W T = M · [ S W T 1 x n . T , S W T 2 x n . T , , S W T L x n . T ] . T M R 1 × L M S W T = [ m 1 g , m 1 h , m 2 g , m 2 h , , m N g , m N h ] · [ S W T 1 g , S W T 1 h , , S W T N g , S W T N h ] . T ,
where N is the max decomposition layer of SWT for limitation, which is the same as the signal length, and m [ 0 , 1 ] represents the weight of each decomposition layer using g and h.
(4) RFF structure unifying: The MSWT is still a complex-valued feature with a real part and an imaginary part, making it challenging to unify the structures of different modal features. In this paper, we focus more on unifying the structures of different modal signals for SEI while maintaining the effectiveness of each modal RFF feature. To accomplish this, we incorporate FFT to further extract RFF features for enhanced frequency information [3] and apply a modulus operation to unify the structures. The MSWT can be redefined as follows:
MSWT m = r e ( f f t ( M S W T ) ) 2 + i m ( f f t ( M S W T ) ) 2 2 ,
where f f t is the fast Fourier transform, r e ( ) and i m ( ) are the real and imaginary parts of the complex-valued feature.
(5) Special particle swarm optimization (SPSO): M S W T m has successfully addressed the challenges of unifying and enhancing the effectiveness of RFF features from multiple modalities. However, the problem of determining the appropriate number and weights of layers remains unresolved due to the inefficiency of manual adjustment. In order to minimize parameter adjustment efforts and optimize the effectiveness of RFF features, it is necessary to implement an adaptive algorithm in M S W T m to determine and unify the best combination of layers and their respective weights for multi-modal signal features. This algorithm can adaptively achieve high RFF performance while using fewer layers and be applicable to all modal signals. To address this issue, we propose an SPSO, which is an adaptive selection algorithm for M S W T m based on PSO [26]. As in Figure 2, the best M S W T m requires appropriate layer decomposition and proper weights for each layer, including g and h. The SPSO selects these parameters; the details are as follows. (This paper does not seek to optimize the traditional selection algorithm, PSO, but only applies it to suit our RFF extractor).
1:
Our dataset has four modal signals and each mode has three devices. Each device introduces 20 labeled signals to join the SPSO selecting procedure. The details of the dataset are presented in Table 1;
2:
The initial l is set to 1 and SPSO reaches the best clustering results by selecting { m 1 g , m 1 h } , 0 m 1 g 1 , 0 m 1 h 1 . The features are clustered for different emitters, and the performance is evaluated by the mean distance between each point and its clustering center. The minimum mean distance is denoted as d s 1 , when l = 1 ;
3:
l = l + 1 , and SPSO selects { m 1 g , m 1 h , , m l g , m l h } , 0 m i g 1 , 0 m i h 1 , as above. The minimum mean distance is denoted as d s l ;
4:
Repeat step 3 until d s l > ( d s l 1 + d s l 2 + d s l 3 ) / 3 i f l > 3 , selecting l = l 3 ; i f l < 3 o r l = 3 (some modal signals cannot be decomposed to more than 3 layers); select l when obtaining the minimum d s l . The mean distance is calculated as follows:
d s = ( t = 1 r j = 1 f ( p t j c k j ) 2 2 ) / r ,
where r is the total signal number in clustering, f is the unifying feature length, and c k is the clustering center for p t .
(6) Energy balancing and multiple synchrosqueezed wavelet transformation of energy unified (MSWTEu): Combining the above steps, the RFF features of different modal signals can be unified into a unified and efficient structure. However, the unification of different modal RFF features is still inadequate, as the inherent properties and distributions are not fully unified. Moreover, the SWT energy explosion issue remains unresolved. Reference [13] proposed that energy regularization for combining SWT layers can enhance the RFF features. Furthermore, an energy-balancing strategy can unify the semantics of different modal signals and improve RFF, as energy accumulation and computation have always been important for signal processing and analysis in emitter identification. We expand upon our previous work [13] to accommodate multi-modal signals and simplify the energy-balancing algorithm, resulting in increased recognition speed. Subsequent experiments have compared the complexity and time cost between ISWTE [13] and MSWTEu. With energy balancing, the performance of MSWTEu also exhibits significant improvements in identification, as depicted in the following experiments. The implementation details of energy balancing by MSWTEu are as follows.
The total energy and average energy can be calculated:
g i ( w ) = S W T i g ( w ) = f f t ( m i g S W T i g ( n ) ) ,
h i ( w ) = S W T i h ( w ) = f f t ( m i h S W T i h ( n ) ) ,
g i ( w ) = S W T E i g ( w ) = w = 1 w = N g i 2 ( w ) j = 1 j = l w = 1 w = N g j 2 ( w ) ,
h i ( w ) = S W T E i h ( w ) = w = 1 w = N h i 2 ( w ) j = 1 j = l w = 1 w = N h j 2 ( w ) ,
E t o t a l g = j = 1 j = l w = 1 w = N g j 2 ( w ) ,
E t o t a l h = j = 1 j = l w = 1 w = N h j 2 ( w ) ,
E a v e g = j = 1 j = l w = 1 w = N g j 2 ( w ) / l ,
E a v e h = j = 1 j = l w = 1 w = N h j 2 ( w ) / l ,
where g i and h i represent the detailed and approximation information of each decomposition layer, respectively, and are unified in the structure according to (6). On the other hand, g i and h i are the relative weight values, which represent the i th weight value among the total g and h values across the l layers. By integrating these relative values into MSWTE and substituting MSWT with them, we can effectively attain the targeted result:
M S W T E = [ g 1 , h 1 g l , h l ] . T .
Based on MSWTE, MSWTEu is designed to unify and balance the energy of each layer. MSWTE is proposed to calculate the energy of each layer to a relative value, which effectively resolves the energy explosion problem. Moreover, MSWTEu proposes energy normalization based on the thought of normalization, which not only reduces the difference value between the maximum and minimum energies but also ensures a consistent energy across each layer.
M S W T E u = [ g 1 E a v e g E t o t a l g 2 , h 1 E a v e h E t o t a l h 2 , , g l E a v e g E t o t a l g 2 , h l E a v e h E t o t a l h 2 ] . T .
Based on the above, when confronted with multi-modal signals, the MSWTEu demonstrates higher efficiency, better adaptivity, and fewer limitations compared to other RFF feature extractors [7,8,13,27], which have also been verified by subsequent experiments. Unlike these extractors, which are only based on one typical signal mode, requiring extensive manual tuning or having high complexity, the MSWTEu offers a unified approach to extracting universal RFF features from different signal modes. This ultimately enables the SEI system to identify emitters based on a wider range of modal signals without the need for complex RFF extractors and additional classifiers.

3. RFF Feature Transferring—FTGAN

As depicted in Figure 3, there are four signal modes, and each mode can be transmitted by different emitters. Our proposed method can accurately identify the signal types and emitter IDs of all modal signals using only one modal signal, the SEI classifier. Specifically, we train an SEI classifier for the known modal RFF features, denoted as Mode 1, which has label and signal information. Then we utilize the network of classifier, excluding the softmax layer, as an identifying feature extraction network for all modal RFF features. Although the unknown modal (Mode 2/3/4) RFF features miss prior knowledge of labels and signals’ information, features extracted by the classifier network can be directly clustered into different emitters. With some modifications, the clustering can also be grouped into different signal modes. Some references [27,28] have also used clustering for evaluating and testing SEI performance. However, it is insufficient at extracting identifying features for all modal signals relying only on the network of one single modal SEI classifier. Therefore, we propose FTGAN to transfer the distribution of RFF features from unknown modal signals to match the distribution of the known modal signal.
There are three steps to complete the final clustering based on the above ideas: training the Mode 1 SEI classifier to identify feature extraction; training FTGAN to transfer unknown modal RFF features to match the known modal RFF features; combining the identifying feature extraction network and FTGAN to obtain identifying features from all modal signals to cluster emitters and signal modes.
Training the Mode 1 SEI classifier and obtaining the identifying feature extractor: To develop an SEI classifier for identifying emitters based on known modal signals, we utilize a ResNet model comprising three ResNet blocks and two dense layers. The ResNet model exhibits superior performance in managing gradient explosion compared to traditional convolutional networks. Commonly used networks for feature extraction, such as AlexNet, VGG, or GoogLeNet, have numerous layers, which might be excessive given the low dimensionality of signal features. Therefore, we adopt a more efficient network with fewer layers to build the SEI classifier, which also demonstrates excellent performance in classifying. Moreover, based on the classifier, we utilize its network up to the point before the softmax layer as the identifying feature extractor.
Training FTGAN and matching all modal RFF features: GANs have become popular in graph generation, but their architectures are not well-suited for waveform direct generation. The FTGAN has been inspired and upgraded based on GANs [13,22,29]. Figure 4 depicts the architectures of a normal GAN and FTGAN. Unlike traditional GAN designs, the FTGAN generator focuses on learning only the differences between different modes of signals. This approach allows for simpler networks and is more suitable for waveform calculation. Moreover, since the differences between different modes of signals have a lower information entropy compared to complete signals, the FTGAN is proposed to generate the feature difference between the features of known modal signals and unknown modal signals. Removing the feature differences from unknown modal signal features can achieve features matching with the known modal, which improves the accuracy and efficiency of waveform feature generation. The training procedure of FTGAN is as depicted in Figure 4: (1) There are two kinds of inputs, the known mode Mode 1 features denoted as x and the unknown mode Mode 2/3/4 features denoted as x . It is important to note that no labels are required during the training process of FTGAN; (2) x is fed into F T G A N g , which outputs the generated differences F T G A N g ( x ) ; (3) The unknown modal features by removing the differences obtain x F T G A N g ( x ) to match with the known modal features x; (4) The discriminator is improved by enhancing the inputs. Normally, traditional G A N d is to calculate the similarity between x and G A N g ( x ) to determine whether inputting samples are real or generated ‘real’. G A N g ( x ) is the generated ‘real’ samples by GAN’s generator. Then, the similarity is used to update and enhance the parameters of G A N g . The FTGAN is proposed to input ( x , x x ) and ( x F T G A N g ( x ) , F T G A N g ( x ) ) to push the discriminator and also utilize the enhanced discriminator to encourage the generator. These inputs can calculate the similarity between real samples and generated ‘real’ samples, as well as the similarity between real differences and generated ‘real’ differences; (5) Three loss functions are added to improve the performance of generators and enhance the realism of the generated features.
The entire process for clustering emitters and signal modes: With the above steps, we can easily recognize different emitters and signal modes with high accuracy by achieving robust features for all modal signals. The entire process can be summarized as follows: (1) the MSWTEu extracts RFF features and unifies the structures of features for all modal signals; (2) all modal RFF features transformed by FTGAN are in a distribution matching that of Mode 1, enabling the Mode 1 SEI classifier network to extract clear identifying features for all modal signals; (3) With the identifying features extracted by the trained Mode 1 classifier network, all modal signals can be clustered by the emitter identification or signal mode. The following are the FTGAN details of the description and handling. The entire process can also be seen in Algorithm 1.
Algorithm 1 The entire process for identifying emitters.
  • Step1: Train the Mode 1 SEI classifier and obtain the identifying feature extractor
  • Input: Known Mode 1 sample MSWTEu features x and labels
  • Output: Identifying feature extractor e x m o d e l
  • 1: Building Mode 1 classifier, as depicted in Figure 3
  • 2: Training Mode 1 classifier by x and labels, as depicted in Figure 3
  • 3: Obtaining e x m o d e l from the trained Mode 1 classifier, as depicted in Figure 3
  • Step2: Train FTGAN
  • Input: Known Mode 1 sample MSWTEu features, denoted as x, unknown Mode 2/3/4 sample MSWTEu features, denoted as x
  • Output: The “Mode 1” MSWTEu features x transformed from Mode 2/3/4 MSWTEu features x
  • for the number of training iterations do
  •     Sample minibatch of m Mode 1 samples  x 1 , , x m from D b
  •     Sample minibatch of m Mode 2/3/4 samples  x 1 , , x m from D t
  •     Calculate FTGAN g ( x ) and L FTGAN d by x, x , F T G A N g , and (20b)
  •      Update FTGAN d parameters
  •     Calculate FTGAN g ( x ) and L FTGAN g by x, x , F T G A N g , and (20a)
  •      Update FTGAN g parameters
  • end for
  • x = x F T G A N g ( x )
  • Step3: Identify Emitters
  • Input: Mode 1/2/3/4 sample MSWTEu features x m u l , emitter number N e m i t t e r
  • Output: SEI classifier c l m o d e l and emitter i d
  • 1: Load e x m o d e l
  • 2: Transferring multi-modal features to match Mode 1 features:
  •      x m u l = x m u l F T G A N g ( x m u l )
  • 3: Extracting identifying features: x m u l i = e x m o d e l ( x m u l )
  • 4: Building clustering model c l m o d e l as in Figure 3
  • 5: Unsupervised training c l m o d e l by x m u l i , N e m i t t e r
  • 6: Randomly selecting 10 samples from each cluster for labeling
  • 7: i d = c l m o d e l ( x m u l i )
The unlabeled samples from Mode 1 belong to the base domain, with their MSWTEu feature distribution represented as D b . The unlabeled samples from Modes 2/3/4 exist within the transferred domain, characterized by their MSWTEu feature distribution D t . The FTGAN aims to train D t D b , thereby minimizing and maximizing the adversarial objective function min G max D V ( D , G ) :
V ( D , G ) = E x [ l o g ( D ( x ) ] + E x [ l o g ( 1 D ( G ( x ) ) ) ] ,
where x D b , x D t , D is the discriminator function, and G is the generator function of FTGAN.
Normally, GAN’s [30] loss functions of D and G are calculated as follows:
L d = E x L b ( Y 0 , D ( x ) ) + E x L b ( Y 1 , D ( G ( x ) ) ) ,
L g = E x L b ( Y 0 , D ( x ) ) + E x L b ( Y 1 , D ( G ( x ) ) ) ,
where L d denotes ascending and L g denotes descending while training GANs; Y 0 ’s size is D ( x ) , filling with 0, and Y 1 ’s size is D ( G ( x ) ) , filling with 1; L b is the popular binary function for loss calculation. The improving details and loss functions of FTGAN are optimized as follows:
(1) Upgrading L d and L g : As depicted in the previous part, FTGAN is enhanced to be more robust by improving the inputs. Therefore, L d and L g are both upgraded:
L d = E x L b ( Y 0 , D ( x x x ) ) + E x L b ( Y 1 , D ( G ( x ) x G ( x ) ) ) ,
L g = E x L b ( Y 0 , D ( x x x ) ) + E x L b ( Y 1 , D ( G ( x ) x G ( x ) ) ) ,
where x is the known modal RFF feature, x is the unknown modal RFF feature, x x is the real RFF feature difference between the known mode and the unknown mode, G ( x ) is the generated feature difference, x G ( x ) is the generated known modal RFF feature, x x is the real difference; F T G A N d simultaneously calculates the similarity between ( x x , x ) and the real, and the similarity between ( G ( x ) , x G ( x ) ) and the fake; conversely, F T G A N g attempts to generate ‘real’ samples that cannot be discriminated by F T G A N d . By calculating losses on the combination of differences and samples, the G and D of FTGAN can be trained more robustly. The enhanced inputs could significantly improve the generation ability of ‘real’ differences by FTGAN, enabling the transfer of unknown modal features to known modal features.
(2) Transfer loss L t r a n s f e r : To ensure the accuracy of the generated differences between D t a n d D b , the L t r a n s f e r is added and the FTGAN is trained by descending L t r a n s f e r :
L t r a n s f e r = E x L m a e ( G ( x ) , x x ) + E x L m a e ( x G ( x ) , x ) ,
where L m a e is the mean absolute error (MAE) loss function, which performs better than the mean squared error (MSE) in the signal waveforms [13]. By decreasing the loss between the generated differences and real differences, and the loss between generated known modal features and real known modal features, L t r a n s f e r allows G to generate more reliable ‘real’ differences.
(3) Amplitude loss L a m p l i t u d e : To generate a more appropriate difference, the amplitude of generating value should be limited:
L o s s n m = i = 1 b m N m ( G ( x i ) ) / b m i = 1 b m N m ( x i x i ) / b m ,
L a m p l i t u d e = m a x Y 0 , L o s s n m ,
where N m ( ) is the binary regularization function, Y 0 ’s size is as L o s s n m , filling with 0, and b m is the batch size. Moreover, (18a) and (18b) allow the generated differences G ( x ) to fit the amplitude distribution of real differences x x .
(4) Smooth transition loss L t n : To smooth the MSWTEu features of the transferred signals, the FTGAN optimized it and descended it while training:
L t n = E x L m a e ( i = 1 b m G ( x i ) / b m G ( z ) ) ,
where x e p o c h = ( x 1 , x 2 , , x b m ) is a sample set of one epoch of x , and z x e p o c h . The loss function is added to improve the distribution balance of different modal features.
(5) The FTGAN’s total loss function: Above all, combining the added three loss functions and improved G and D functions, the final loss functions of FTGAN are as follows:
L F T G A N g = α L g + β L t r a n s f e r + L a m p l i t u d e + L t n ,
L F T G A N d = L d ,
where α = 0.5 , and β = 10 .

4. Experiments and Discussion

4.1. Data Acquisition and Experimental Setup

Dataset description: Four modal communication signals, ads-b, Wi-Fi, and replied signals rm1 and rm2, were utilized in the experiments. Each modal signal was transmitted by three emitters. Table 1 provides details about the data employed. Ads-b signals were collected from flying airplanes over a period of 12 months, and pre-processed by us to an SNR of 10 dB. The collector used was the Tektronix RSA6120B real-time spectrum analyzer. At a frequency of 1090 MHz, approximately 5 modal signals were collected from the air acquisition. We screened ads-b signals from them and labeled them using the airplane’s International Civil Aviation Organization (ICAO) address. ICAO serves as the only global identifier [15,16], and can be obtained by decoding through our Simulink codes. Wi-Fi signals were obtained from a public dataset. The use of ads-b and Wi-Fi signals has been observed in previous SEI references [15,16,21,27]. Communication reply signals—rm1 and rm2—are two modal communication signals that use pulse position modulation (PPM). Rm1 and rm2 are similar to ads-b and are used to reply to interrogating signals in air traffic control, which are the ads-b a mode and ads-b c mode, respectively. Traditionally, the ads-b signal is broadcasted and easily received. The Ads-b a/c mode carries little information, and does not carry the airplane’s ICAO address; it is sent at specific time points. Therefore, we cannot receive enough signals with the a/c mode to label them by their ICAO. The ads-b s mode, a mode, and c mode employ different data links and do not share the same signal transmitters in the same airplanes. Therefore, we can simulate the a/c mode by other emitters, which are called rm1 and rm2 in this paper. Based on ads-b, rm1, and rm2, we can completely simulate CATCRBS and conduct SEI experiments for CATCRBS. We can obtain the identification of airplanes by SEI, ignoring the effects of different signal modes and emitters. With this, our SEI system can completely replace the identification-decoding module within CATCRBS. To explore the effects of different modal signals transmitted by the same emitters, rm1 and rm2 are generated by the same three Tektronix AWG 5000 series arbitrary waveform generators. These signals are received by the Teledyne LeCroy WaveRunner 9404 oscilloscope and used to build the simulated dataset.
Ads-b, rm1, and rm2 have the same modulation PPM, but with different structures. We utilize their preamble pulses for SEI. The preamble pulses of communication signals without any modulated information are more efficient and less susceptible to deception when identifying emitters. With the same modulation, the proposed MSWTEu can also recognize the signal modes, as shown in later experiments. Wi-Fi signals consist of continuous pulses, and we adopt each pulse as a single signal for SEI.
Training procedure: Firstly, RFF features are extracted by MSWTEu from each sample. Non-Wi-Fi signals select preamble pulses, while Wi-Fi signals select one pulse. Secondly, the FTGAN is trained with 3000 unlabeled ads-b MSWTEu (the known mode) and 9000 unlabeled non-adsb MSWTEu (the unknown modes). Thirdly, the Mode 1 SEI classifier is trained with 900 labeled ads-b MSWTEu, as shown in Figure 3. Fourthly, with FTGAN, MSWTEu features of all modes are transferred to match the Mode 1 distribution, and all layers, except for the softmax of the Mode 1 classifier, are used to extract identifying features for all modes. Finally, a clustering model is employed to recognize these 12 emitters based on the identifying features without pre-recognizing signal types and supervised training. If signal mode recognition is required, the clustering model can also be utilized for that purpose.
(Note: Although rm1-1∼3 and rm2-1∼3 share the same emitters, the clustering process still divides them into 6 clusters instead of 3. In other words, in our experiments, the final clustering number for SEI is 12. There are two main reasons for this: (1) Different modal signals from the same wireless devices may actually be from different emitters, such as the actual ads-b a/c/s modes. (2) Clustering the same emitter with different modal signals into different clusters can help the SEI system simultaneously identify signal modes and sources. For a convenient description, in the later experiments, adsb-1, adsb-2, adsb-3, wifi-1, wifi-2. wifi-3, rm1-1, rm1-2, rm1-3, rm2-1, rm2-2, and rm2-2 are used as separate transmitter IDs, representing individual emitters, respectively).
However, for a clear classification, 10 labeled samples per cluster are required for label correction. It is worth mentioning that signal labeling should be corrected in cross-domain exploring, which is different from image cross-domain exploration [22,31], while maintaining a similar inherent structure. The cross-domain exploration problem in signal analysis also exists in other areas without any similar inherent properties, such as the conversion of texts to graphs. With proper label correction, the proposed structure of MSWTEu+FTGAN can significantly improve the accuracy of SEI. During the experiments, 300 samples were selected for evaluation by each emitter, with a total of 3600 samples evaluated (4 modes × 3 emitters × 300).

4.2. Results and Discussions

Extractors for identifying devices: The first experiment involves confirming the unifying feature performance and improving RFF features by MSWTEu. Figure 5 shows that the MSWT features have achieved structure unification for different kinds of signals. Moreover, Figure 5 depicts that the MSWTEu has clear improvement in limiting noises and interferences. The red dashed box in Figure 5 shows that the noises have been well inhibited by MSWTEu. Figure 6 shows the MSWTEu’s clustering performance in identifying 3 emitters under ads-b signals with different SNRs, which is obviously better than MSWT without energy unifying under lower SNRs. Moreover, the performances of our proposed methods, which use SPSO for adaptive parameter determination, are better than ISWTE [13], which is the previous RFF feature extractor we propose to counteract heavy noise. The ISWTE is manually adjusted based on limited decomposing layers. Figure 7 depicts the boxplot performance for identifying 12 emitters. MSWTEu performs far better than MSWT in the different modal signals for SEI, although MSWT is only a bit worse than MSWTEu at identifying three emitters using the same modal signal. Above all, MSWT succeeded at unifying RFF features and enhancing SEI performance, but did not show more robustness in identification across multi-modal signals. Furthermore, MSWTEu achieved better performance in SEI across multi-modal signals based on MSWT.
Complexity of RFF extractors: We compared the complexity of our MSWT and MSWTEu with MAp and ISWTE, which are RFF extractors that were introduced in recent references [13,27]. Table 2 illustrates the comparative results, where M represents MSWT and MEu represents MSWTEu. The ‘time’ in Table 2 represents the time taken by these extractors to handle 3000 signals when N = 1024 . (N is the length of multi-modal signals after unification). M is completely based on wavelet functions with a time complexity of O ( N l o g N ) , and is slightly smaller than that of MEu. MAp in [27] claims to have a complexity of O ( N 2 ) , but its actual complexity is closer to O ( N 3 ) , which requires more time for feature handling, as shown in Table 2. Although MEu has a higher time complexity than M, the additional time cost is only 0.1 s. Moreover, they share the same space complexity. Compared to ISWTE, MEu demonstrates a faster processing speed despite having the same time complexity. Our MSWTEu has obvious advantages in time and space complexity when compared to many RFF extraction algorithms in signal processing. Although MEu is a little worse than M in complexity, the identifying performances in multi-modal signals or under low SNR have obvious advantages, as shown in Figure 6 and Figure 7. Moreover, the FTGAN output dimensionality is 20, further contributing to the smaller space complexity in the final clustering.
Open-set signal mode identification by RFF extractors: To investigate the generality of these RFF extractors across multi-modal signals, we conducted experiments using M, MEu, and MAp to handle the four modal signals. If these features are extracted effectively and the multi-modal features are aligned properly, clustering for signal modes based on these features can also be effectively achieved. Therefore, we evaluated their performances in multi-modal signals by measuring the accuracy of mode identification through clustering. We introduce MAp since [27] proposed an RFF feature extractor similar to ours, which can also achieve emitter identification through clustering. However, all the features of rm1 calculated by MAp are zero, indicating that MAp is unsuitable for rm1 processing. Apart from rm1, the average accuracy of MAp is 50% in regard to recognizing signal modes. In contrast, MSWT achieves an accuracy of 92.6%, and MSWTEu achieves 98.1%. Figure 8 depicts the accuracy comparison concerning the recognition of signal modes across different methods. Figure 9 presents the accuracy confusion matrix for various methods in recognizing signal modes, using the same clustering algorithm. The MSWTEu demonstrates exceptional generalizability to different modal signals and performs consistently well in feature extraction, exhibiting the highest accuracy and most stable performance in signal mode clustering.
SEI performance comparison by MSWT, MSWTEu, and MSWTEu+FTGAN:Table 3 illustrates the improvement of each proposed method in identifying emitters. In the experiments, k-means is used for clustering. M uses MSWT+k-means and MEu uses MSWTEu+k-means to classify emitters. MEu+ FTGAN is as depicted in Figure 3 for classification. When the classified number is 3 in Table 3, k-means is only used to identify 3 emitters for the same modal signal. When the classified number is 12 in Table 3, the k-means is utilized to directly distinguish 12 emitters for all modal signals. MSWTEu performs significantly better than MSWT in classifying emitters for the same modal signal but FTGAN does not improve the accuracy for the same modal signal. In the comprehensive recognition of twelve emitters by different modal signals, MSWT only achieves 69% accuracy, and rm2-1∼3 is almost not recognized, which is also shown in Table 3 and Figure 7a. MSWTEu improved by 26% while recognizing rm2-1∼3. However, relying solely on feature extraction, the integrated performance in identifying twelve emitters through different modal signals is still not enough. Although FTGAN is trained unsupervised and no extra labeling work is required, the MSWTEu+FTGAN has nearly 20% improvement compared with MSWT, achieving 86% comprehensive accuracy. In conclusion, MSWTEu offers an improvement in SEI for both the same modal signals and different modal signals when compared to MSWT. This demonstrates the effectiveness of the energy-balancing strategy for the RFF feature extractor. The SEI effect by FTGAN is not improved in the same modal signal compared to MSWTEu since the FTGAN is optimized based on transferring features for multi-modal signals. However, the features transferred by FTGAN show significant improvement in SEI with different modal signals compared to MSWTEu, proving that the multi-modal signal features have been successfully transferred to a matching distribution by the FTGAN. Figure 10 also shows the comparisons between MSWT, MSWTEu, and MSWTEu+FTGAN in terms of recognizing emitters with the same modal signals and multi-modal signals. Figure 10 shows the processing time and accuracy recognition among different methods. In Figure 10a, although the processing time of MSWTEu is longer than that of MSWT, the average accuracy recognition is clearly higher than that of MSWT, confirming that energy balance has an effect on feature extraction but at some cost. Moreover, in Figure 10b, FTGAN shows a clear improvement in recognizing emitters across different signal modes, which confirms that FTGAN successfully matches the multi-modal signal features but with the same time costs.
The comprehensive SEI performance:Table 4 depicts the comparison performances of different methods in identifying 12 devices by multi-modal signals. We introduce MAp [27] to compare with our MSWT and MSWTEu. MAp is an RFF extractor and utilizes clustering for identification. MAp is designed for Wi-Fi [27], but it is not suitable for other modal signals. As shown in Table 4, it fails to identify emitters by non-Wi-Fi signals and achieves only 35% average accuracy in SEI by multi-modal signals. DTN is a representative cross-domain method proposed by [22]. In DTN, ads-b is also the known mode. Through training and fine-tuning via multi-modal signal features extracted by MSWTEu, DTN only has 63% accuracy, which is even lower than MSWT (DTN labels have been corrected). The datasets in [22] consist of multi-modal graphs. Although different modal graphs have different modal features, they still share inherent features for identifying objects. Therefore, our experiments show that the method in [22] does not achieve the desired cross-domain SEI in the signal domain. TFL [21] trains a base classifier for one modal signal and then fine-tunes the last few layers to fit other modal signals. Table 4 also suggests that TFL does not perform well in SEI, confronting multi-modal signals from different emitters. In [21], TFL attempts to identify emitters across domains, from ads-b to Wi-Fi. However, the final accuracy in the experiment of [21] is also unsatisfactory, coming across more as an exploratory discussion. Due to the unique wave structure of signals, our work proposes the extraction of a unified and matched feature set for different modal signals, using clustering for emitter identification. Although the clustering needs 10 samples for label correction, the TFL needs more labeled samples for fine-tuning but still has worse results. Moreover, by TFL, a new modal signal requires a new, fine-tuned classifier, which wastes more time and resources compared to the proposed MSWTEu+FTGAN. Figure 11 depicts the comparison curves of the different methods used in identifying 12 devices through multi-modal signals. The red line represents our method, which outperforms all other methods in all modal signals.
In conclusion, identifying wireless devices via cross-modal signals may not be suitable for traditional cross-domain and transfer learning. These methods perform poorly and also require label correction in SEI by cross-modal signals. Therefore, by obtaining a robust RFF distribution across multi-modal signal domains, cross-modal SEI may be better suited for clustering. Moreover, by using a few labels for label correction, we can achieve a higher SEI accuracy. Our MSWTEu and FTGAN unify the RFF features and match the final feature distributions for multi-modal signals, leading to better results in SEI through clustering and significant progress in cross-modal SEI. In Figure 12, we present an overview of the final performance through the confusion matrix for 12 emitters across 4 modal signals. We can see that different modes have been effectively clustered, making it hard for the same mode to be clustered into other modes. Moreover, the comprehensive accuracy of emitter identification aligns with our projections.

5. Conclusions

In this paper, we addressed the challenge of how an SEI system can identify wireless devices across multi-modal signals using only a single RFF extractor and a single SEI classifier. In typical scenarios that encounter multi-modal signals, no prior works have efficiently proposed an SEI system for identifying emitters without the need for different extractors and classifiers for each modal signal. Specifically, the enhanced SEI system with MSWTEu and FTGAN that we propose addresses the above issues. Confronting multi-modal signals, the MSWTEu unifies the features first, and further optimizes feature energy calculation to improve the comprehensive RFF performance. Moreover, with FTGAN and the proposed training process, the MSWTEu of different modal signals can be transferred to a matching distribution, unifying the final identifying multi-modal signal features for SEI and requiring only one unsupervised clustering for the identifying emitters. Experimental results, evaluated on a mixture of simulated, public, and real received datasets, show that MSWTEu, FTGAN, and the proposed training strategy can significantly improve the SEI performance and save resources when dealing with multiple types of signals. Solely using feature tackling, MSWTEu can reach 87% accuracy under the same modal signals in unsupervised clustering for identifying emitters, 75% accuracy under different modal signals in unsupervised clustering for identifying emitters, and 98.1% accuracy in unsupervised clustering for identifying signal modes. The results are far better compared to existing SEI studies that pay attention to feature tackling. With the inclusion of FTGAN and our proposed training approach, the feature performance in cross-modal emitter identification is further enhanced, rising from 75% to 86% in clustering. In conclusion, we believe that further improvements in feature optimization and matching for multi-modal signals and utilizing these improved features through unsupervised clustering to identify emitters have great potential in improving the comprehensive performance in crossing-modal SEI and reducing manual work, such as labeling.
However, there are some limitations in our work: (1) the simple clustering algorithm used for identifying emitters has limited adaptability when confronted with a constantly changing number of emitters; (2) the proposed methods cannot fully resolve the identification problem for new modal signals that lack labels. This issue still requires some manual labeling for correction; (3) the same emitters with different signal modes cannot be clustered into the same class. This can be seen in the boxplot of Figure 7b in rm1 and rm2, as the rm1 and rm2 modes share the same emitters. Future work on SEI cross-multiple modal signals should be further improved (1) to research a proper dynamic clustering algorithm to replace the simple clustering in our SEI system and solve limitation 1; (2) to research a feature extraction and matching method that is more adaptable and robust, to reduce limitation 2 and resolve limitation (3).

Author Contributions

Data curation, M.Z. and Y.Y.; writing—original draft preparation, M.Z.; writing—review and editing, H.Y. and M.Z.; visualization, M.Z.; model design, M.Z.; validation, M.Z., Y.Y., H.L. and Y.H.; funding acquisition, H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (grant No. U20A20161).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Xia, Z.X.; Wei, Z.; Zhang, H. Review on Security Issues and Applications of Trust Mechanism in Wireless Sensor Networks. Comput. Intell. Neurosci. 2022, 2022, 3449428. [Google Scholar] [CrossRef]
  2. Hall, J.; Barbeau, M.; Kranakis, E. Detection of transient in radio frequency fingerprinting using signal phase. Wirel. Opt. Commun. 2003, 9, 2519–2531. [Google Scholar]
  3. Kennedy, I.O.; Scanlon, P.; Mullany, F.J.; Buddhikot, M.M.; Nolan, K.E.; Rondeau, T.W. Radio Transmitter Fingerprinting: A Steady State Frequency Domain Approach. In Proceedings of the 2008 IEEE 68th Vehicular Technology Conference, Calgary, AB, Canada, 21–24 September 2008; pp. 1–5. [Google Scholar]
  4. Defense Advanced Research Projects Agency (DARPA). Radio Frequency Machine Learning Systems (RFMLS) (HR001117S0043) [EB/OL]. Available online: http://www.fbo.gov/ (accessed on 11 August 2017).
  5. Bremnes, K.; Moen, R.; Yeduri, S.R.; Yakkati, R.R.; Cenkeramaddi, L.R. Classification of UAVs Utilizing Fixed Boundary Empirical Wavelet Sub-Bands of RF Fingerprints and Deep Convolutional Neural Network. IEEE Sens. J. 2022, 22, 21248–21256. [Google Scholar] [CrossRef]
  6. Nie, W.; Han, Z.-C.; Zhou, M.; Xie, L.-B.; Jiang, Q. UAV Detection and Identification Based on WiFi Signal and RF Fingerprint. IEEE Sens. J. 2021, 21, 13540–13550. [Google Scholar] [CrossRef]
  7. Reising, D.; Cancelleri, J.; Loveless, T.D.; Kandah, F.; Skjellum, A. Radio Identity Verification-Based IoT Security Using RF-DNA Fingerprints and SVM. IEEE Internet Things J. 2021, 8, 8356–8371. [Google Scholar] [CrossRef]
  8. Huang, D.Y.; Hou, C.P.; Yang, Y.; Lang, Y.; Wang, Q. Micro-Doppler Spectrogram Denoising Based on Generative Adversarial Network. In Proceedings of the 48th European Microwave Conference (EuMC), Madrid, Spain, 23–27 September 2018; pp. 909–912. [Google Scholar]
  9. Uzundurukan, E.; Ali, A.M.; Dalveren, Y.; Kara, A. Performance Analysis of Modular RF Front End for RF Fingerprinting of Bluetooth Devices. Wirel. Pers. Commun. 2020, 112, 2519–2531. [Google Scholar] [CrossRef]
  10. Zhou, X.Y.; Hu, A.Q.; Li, G.Y.; Peng, L.N.; Xing, Y.X.; Yu, J.B. A Robust Radio-Frequency Fingerprint Extraction Scheme for Practical Device Recognition. IEEE Internet Things J. 2021, 8, 11276–11289. [Google Scholar] [CrossRef]
  11. Agadakos, I.; Agadakos, N.; Polakis, J.; Amer, M.R. Chameleons’ Oblivion: Complex-Valued Deep Neural Networks for Protocol-Agnostic RF Device Fingerprinting. In Proceedings of the 5th IEEE European Symposium on Security and Privacy (IEEE Euro S and P), Electr Network, Genova, Italy, 16–18 June 2020; pp. 322–338. [Google Scholar]
  12. Wang, Y.; Gui, G.; Gacanin, H.; Ohtsuki, T.; Dobre, O.A.; Poor, H.V. An Efficient Specific Emitter Identification Method Based on ComplexValued Neural Networks and Network Compression. IEEE J. Sel. Areas Commun. 2021, 39, 2305–2317. [Google Scholar] [CrossRef]
  13. Zeng, M.; Wang, Z.; Liu, H.; Li, Y.; Yang, H. An Adaptive Specific Emitter Identification System for Dynamic Noise Domain. IEEE Internet Things J. 2022, 9, 25117–25135. [Google Scholar] [CrossRef]
  14. Zhu, Z.; Yi, Z.; Li, S.; Li, L. Deep Muti-Modal Generic Representation Auxiliary Learning Networks for End-to-End Radar Emitter Classification. Aerospace 2022, 9, 732. [Google Scholar] [CrossRef]
  15. Leonardi, M.; Gerardi, F. Aircraft Mode S Transponder Fingerprinting for Intrusion Detection. Aerospace 2020, 7, 30. [Google Scholar] [CrossRef]
  16. Nicolussi, A.; Tanner, S.; Wattenhofer, R. Aircraft Fingerprinting Using Deep Learning. In Proceedings of the 28th European Signal Processing Conference (EUSIPCO), Electr Network, Dublin, Ireland, 23–27 August 2021; pp. 740–744. [Google Scholar]
  17. Leonardi, M.; Di Gregorio, L.; Di Fausto, D. Air Traffic Security: Aircraft Classification Using ADS-B Message’s Phase-Pattern. Aerospace 2017, 4, 51. [Google Scholar] [CrossRef]
  18. Al-Shawabka, A.; Restuccia, F.; D’Oro, S.; Jian, T.; Rendon, B.C.; Soltani, N.; Dy, J.; Ioannidis, S.; Chowdhury, K.; Melodia, T. Exposing the Fingerprint: Dissecting the Impact of the Wireless Channel on Radio Fingerprinting. In Proceedings of the IEEE INFOCOM 2020—IEEE Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020; pp. 646–655. [Google Scholar]
  19. Mkadem, F.; Boumaiza, S. Physically Inspired Neural Network Model for RF Power Amplifier Behavioral Modeling and Digital Predistortion. IEEE Trans. Microw. Theory Tech. 2011, 59, 913–923. [Google Scholar] [CrossRef]
  20. Sa, K.; Lang, D.; Wang, C.; Bai, Y. Specific Emitter Identification Techniques for the Internet of Things. IEEE Access 2020, 8, 1644–1652. [Google Scholar] [CrossRef]
  21. Kuzdeba, S.; Robinson, J.; Carmack, J. Transfer Learning with Radio Frequency Signals. In Proceedings of the 2021 IEEE 18th Annual Consumer Communications & Networking Conference (CCNC), Las Vegas, NV, USA, 9–12 January 2021; pp. 1–9. [Google Scholar]
  22. Tang, H.; Jia, K. Discriminative adversarial domain adaptation. In Proceedings of the 34th AAAI Conference on Artificial Intelligence, New York, NY, USA, 3 April 2020; pp. 5940–5947. [Google Scholar]
  23. Daubechies, I.; Lu, J.F.; Wu, H.T. Synchrosqueezed wavelet transforms: An empirical mode decomposition-like tool. Appl. Comput. Harmon. Anal. 2011, 30, 243–261. [Google Scholar] [CrossRef]
  24. Yu, J.B.; Hu, A.Q.; Li, G.Y.; Peng, L.N. A Robust RF Fingerprinting Approach Using Multisampling Convolutional Neural Network. IEEE Internet Things J. 2019, 6, 6786–6799. [Google Scholar] [CrossRef]
  25. Mo, H.; Kim, S. A Deep Learning-Based Human Identification System with Wi-Fi CSI Data Augmentation. IEEE Access 2021, 9, 91913–91920. [Google Scholar] [CrossRef]
  26. Xue, B.; Zhang, M.; Browne, W.N. Particle Swarm Optimization for Feature Selection in Classification: A Multi-Objective Approach. IEEE Trans. Cybern. 2013, 43, 1656–1671. [Google Scholar] [CrossRef]
  27. Sun, L.T.; Wang, X.; Yang, A.F.; Huang, Z.T. Radio Frequency Fingerprint Extraction Based on Multi-Dimension Approximate Entropy. IEEE Signal Process. Lett. 2020, 27, 471–475. [Google Scholar] [CrossRef]
  28. Xie, C.; Zhang, L.; Zhong, Z. Few-Shot Unsupervised Specific Emitter Identification Based on Density Peak Clustering Algorithm and Meta-Learning. IEEE Sens. J. 2022, 22, 18008–18020. [Google Scholar] [CrossRef]
  29. Gong, J.L.; Xu, X.D.; Lei, Y.K. Unsupervised Specific Emitter Identification Method Using Radio-Frequency Fingerprint Embedded InfoGAN. IEEE Trans. Inf. Forensics Secur. 2020, 15, 2898–2913. [Google Scholar] [CrossRef]
  30. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. In Proceedings of the 28th Conference on Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014; pp. 2672–2680. [Google Scholar]
  31. Zhang, H.; Li, A.; Han, X.; Chen, Z.; Zhang, Y.; Guo, Y. Improving open set domain adaptation using image-to-image translation. In Proceedings of the IEEE International Conference on Multimedia and Expo (ICME), Shanghai, China, 8–12 July 2019; pp. 1258–1263. [Google Scholar]
Figure 1. The process of a traditional SEI system for multi-modal signals and our SEI system for multi-modal signals. Our system aims to unify the extractor and classifier, which improves the efficiency of the system. With this improvement, the system decreases labeling work.
Figure 1. The process of a traditional SEI system for multi-modal signals and our SEI system for multi-modal signals. Our system aims to unify the extractor and classifier, which improves the efficiency of the system. With this improvement, the system decreases labeling work.
Sensors 23 08576 g001
Figure 2. The MSWTEu process. The first column waveform is the time-domain wave; in the second to third columns, the blue wave represents S W T i g , and the red wave represents S W T i h .
Figure 2. The MSWTEu process. The first column waveform is the time-domain wave; in the second to third columns, the blue wave represents S W T i g , and the red wave represents S W T i h .
Sensors 23 08576 g002
Figure 3. The emitter label identification procedure for a single mode (Mode 1) and four different modes (Mode 1/2/3/4). The top right part is the SEI classifier trained for Mode 1 under supervision, which outputs the emitter ID. The bottom right is our multi-modal SEI system, which recognizes the emitter ID for all modal signals. Our multi-modal SEI system uses FTGAN to transfer different mode features to Mode 1 distribution and shares parameters with the Mode 1 SEI classifier.
Figure 3. The emitter label identification procedure for a single mode (Mode 1) and four different modes (Mode 1/2/3/4). The top right part is the SEI classifier trained for Mode 1 under supervision, which outputs the emitter ID. The bottom right is our multi-modal SEI system, which recognizes the emitter ID for all modal signals. Our multi-modal SEI system uses FTGAN to transfer different mode features to Mode 1 distribution and shares parameters with the Mode 1 SEI classifier.
Sensors 23 08576 g003
Figure 4. The architectural comparison between normal GAN and FTGAN.
Figure 4. The architectural comparison between normal GAN and FTGAN.
Sensors 23 08576 g004
Figure 5. The waveform feature comparison between MSWT and MSWTEu; (a,c,e,g) are the MSWT waveform features of rm1, wifi, ads-b, and rm2; (b,d,f,h) are the MSWTEu waveform features of rm1, wifi, ads-b, and rm2; the red boxes are noises not decreased by the MSWT, which do not exist in the MSWTEu waveform features; the horizontal axis displays the frequency and the vertical axis displays the amplitude.
Figure 5. The waveform feature comparison between MSWT and MSWTEu; (a,c,e,g) are the MSWT waveform features of rm1, wifi, ads-b, and rm2; (b,d,f,h) are the MSWTEu waveform features of rm1, wifi, ads-b, and rm2; the red boxes are noises not decreased by the MSWT, which do not exist in the MSWTEu waveform features; the horizontal axis displays the frequency and the vertical axis displays the amplitude.
Sensors 23 08576 g005
Figure 6. The SEI accuracy comparison between MSWTEu, MSWTE, and ISWTE in identifying emitters by ads-b signals under different SNRs.
Figure 6. The SEI accuracy comparison between MSWTEu, MSWTE, and ISWTE in identifying emitters by ads-b signals under different SNRs.
Sensors 23 08576 g006
Figure 7. Boxplot representation of MSWT (a) and MSWTEu (b) for 12 devices; the values of the same red boxes show a similarity which crosses different modes in (a) but not crosses different modes in (b); the horizontal axis represents the signal mode while the vertical axis indicates the value of each feature. For visualization purposes, these features have been reduced to 1 dimension using T-distributed stochastic neighbor embedding (TSNE).
Figure 7. Boxplot representation of MSWT (a) and MSWTEu (b) for 12 devices; the values of the same red boxes show a similarity which crosses different modes in (a) but not crosses different modes in (b); the horizontal axis represents the signal mode while the vertical axis indicates the value of each feature. For visualization purposes, these features have been reduced to 1 dimension using T-distributed stochastic neighbor embedding (TSNE).
Sensors 23 08576 g007
Figure 8. The accuracy comparison for various methods in recognizing signal modes.
Figure 8. The accuracy comparison for various methods in recognizing signal modes.
Sensors 23 08576 g008
Figure 9. The accuracy confusion matrix for various methods in recognizing signal modes, (a) is MAp, (b) is MSWT, (c) is MSWTEu, and the darker color represents the higher accuracy.
Figure 9. The accuracy confusion matrix for various methods in recognizing signal modes, (a) is MAp, (b) is MSWT, (c) is MSWTEu, and the darker color represents the higher accuracy.
Sensors 23 08576 g009
Figure 10. MSWT and MSWTEu performances in recognizing 3 emitters for the same mode, (a) and MSWTEu and MSWTEu+FTGAN performances in recognizing 12 emitters for multiple modes (b).
Figure 10. MSWT and MSWTEu performances in recognizing 3 emitters for the same mode, (a) and MSWTEu and MSWTEu+FTGAN performances in recognizing 12 emitters for multiple modes (b).
Sensors 23 08576 g010
Figure 11. The SEI accuracy through multi-modal signals based on different methods.
Figure 11. The SEI accuracy through multi-modal signals based on different methods.
Sensors 23 08576 g011
Figure 12. The comprehensive SEI accuracy confusion matrix through multi-modal signals based on MSWTEu+FTGAN, and the darker color represents the higher accuracy.
Figure 12. The comprehensive SEI accuracy confusion matrix through multi-modal signals based on MSWTEu+FTGAN, and the darker color represents the higher accuracy.
Sensors 23 08576 g012
Table 1. Details of experimental signals.
Table 1. Details of experimental signals.
Signal ModeTransmitter
ID
DetailsTraining
ads-b
(ads-b s mode)
adsb-1
adsb-2
adsb-3
Aircraft communication signals, collected from
real aircraft by a real-time spectrum analyzer
over a 12-month period;
The preamble pulses are for SEI, including
4 pulses for each signal;
f c = 1090  MHz, f s = 150  MHz, PPM
Modulation, SNR = 10 db.
3000 unlabeled samples in FTGAN
training as the known mode;
900 labeled samples in
classifier training;
900 labeled samples in
classifier evaluation.
Wi-Fiwifi-1
wifi-2
wifi-3
Public dataset [18] part 3 raw data; wifi-1∼3 are
from 3 antennas of the same device;
One transmitted pulse is for SEI per one signal;
IEEE 802.11 a / g , 2.4 GHz, 20 MS/s, BPSK
Modulation, SNR = 5 db.
9000 unlabeled samples
(3000 per 1 mode) in FTGAN
training as the unknown modes;
2700 labeled samples
(300 per 1 mode per 1 emitter)
(Wi-Fi, rm1, rm2) in
classifier evaluation.
rm1 and rm2
(ads-b a/c
mode)
rm1-1
rm1-2
rm1-3
rm2-1
rm2-2
rm2-3
Aircraft communication replying signals,
generated by three waveform generators
and collected by an oscilloscope;
rm1-1∼3 and rm2-1∼3 share the same generators;
The preamble pulses are for SEI, including
2 pulses for each signal; f c = 1030  MHz,
f s = 1030  MHz, PPM modulation, SNR = 10 db.
Table 2. The complexity comparison between different RFF extractors.
Table 2. The complexity comparison between different RFF extractors.
MethodsTime ComplexityTimeDimension
M O ( N l o g N ) 0.9 s 6 N
MEu O ( N 2 ) 1 s 6 N
ISWTE O ( N 2 ) 10 s 6 N
MAp O ( N 3 ) 1 d4
MEu+FTGAN20
Table 3. SEI accuracy of MSWT, MSWTEu, and MSWTEu+FTGAN confronting one modal signal and multi-modal signals.
Table 3. SEI accuracy of MSWT, MSWTEu, and MSWTEu+FTGAN confronting one modal signal and multi-modal signals.
Classify
Number
3 Emitters under the Same Modal Signals12 Emitters under 4 Modal Signals
Device
ID/Method
MMEuMEu+FTGANMMEuMEu+FTGAN
rm1-1∼30.750.990.8650.670.680.82
wifi-1∼30.970.960.8150.770.740.83
adsb-1∼30.670.7720.990.970.961
rm2-1∼30.410.7720.780.340.60.79
Avg 30.700.870.865///
Avg 12///0.690.750.86
Table 4. SEI accuracies of different methods confronting multi-modal signals.
Table 4. SEI accuracies of different methods confronting multi-modal signals.
Methodrm1-1∼3wifi-1∼3adsb-1∼3rm2-1∼3Avg 12
M0.670.770.970.340.68
MEu0.680.740.9640.60.75
MAp0.290.420.340.360.35
MEu+DTN0.590.420.950.550.63
MEu+TFL0.330.610.860.560.59
MEu+FTGAN0.820.831.00.790.86
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zeng, M.; Yao, Y.; Liu, H.; Hu, Y.; Yang, H. A Specific Emitter Identification System Design for Crossing Signal Modes in the Air Traffic Control Radar Beacon System and Wireless Devices. Sensors 2023, 23, 8576. https://doi.org/10.3390/s23208576

AMA Style

Zeng M, Yao Y, Liu H, Hu Y, Yang H. A Specific Emitter Identification System Design for Crossing Signal Modes in the Air Traffic Control Radar Beacon System and Wireless Devices. Sensors. 2023; 23(20):8576. https://doi.org/10.3390/s23208576

Chicago/Turabian Style

Zeng, Miyi, Yue Yao, Hong Liu, Youzhang Hu, and Hongyu Yang. 2023. "A Specific Emitter Identification System Design for Crossing Signal Modes in the Air Traffic Control Radar Beacon System and Wireless Devices" Sensors 23, no. 20: 8576. https://doi.org/10.3390/s23208576

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop