Next Article in Journal
A Waveform Dataset in Continuous Mode of the Montefeltro Seismic Network (MF) in Central-Northern Italy from 2018 to 2020
Previous Article in Journal
Database of Metagenomes of Sediments from Estuarine Aquaculture Farms in Portugal—AquaRAM Project Collection
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Data Descriptor

Spectrogram Data Set for Deep-Learning-Based RF Frame Detection

1
Division Engineering of Adaptive Systems EAS, Fraunhofer Institute for Integrated Circuits, 01187 Dresden, Germany
2
Department of CSE, LNM Institute of Information Technology, Jaipur 302031, India
*
Author to whom correspondence should be addressed.
Data 2022, 7(12), 168; https://doi.org/10.3390/data7120168
Submission received: 1 November 2022 / Revised: 18 November 2022 / Accepted: 19 November 2022 / Published: 23 November 2022

Abstract

:
Automated spectrum analysis serves as a troubleshooting tool that helps to diagnose faults in wireless networks such as difficult signal propagation conditions and coexisting wireless networks. It provides a higher monitoring coverage while requiring less expertise compared with manual spectrum analysis. In this paper, we introduce a data set that can be used to train and evaluate deep learning models, capable of detecting frames from different wireless standards as well as interference between single frames. Since manually labeling a high variety of frames in different environments is too challenging, an artificial data generation pipeline was developed. The data set consists of 20,000 augmented signal segments, each containing a random number of different Wi-Fi and Bluetooth frames, their spectral image representations and labels that describe the position and type of frame within the spectrogram. The data set contains results of intermediate processing steps that enable the research or teaching community to create new data sets for specific requirements or to provide new interesting examination examples.
Data Set License: CC BY 4.0

1. Introduction

Due to the flexibility and mobility offered by wireless networks, they play an increasingly important role in industrial applications, consumer electronics, medical monitoring and building automation. However, it is challenging to meet the demanding latency and reliability requirements that are particularly prevalent in an industrial context. Applications that fail to meet these requirements suffer degradation, leading to economic losses and system failures. Inadequate signal propagation conditions due to path loss, multipath propagation, nlos and interference caused by co-existence problems with other wireless networks are among the most common sources of failures that lead to long service downtimes. Especially, in the ism frequency bands, the available channel capacity gets exhausted quickly, which is why co-existence problems have become a major problem. Techniques for mitigating these problems, as implemented in the UMTS (CDMA), Bluetooth (FHSS) and Wi-Fi6 and 5G (OFDMA) communications standards, cannot avoid them completely.
System failures caused by faults within the wireless communication system are difficult to detect and can lead to severe financial losses. Time-efficient troubleshooting and preventive countermeasures can only be guaranteed if troubleshooting equipment is used to constantly monitor the wireless communication networks. Intelligent spectrum analysis provides deep insight into the physical layer, enabling the elimination of co-existence problems between individual communication participants.
Many spectrum analysis approaches already exist in the literature. Most of these approaches [1,2,3,4,5,6] employ machine learning algorithms for automatic frame detection and classification. Some of the proposed approaches use advanced image recognition and object detection algorithms. Our proposed data set has already been evaluated [7] with a state-of-the-art object detection algorithm for the detection of co-existing standards and collisions in the network. Although ML-based approaches show significant improvement in accuracy compared with heuristic approaches, generating a substantial labeled data set is challenging. Especially, in the case of machine-learning-based object detection systems, the availability of labeled data is of prime importance. Capturing spectrograms of a real wireless communication in a shielded environment would be an intuitive idea. An rf channel emulator could provide the required variety of different environments. However, manual labeling is too time-consuming. Determining the timestamps and duration of frames sent is not feasible, as real traffic at this level is difficult to measure or reproduce. To generate labeled spectrograms, we have developed a data augmentation pipeline. Using pre-recorded frames of different rf standards, a policy specific to the communication standard randomly arranges these frames and modifies them using a channel model. The precise location of each frame is stored as a part of the label in the data set. In this way, the exact position and dimension of each frame is known, while maintaining the resemblance of spectral representations to the real measurements. Moreover, collisions between multiple frames can be generated in this way, which are difficult to detect in a real environment.
The proposed approach can be used to fully automate the generation of a large data set of spectrograms. We even provide Python scripts to generate the data set using user-specific parameters such as image resolution or color map. The scripts can also be used for conversion of the labels into a different format.

2. Data Set Description

This section gives an overview of the published data set and describes how it is structured. It consists of 20,000 spectrograms, which contain a total of about 362,780 Wi-Fi frames, 21,340 frames of each BLE 1 M Hz and BLE 2 M Hz , as well as BT and 77,600 collisions between different frames.
Figure 1 contains an example of the spectrograms that form the training data. The horizontal axis represents the time in μ s, and the vertical axis represents the frequency in MHz. The brightness encodes the reception strength in dBm and was colored using the viridis scale. The spectrogram depicted corresponds to a sampling rate of 60 M S per second and includes CCK (e.g., frame 1) and OFDM (frame 3) modulated 802.11 frames, as well as 2 M Hz BLE (frame 2) and BT (frame 4) frames. Frame 5 is a 20 M Hz Wi-Fi frame that is only partially within the acquisition bandwidth of the spectrogram. The bottom figure illustrates the same spectrogram, with the individual frames marked by colored bounding boxes. The thick red lines illustrate the areas where two or more frames overlap, referred to as collisions.
The data set not only contains the fully labeled simulated spectrogram frames but also files from two intermediate processing steps. First, there are separate frames recorded at a high snr that serve as input to the augmentation pipeline. Second, there are chunks of the time-domain signal in the form of complex values, available as raw binary data, from which final spectrogram images can be generated. The data set is stored in the following directory tree:
Data 07 00168 i001
The ./singlepacketsamples/ directory contains all the frames used by our automated pipeline to assemble them into the augmented spectrograms. These images can also be reused for creation of a new or modified pipeline with a different focus. Within the filenames, the properties of each frame were encoded, and therefore represent part of the labels. The binary files *.packet contain the complex-valued time signal samples of a single frame, consisting of pairs of 32-bit floating point numbers. The configuration file ./singlepacketsamples/configpacketcapture.toml contains parameters used for generating and capturing the single frames of the respective RF standard, described in more detail in Section 3.1.
The ./mergedpackets directory contains sample sections populated with a selection of single frames to which a simulated channel model was applied. These single-packet samples have the same format as the time-domain signal files described earlier. All corresponding labels are available as csv files with the same identification number. These labels contain detailed information about all frames within the signal section, see Table 1.
The ./results/ directory contains the final spectrogram images. They are stored in png file format together with the corresponding label files in Yolov4 format. In addition, there is a duplicate of each spectrogram containing the bounding box visualizations for all object classes. These files were generated using a Yolov4 object detector specifically trained for this purpose. However, all images containing bounding boxes are only used for evaluation of the generated results and must not be used for training.
The next sections provide a detailed description about how the published data set was created.

3. Data Set Generation

Figure 2 provides an overview of the developed data generation and enhancement pipeline, which is described in detail in the following sections.
At the first stage, parameterizable frames of different RF standards are generated by a Rhode and Schwarz® SMBV100B vector signal generator [8] and recorded by a SDR; details are explained in Section 3.1. Individual frames are then extracted from the recorded signal and stored separately. The second stage is called data augmentation; details are explained in Section 3.2. Here, the signals of the individual frames are randomly arranged according to a policy corresponding to the RF standards to form larger signal sections of 4.5   m s . During composition, a separate channel model is applied to each frame. Since the positioning of the frames is random, a number of frame collisions occur (shown in red in Figure 2), and the overlapping regions are labeled separately. This approach allows to simulate different RF environments whose appearance closely resembles real scenarios and significantly increases the variability within the data set. Subsequently, the composite signal sections are converted into actual spectrogram images, which are stored together with accurate labels; details are explained in Section 3.3. By following the method described above, we generated 20,000 spectrograms representing the given data set.
The following Section 3.1, describes the setup for generating and recording the various 802.11, Bluetooth® Low Energy (BLE) and Bluetooth® Classic (BT) frames, which serve as input to the actual augmentation pipeline.

3.1. Single-Frame Acquisition

A Rhode and Schwarz® SMBV100B vector signal generator [8] is used to generate raw single-frame signals. The parameters of these frames are defined in a configuration file (./singlepacketsamples/configpacketcapture.toml) and read by a Python program, which automates the process of signal generation and capture. To record the generated signals, a USRP N310 SDR from Ettus Research® [9] is used, which is connected to the vector signal generator via a coaxial cable to ensure sufficient SNR. All generated signals are recorded at a fixed sampling rate of 125 M S / s . From the measurements in form of a continuous sampling stream, individual frames are isolated using a simple threshold-based method and stored in separate files. Figure 3 shows the setup of the hardware components. In Figure 3, the output of the signal generator is connected to an input of the USRP via a coaxial cable. Additionally, an output of the USRP is directly connected to an input to loop back the signal of the simulated spectrogram. See Section 3.3.2 for more details on spectrogram generation using the USRP pipeline.
Table 2 displays the variety of parameters of different RF standards used for creation of the data set. They can also be found in the data set directory tree in ./singlepacketsamples/configpacketcapture.toml.
After the frame records of all defined rf standards are available for the defined parameter combinations, they are stored in ./singlepacketsamples/RF-Standard. The following Section 3.2 describes the data augmentation pipeline for creating simulated rf environments. It was implemented entirely in software, using single frames as an input to generate the actual training data set.

3.2. Generation of Simulated Signal Environments

To create spectrograms with a realistic appearance, we combine the signal of single frames that was recorded as described in Section 3.1 into a mutual signal segment. To maximize the diversity of the data set, different channel models are applied to each frame, defined by a set of random parameters. Table 3 lists all parameters that affect the spectrogram segments. The described parameters specify the distribution of the frames within the dataset to be created, representing the ratio of the individual standards in the entire dataset or the minimum and maximum values of a uniform distribution of values. For instance, the value of the "ratio of a spectograms without a frame" parameter highlights that there are 3 of spectrograms in the data set that only contain background noise.
Since the relative width of an RF frame within the spectrogram image changes with the different sampling rates, the data set should contain several rates. In this way, a model can be trained that can generalize the appearance when an sdr with a configurable sampling rate is used. All sampling rates supported by the sdr can be covered either by training a single model or by separate models. The sampling rates are defined in a tuple, where the number of generated signal segments is equal for all rates. To generate the present data, we chose the sampling rates supported by the usrp N310, which correspond to integer divisions of its master clock. Since all frames are captured at the maximum sampling rate of 125 M S / s , they are downsampled to the target rate when added to a section.
In addition to having the ratio of the included rf Standard defined, the chance of selecting a frame with a higher bandwidth than the target sample rate of the spectrogram section is hard-coded to be 1%. The ratio of the remaining frame parameters such as payload and modulation scheme distribution is implicitly defined by single-frame generation parameters; see Table 2. The imbalance in rf standards in favor of Wi-Fi over bt and ble was intentional, since the effects of the different channel models on wideband signals are greater, and long signals sent with a higher bandwidth also allow for much more complex collision patterns. This leads to a higher diversity compared with narrowband standards, which has to be considered when creating the data set.
The section length of 4.5   m s per spectrogram was iteratively found suitable for detecting even the shortest frames when the image is compressed to the target resolution while maintaining the real-time capability of the detection algorithm [7]. The number of frames between 18 and 25 to be added into a section is considered high for that length. Since the position of a frame is picked randomly, the diversity of interframe spacings and frame collision pattern is high. As shown in the flowchart in Figure 4, a new frame and position is selected once the maximum number of collisions is reached. Figure 4 illustrates the algorithm for combining individual frames into artificial spectrograms and how the random parameters are used.
The range of awgn to be added was chosen experimentally so that all frames are of varying strength and are not lost in background noise due to too low snr. This background noise is added before the spectrograms are created using the software-based method.
Compared with Table 3, which contains the parameters for the main spectrogram configuration, Table 4 defines the channel model parameters per single frame before it is added to a spectrogram segment.
The first parameter in the list describes the limits of gain values that are picked from a uniform distribution and applied on a frame when being added to the spectrogram section. In this way, different distances between the transmitting nodes and the spectrum analyzer’s observation position are simulated. To achieve the same mean spectral power density for all standards, the signal is amplified according to its bandwidth and normalized to 1 M Hz .
g a i n = r a n d o m _ g a i n + 10 log 10 s i g n a l _ b a n d w i t h 1   M Hz d B
The next parameter is the list of frequency offsets, which defines the center frequency shift of a frame from the horizontal center of the spectrogram. Only 20 of the frames are shifted from the center frequency to avoid too many frames being only partially within the sampling bandwidth. In order to apply a frequency shift to the time signal of a frame, it is upconverted by multiplying it with a sinusoidal signal having a frequency corresponding to the shift. It is then downconverted back to the target sampling rate, where aliasing effects are avoided by the upconversion and downconversion.
The remaining parameters are used to tune the multipath propagation model individually for each frame. A simpler model consisting only of a reflection path and random Doppler shift due to moving objects is applied to 50 of the frames. Figure 5 shows a spectrogram containing only the frames generated using these models. In particular, for the frames labeled 1, 2 and 3, clear frequency-selective fading effects of different simulated Doppler velocities are observed, but no frequency spreading is seen.
A more complex frequency-selective fading model is applied [10,11] to the other half of the frames. It simulates an environment with reflective objects and surfaces and is called Rician fading. A random combination of the K-factor, the number of mpc and a set of random sub-parameters for each reflected ray (time delay, size and standard and maximum deviation of pdp) describes a user-defined rf channel for each frame. The K-factor defines the ratio of the determined direct line-of-sight signal power to the scattered signal power. For low K-factors, the line-of-sight component becomes sub-dominant, approaching Rayleigh fading.
Different combinations of model parameters can have different effects on the spectral signal appearance, as shown in Figure 6.
Some significant channel parameters and their spectral representation for different frames types are marked in Figure 6. The corresponding channel parameters are listed in Table 5.
Frame 1 (20 MHz   802.11 b/g: CCK1 modulation) is severely affected by the effect of Doppler spreading and propagation along different paths, resulting in spectral broadening. The channel parameters of Frame 2 (20 MHz   802.11 n, MCS = 3) have a higher Doppler shift than Frame 1. However, Frame 2 has only a reflection component with no delay variation and has effectively been modeled as a determined line-of-sight transmission. Frame 3 (40 MHz   802.11 ac, MCS = 1) is an example of a strong influence by MPC. A clear fading is evident, showing a wide variation over time. Frame 4 is a 2 MHz BLE data frame with only one reflection that exhibits a delay deviation over time, which together with the high K-factor results in a slight frequency spread.
After the augmentation of the signal sections, they are still present in the form of complex numbers in binary format. The last step consists of generating spectrogram images.

3.3. Spectrogram Generation

The prior composition of frames into segments is performed on time-domain signals. To generate a time-spectral representation of these sections that can be used by image processing algorithms, two different methods are employed. The first spectrogram calculation is completely software-based. It provides full control over the added noise, and its fft output shows higher magnitude resolution compared with the hardware-based generation. The second method consists of the same pipeline that is used within the actual measurement and processing stage. It utilizes a usrp N310 for sending, recapturing and a hardware-based calculation of the fft from the time-domain signal inside the on-board fpga. While this approach allows for results that are very close to a real measurement system, the software-based generation of spectrograms provides better reusability due to its generic approach. For this reason, 80 of the spectrograms were obtained using the software-based method.
Regardless of the method used, the resulting spectrogram images must be reduced to a temporal resolution which guarantees real-time frame detection inference. Here, it must be taken into account that as little information as possible is lost.
The following subsections describe the individual components of the two spectrogram generation methods. Both of them take the time-domain signal of 4.5   m s sections as input, including combined frames with individual channel models, as described in Section 3.2.

3.3.1. Software-Based Spectrogram Generation

All sample segments that should be processed by this method have a noise level defined in their respective labels file, and the awgn was already added in order to simulate the thermal noise of a receiving amplifier.
The Python library matplotlib [12] is used to generate the spectrograms, scale them to the target resolution and save them as lossless PNG images. The ‘Hanning’ window is used for interpolation and ‘viridis’ is used for the colormap. A three-channel colormap shows better results than gray scale with our detection algorithm. Additional Python programs are provided, allowing a convenient regeneration of images for the use of different color maps or resolutions.

3.3.2. SDR Loopback Spectrogram Generation

Figure 7 shows the transmit–receive loopback pipeline used to generate spectrograms with the characteristic properties of a radio communication.
All function blocks within the usrp are implemented in hardware and controlled by the host PC. The transmit and receive pipelines must be synchronized precisely. The resolution of the magnitude of the hardware fft is lower compared with the software-based implementation by matplotlib. In order to preserve the low-intensity parts of the spectrum, an additional gain block shifts the range before calculating the squared power density. With this method, the noise floor is created by the actual hardware amplifier and not added artificially, making it more in line with reality.
The following Figure 8, Figure 9, Figure 10 and Figure 11 are further examples of spectrograms to give an impression about the appearance of the different parameters such as sample rate and method of spectrogram generation. The difference in appearance when using the hardware-based method is visible in Figure 8 and Figure 9 in direct comparison with Figure 10 and Figure 11 or previously shown spectrograms.

4. Discussion

In this paper, we present an extensive data set of spectrograms that can be used as training data for detecting frames of several rf standards. The data set is fully and extensively labeled, making it suitable for training machine learning models using supervised learning techniques, as well as for verifying the detection accuracy of computer vision algorithms. The data set was created using a novel method that records real radio signals of individual frames transmitted under controlled conditions and creates spectrograms using a specially designed data augmentation pipeline. In contrast to the direct recording of a transmission, the augmentation pipeline offers the possibility to freely parameterize the channel models to be used. This procedure greatly improves the versatility of the environmental conditions represented within the data set. Frames can be selectively placed in the spectrum, allowing the recreation of specific scenarios and extensive labeling of each image.
This data set was initially developed to train an object detection system that can recognize single frames from different rf standards and collisions between frames. Considering this objective, these specially tailored spectrograms differ from real measurements. The main deviation is inaccurate representation of temporal dependencies between frames within a spectrogram. Communication between the individual participants of a network is governed by certain definitions of a communication standard, which can be recognized in the spectrum in a condensed form. This includes the duration and time interval between individual frames, as well as switching to a different channel or to a different bandwidth. For example, frames sent in rapid succession from one network node are subject to almost the same channel characteristics, and thus there are no sudden changes in the reception power at the receiving node.
These restrictions on the use of the data set can be addressed in the future. Functional extensions are straightforward, since we not only provide a single data set but also the underlying raw frames and the essential Python programs to generate the spectrograms. Simulator coupling or a more simple rule-based approach could be used to model temporal dependencies of a communication realistically. Subsequent developments include the generation of labeled spectrogram videos to enable the training of specific models for real-time spectral analysis and an extension of the frames database to include sources of interference, such as microwaves or radar systems that can be used for generating spectrograms. Possible research objectives for further extensions of the data set could include the automatic detection of unwanted signals in licensed frequency bands or device identification based on spectral features to enhance authentication security.
The dataset can be a potential platform for testing and validating future co-existential analysis methods. It can be used for analyzing techniques proposed for qos degradation issues or interference detection. In its current form, this data set offers a wide range of possibilities for testing and developing algorithms in the area of advanced ML and DL models for image recognition and computer vision in the context of research and teaching. It serves as a valuable and easy-to-use reference data set.

Author Contributions

J.W. is responsible for conceptualization, software, data set preparation, validation, visualization and writing; U.W. contributed to conceptualization, software and writing; V.J. contributed to writing. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Federal Ministry of Education and Research of the Federal Republic of Germany (BMBF) within the PENTA project “SunRISE” (https://www.project-sunrise.eu/ accessed on 18 November 2022) under the Project Number 16ES0974 and in cooperation with the Center for Analytics–Data–Applications (ADA-Center) which is supported by the Bavarian Ministry of Economic Affairs, Regional Development and Energy within the framework of “BAYERN DIGITAL II” (20-3410-2-9-8).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data set described in this document is archived under https://fordatis.fraunhofer.de/handle/fordatis/287 (accessed on 18 November 2022) or http://dx.doi.org/10.24406/fordatis/216 (accessed on 18 November 2022). Corresponding helper scripts are available under https://gitlab.cc-asp.fraunhofer.de/ifk_public/sunrise/public-mdpi-dataset-helper-scripts/-/tree/dataset_20220711 (accessed on 18 November 2022).

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Fehske, A.; Gaeddert, J.; Reed, J.H. A new approach to signal classification using spectral correlation and neural networks. In Proceedings of the First IEEE International Symposium on New Frontiers in Dynamic Spectrum Access Networks, Baltimore, MD, USA, 8–11 November 2005. [Google Scholar]
  2. Yu, J.; Alhassoun, M.; Buehrer, R.M. Interference classification using deep neural networks. In Proceedings of the IEEE 92nd Vehicular Technology Conference (VTC2020-Fall), Victoria, BC, Canada, 18 November–16 December 2020. [Google Scholar]
  3. Toma, A.; Nawaz, T.; Gao, Y.; Marcenaro, L.; Regazzoni, C.S. Regazzoni, Interference mitigation in wideband radios using spectrum correlation and neural network. IET Commun. 2019, 13, 1336–1347. [Google Scholar] [CrossRef]
  4. O’Shea, T.J.; Corgan, J.; Clancy, T.C. Convolutional radio modulation recognition networks. In International Conference on Engineering Applications of Neural Networks; Springer: Cham, Switzerland, 2016; pp. 213–226. [Google Scholar]
  5. O’Shea, T.J.; Roy, T.; Clancy, T.C. Over-the-air deep learning based radio signal classification. IEEE J. Sel. Top. Signal Process. 2018, 12, 168–179. [Google Scholar] [CrossRef] [Green Version]
  6. Mennes, R.; Claeys, M.; De Figueiredo, F.A.; Jabandžić, I.; Moerman, I.; Latré, S. Deep learning-based spectrum prediction collision avoidance for hybrid wireless environments. IEEE Access 2019, 7, 45818–45830. [Google Scholar] [CrossRef]
  7. Wicht, J.; Wetzker, U.; Frotzscher, A. Deep Learning Based Real-Time Spectrum Analysis for Wireless Networks. In European Wireless 2021, Proceedings of the 26th European Wireless Conference, Verona, Italy, 10–12 November 2021; VDE: Berlin, Germany; pp. 1–6.
  8. Rhode & Schwarz R&S® SMBV100B. Available online: https://www.rohde-schwarz.com/uk/products/test-and-measurement/vector-signal-generators/rs-smbv100b-vector-signal-generator_63493-519808.html (accessed on 29 July 2022).
  9. USRP N310. Ettus Research, a National Instruments Brand. Available online: https://kb.ettus.com/N300/N310 (accessed on 29 July 2021).
  10. Alimohammad, A.; Fard, S.F.; Cockburn, B.F.; Schlegel, C. An Accurate and Compact Rayleigh and Rician Fading Channel Simulator. In Proceedings of the VTC Spring 2008–IEEE Vehicular Technology Conference, Singapore, 11–14 May 2008; IEEE: New York, NY, USA, 2008; pp. 409–413. [Google Scholar] [CrossRef]
  11. Ren, F.; Zheng, Y.R. A low-complexity hardware implementation of discrete-time frequency-selective Rayleigh fading channels. In Proceedings of the 2009 IEEE International Symposium on Circuits and Systems, Taipei, Taiwan, 24–27 May 2009; pp. 1759–1762. [Google Scholar] [CrossRef]
  12. Caswell, T.A.; Droettboom, M.; Hunter, J.; Lee, A.; Firing, E.; Stansby, D.; Klymak, J.; de Andrade, E.S.; Nielsen, J.H.; Varoquaux, N.; et al. matplotlib/matplotlib: REL: v3.1.1 (v3.1.1). Zenodo. 2019. Available online: https://doi.org/10.5281/zenodo.3264781 (accessed on 18 November 2022).
Figure 1. Example of a simulated spectrogram. It was generated using the software-based method and has 60 M Hz bandwidth. Sub-figure (b) depicts the same spectrogram as sub-figure (a), but with colored bounding boxes to illustrate the labeled RF frames.
Figure 1. Example of a simulated spectrogram. It was generated using the software-based method and has 60 M Hz bandwidth. Sub-figure (b) depicts the same spectrogram as sub-figure (a), but with colored bounding boxes to illustrate the labeled RF frames.
Data 07 00168 g001
Figure 2. General structure of our automated pipeline.
Figure 2. General structure of our automated pipeline.
Data 07 00168 g002
Figure 3. Hardware setup used to generate and capture single-frame signals. Host PC with a 10 G / s network card (left), SMBV100B vector signal generator (right) and the USRP N310 on top.
Figure 3. Hardware setup used to generate and capture single-frame signals. Host PC with a 10 G / s network card (left), SMBV100B vector signal generator (right) and the USRP N310 on top.
Data 07 00168 g003
Figure 4. Methodology used for spectrogram augmentation.
Figure 4. Methodology used for spectrogram augmentation.
Data 07 00168 g004
Figure 5. Example spectrogram with 125  M Hz bandwidth. Models with only frequency-selective or flat-fading characteristics but no time variance or frequency spreading were applied to all frames. Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Figure 5. Example spectrogram with 125  M Hz bandwidth. Models with only frequency-selective or flat-fading characteristics but no time variance or frequency spreading were applied to all frames. Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g005
Figure 6. Example of a spectrogram with 45 M Hz bandwidth. The Doppler effect and more complex multipath propagation was simulated. Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Figure 6. Example of a spectrogram with 45 M Hz bandwidth. The Doppler effect and more complex multipath propagation was simulated. Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g006
Figure 7. Pipeline that is not only used during the actual measurement and inference stage but also during training data generation for having samples with appearances close to actual measurements.
Figure 7. Pipeline that is not only used during the actual measurement and inference stage but also during training data generation for having samples with appearances close to actual measurements.
Data 07 00168 g007
Figure 8. Example of a simulated spectrogram, created using the usrp loopback at 45 M S / s . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Figure 8. Example of a simulated spectrogram, created using the usrp loopback at 45 M S / s . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g008
Figure 9. Example of a simulated spectrogram, created using the usrp loopback at 60 M S / s . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Figure 9. Example of a simulated spectrogram, created using the usrp loopback at 60 M S / s . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g009
Figure 10. Example of a simulated spectrogram, created using the software-based method with a bandwidth of 25 M Hz Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF-frames. The bold red boxes mark overlapping areas of collided frames.
Figure 10. Example of a simulated spectrogram, created using the software-based method with a bandwidth of 25 M Hz Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF-frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g010
Figure 11. Example of a simulated spectrogram, created using the software-based method with a bandwidth of 60 M Hz . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Figure 11. Example of a simulated spectrogram, created using the software-based method with a bandwidth of 60 M Hz . Sub-figure (b) depicts the same spectrogram as sub-figure (a) but with colored bounding boxes to illustrate the labeled RF frames. The bold red boxes mark overlapping areas of collided frames.
Data 07 00168 g011
Table 1. Description of the label file columns.
Table 1. Description of the label file columns.
ColumnDescription
IDIndex given to a frame when added; when describing a collision it refers the involved frames indices, connected by a ‘-’
sample_position_startNumber of samples after which the frame is added to the section
sample_position_endNumber of samples after which the added frame ends
pdu_lengthPayload of the frame in number of bytes (empty when collision)
LevelSignal level of the frame in dB with an arbitrary reference
BandwidthBandwidth of the frame signal
freq_offsetFrequency shift from center frequency of the sample range
ClassString giving the class name [WLAN, BT_classic, BLE_1MHz, BLE_2MHz, collision]
rf_stdString giving the specific Wi-Fi standard (empty if not Wi-Fi)
WLAN_mcsModulation coding scheme of a Wi-Fi frame (empty if not Wi-Fi)
BT_packet_typeModulation coding scheme of a BT or BLE frame (empty if not BT or BLE)
noise_lvlMean magnitude of noise level of the complete section or a string “usrp_txrx_loop” in case of hardware-induced noise
sample_rateSample rate of the complete section
samples_totalNumber of samples within the complete section
doppler_speed_kmhSpeed of objects that would produce the emulated Doppler effect on that frame
k_factorChannel model parameter of the individual frame (ratio of line-of-sight signal power over the scattered signal power)
multipath_componentsChannel model parameter of the individual frame (number of reflections)
PDP_delaysChannel model parameter of the individual frame (delay (in samples) for arriving reflected Ray)
PDP_delay_max_devChannel model parameter of the individual frame (maximum deviation of delay per reflection)
PDP_delay_std_devChannel model parameter of the individual frame (step-size Gaussian standard deviation per reflection)
PDP_magChannel model parameter of the individual frame (magnitude of each arriving reflected Ray)
IdentifierUnique identifier used within the filename
Table 2. RF standard parameters used for creation of the data set.
Table 2. RF standard parameters used for creation of the data set.
Communication StandardsFrame ParametersTotal Number of Frames Per Standard
IEEE 802.11 b/gPayload length480 (147 × 20 M Hz +
IEEE 802.11 nmcs144 × 40 M Hz +
IEEE 802.11 acFrame bandwidth189 × 80 M Hz )
Packet type (Data, beacon, trigger or sounding frames)
blePayload length29 (8 × 2 M Hz +
Channel type (ADV, DATA)21 × 1 M Hz )
Packet type (DATA, AIND)
Packet format (L1M, L2M, LCOD)
btPayload length29
Packet type (DHx: 1Mbps; ADHx: 2Mbps, AEDHx: 3Mbps)
Channel type (ADV, DATA)
Table 3. Parameters used for generating the signal sections and resulting spectrograms as defined in ./configtrainingdata.toml.
Table 3. Parameters used for generating the signal sections and resulting spectrograms as defined in ./configtrainingdata.toml.
ParameterValue
sample rates(25, 45, 60, 125) M S / s
number of spectrograms per sample rate5000
ratio of RF standards [Wi-Fi]0.85
ratio of RF standards [BT]0.05
ratio of RF standards [BLE (1 M Hz )]0.05
ratio of RF standards [BLE (2 M Hz )]0.05
time section per spectrogram 4.5   m s
number of frames per spectrogram [min, max](18, 25)
maximum number of frame collisions4
ratio of spectrograms without a frame0.03
ratio of spectrograms with a single frame0.1
ratio of spectrograms generated using a usrp0.2
amplitude range of added noise [min, max](0.0055, 0.0065)
resolution of the spectrogram images [x, y](1024, 192)
Table 4. Random channel model applied to a single frame before being added to a spectrogram.
Table 4. Random channel model applied to a single frame before being added to a spectrogram.
ParameterValue
gain per frame [min, max][ 20 , 6]
frequency offsets 65 M Hz …65 M Hz , step size = 5 M Hz
ratio of frames with frequency offset0.2
channel model [max k factor]10
channel model [is ricean]true
channel model [max doppler speed]20 k m / h
channel model [max multi-path components]6
channel model [delay standard deviation]0.0
ratio of frames with multi-path components0.5
Table 5. RF standard parameters used for creation of the data set.
Table 5. RF standard parameters used for creation of the data set.
Frame IDFading Model Parameter
1Doppler_speed = 17 k m / h
K-factor = 5
Multipath components = 6
PDP delays = [1 2 3 4 5 6] samples
PDP delay_max_dev = [0.1 0.2 0.3 0.4 0.5 0.6]
PDP delay_std_dev = [0.0121 0.0072 0.0101 0.0095 0.0053 0.0061]
PDP magnitude = [0.97 0.79 0.79 0.67 0.54 0.38]
2Doppler_speed = 19 k m / h
K-factor = 7
Multipath components = 1
PDP delays = [1]
PDP delay_max_dev = [0]
PDP delay_std_dev = [0]
PDP magnitude = [1]
3Doppler_speed = 7 k m / h
K-factor = 6
Multipath components = 5
PDP delays = [1 2 3 4 5]
PDP delay_max_dev = [0.1 0.2 0.3 0.4 0.5]
PDP delay_std_dev = [0.0096 0.0195 0.0067 0.0049 0.0033]
PDP magnitude = [0.87 0.78 0.75 0.59 0.5]
4Doppler_speed = 5 k m / h
K-factor = 10
Multipath components = 1
PDP delays = [1]
PDP delay_max_dev = [0.1]
PDP delay_std_dev = [0.0037]
PDP magnitude = [0.9]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wicht, J.; Wetzker, U.; Jain, V. Spectrogram Data Set for Deep-Learning-Based RF Frame Detection. Data 2022, 7, 168. https://doi.org/10.3390/data7120168

AMA Style

Wicht J, Wetzker U, Jain V. Spectrogram Data Set for Deep-Learning-Based RF Frame Detection. Data. 2022; 7(12):168. https://doi.org/10.3390/data7120168

Chicago/Turabian Style

Wicht, Jakob, Ulf Wetzker, and Vineeta Jain. 2022. "Spectrogram Data Set for Deep-Learning-Based RF Frame Detection" Data 7, no. 12: 168. https://doi.org/10.3390/data7120168

Article Metrics

Back to TopTop