Next Article in Journal
Towards Latency Bypass and Scalability Maintain in Digital Substation Communication Domain with IEC 62439-3 Based Network Architecture
Previous Article in Journal
Enhancing Targeted Minority Class Prediction in Sentence-Level Relation Extraction
Previous Article in Special Issue
Melding Fog Computing and IoT for Deploying Secure, Response-Capable Healthcare Services in 5G and Beyond
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Active 2D-DNA Fingerprinting of WirelessHART Adapters to Ensure Operational Integrity in Industrial Systems

Department of Electrical and Computer Engineering, US Air Force Institute of Technology, Wright-Patterson AFB, OH 45433, USA
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(13), 4906; https://doi.org/10.3390/s22134906
Submission received: 26 May 2022 / Revised: 22 June 2022 / Accepted: 24 June 2022 / Published: 29 June 2022
(This article belongs to the Special Issue Internet of Things for Industrial Applications)

Abstract

:
The need for reliable communications in industrial systems becomes more evident as industries strive to increase reliance on automation. This trend has sustained the adoption of WirelessHART communications as a key enabling technology and its operational integrity must be ensured. This paper focuses on demonstrating pre-deployment counterfeit detection using active 2D Distinct Native Attribute (2D-DNA) fingerprinting. Counterfeit detection is demonstrated using experimentally collected signals from eight commercial WirelessHART adapters. Adapter fingerprints are used to train 56 Multiple Discriminant Analysis (MDA) models with each representing five authentic network devices. The three non-modeled devices are introduced as counterfeits and a total of 840 individual authentic (modeled) versus counterfeit (non-modeled) ID verification assessments performed. Counterfeit detection is performed on a fingerprint-by-fingerprint basis with best case per-device Counterfeit Detection Rate (%CDR) estimates including 87.6% < %CDR < 99.9% and yielding an average cross-device %CDR ≈ 92.5%. This full-dimensional feature set performance was echoed by dimensionally reduced feature set performance that included per-device 87.0% < %CDR < 99.7% and average cross-device %CDR ≈ 91.4% using only 18-of-291 features—the demonstrated %CDR > 90% with an approximate 92% reduction in the number of fingerprint features is sufficiently promising for small-scale network applications and warrants further consideration.

1. Introduction

The adoption of wireless technologies is an ongoing trend that has “major implications” for achieving Internet of Things (IoT), Industrial Internet of Things (IIoT), and Fourth Industrial Revolution (IR 4.0) objectives [1]. Among the objectives is a desire to minimize or entirely eliminate human intervention through increased reliance on automation [2]. This includes using real-world virtualization in industrial system frameworks that integrate IoT/IIoT devices, computation, networking, and physical processes [3]. While the inherent efficiencies, cost savings, etc., brought about by hands-off automation, are certainly desirable, benefits are only realized if the supporting devices achieve their full life expectancy [4] and operational cyber security risks are mitigated. These risks include an increase in the number of wireless doorways for entering the operational attack surface that has expanded with the adoption and networked integration of IoT/IIoT devices [5]. This is particularly important when considering the potential for catastrophic failure in applications involving safety and the potential loss of life.
There have been decades of research and demonstration activity addressing the use of Radio Frequency Fingerprinting (RFF) to provide post-deployment operational protection of fielded electronic, electrical, and electromechanical devices. The various RF fingerprint features, extraction methods, and discrimination algorithms considered are aptly summarized by the most recent review provided in [6]. From the RF fingerprinting protection perspective, the methods addressed in [6] are collectively referred to herein as passive fingerprinting methods. That is, the fingerprint features are extracted from collected emissions of a component, subassembly, or device that is actively operating and performing its intended function. As evident in [6], passive fingerprinting methods are the most prolific and most appropriate for post-deployment protection within the field operation phase of a device’s technical lifespan [4].
The focus here is on active fingerprinting which is much less prolific and uses features extracted from stimulated responses of non-operating components, subassemblies, or devices. Active fingerprinting is most appropriate for pre-deployment protection (e.g., counterfeit detection) within the near-cradle phase of a device’s technical lifespan [4]. Among the numerous RFF citations in [6] are the earliest RF fingerprinting works [7,8,9,10] that formed the basis for a decade of subsequent activity involving passive DNA fingerprinting of wireless communication devices [4,11,12,13,14,15,16,17]. More recently, there has been a transition to active DNA fingerprinting [18,19,20,21] with work in [19] representing the first application of active DNA fingerprinting to wireless communication devices. Community interest in the first-look 1D-DNA fingerprinting work in [19] using WirelessHART devices motivated the next-step activity presented here and consideration of active 2D-DNA fingerprinting to provide performance improvement. In this case, 1D is referring to fingerprints generated from a single domain response (e.g., instantaneous time) and 2D is referring to fingerprints generated from a multi-domain response (e.g., time-frequency).

1.1. Operational Motivation

Wireless Highway Addressable Remote Transducer (WirelessHART) signaling continues to emerge as a key enabling technology for achieving the level of IoT/IIoT network integration required for effective operations. WirelessHART is one of the two most widely used industrial international standards, with the estimated number of deployed field devices reaching into the tens-of-millions [2]. Its proliferation is expected to continue as manufacturers produce hundreds-of-thousands of devices that are introduced into the supply chain each year [22]. The widespread adoption of WirelessHART throughout European and North American industries is attributable to several key factors. Some of these include [1,2,22]:
  • Operation based on the legacy wired HART protocol—users can take full advantage of prior experience, prior training, prior tool purchases, etc.;
  • Considerable reduction in deployment, installation, and maintenance cost—minimal to no additional infrastructure cabling required;
  • Network architectural flexibility—expansion easily accommodated by adding connectivity to additional field devices and/or other nearby networks;
  • Dramatic reduction in device commissioning times—device benchtop programing and field installation completed in hours versus days.
The demonstrations here are based on WirelessHART signaling given its widespread adoption and use to achieve IoT/IIoT and IR 4.0 objectives. The focus here is on relatively small-scale networks consisting of a gateway and five or more field devices. A five-device network configuration provides “sufficiently redundant mesh network operation” [22] while taking advantage of architectural flexibility to provide the required sub-network and cross-network interconnectivity within a “classical Industrial Wireless Sensor Network (ISWN) architecture” [2]. As with large-scale networks, care must be taken in small-scale WirelessHART networks supporting critical applications involving personnel safety. This is particularly important when considering the hundreds-of-thousands of WirelessHART devices that enter the supply chain annually [22].
As with all electronic devices (integrated circuits, circuit boards, sub-assemblies, etc.) entering the supply chain, there is concern that operational integrity may be compromised through the introduction of counterfeits. The extent of compromised device effects ranges from degraded functionality (slow, sluggish, inefficient operation) to premature failure (lifecycle termination)—both manifesting possible catastrophic results. The potential for this occurring can be minimized by employing near-cradle [4] pre-deployment protection measures to detect devices that have been intentionally modified or counterfeited before placing them into service. Near-cradle counterfeit protection is at one lifecycle extreme and includes two fundamentally different Radio Frequency (RF) methods: (1) active RF Identification (RFID) that exploits interrogated responses of onboard functionality that is intentionally embedded at the time of manufacture [23,24], and (2) active RF-DNA fingerprinting that exploits inherent response uniqueness resulting from component, device architecture, and manufacturing process differences [18,19,20,21].

1.2. Technical Motivation

The development and demonstration of active DNA fingerprinting methods in [19,21] to support pre-deployment protection was complementary to prior passive DNA fingerprinting works supporting post-deployment operational protection [4,11,12,14,15,16,17]. The distinction between passive and active DNA fingerprinting is not based on device operating conditions but rather the DNA generation process employed. Passive DNA fingerprinting uses features extracted from transmitted responses (wired or wireless) of an operating device that is physically connected within a system and performing its intended function—passive DNA fingerprinting is a method for providing operational protection during a device’s technical lifecycle [4]. Active DNA fingerprinting uses features extracted from externally stimulated device responses of an uninstalled non-operably connected device—active DNA fingerprinting is a method for providing near-cradle protection at the onset of a device’s technical lifecycle [4].
Active DNA fingerprinting was pursued here given prior success in microwave system state estimation [18,20] and integrated circuit discrimination [21]. These previous works demonstrated benefits for using wideband (energy spanning 100s to 1000s of MHz) stimulation signals as a means to increase source-to-device electromagnetic interaction and induce fingerprint feature variability and uniqueness. The degree of source-to-device interaction is fundamentally driven by the stimulating signal characteristics (e.g., time varying amplitude, phase, and/or frequency), the device’s electrical/electronic architecture complexity, and resultant higher-order multi-frequency intermodulation effects. The active DNA fingerprinting work in [21] introduced a wideband Stepped Frequency Modulated (SFM) stimulus signal that was subsequently adopted for WirelessHART adapter discrimination work in [19]. The active DNA work in [19] considered discrimination of the same four Siemens AW210 [25] and four Pepperl+Fuchs Bullet [26] WirelessHART adapters used for passive DNA discrimination assessments in [4,14].
The active DNA fingerprinting work in [19] used an SFM stimulus signal with one-dimensional DNA (1D-DNA) features extracted from instantaneous time domain device responses. Considering the eight available WirelessHART adapters, 8-choose-6 = 28 different 6-class Multiple Discriminant Analysis (MDA) models were developed to represent six authentic devices. The two held-out devices were introduced as counterfeits, and a total of 28 × 6 × 2 = 336 individual counterfeit device ID verification assessments performed. Collectively considering all 28 models and the 336 individual assessments, a Counterfeit Detection Rate of %CDR ≈ 92% was demonstrated. While community feedback for work in [19] was generally favorable, there was also a clear desire to improve upon the demonstrated %CDR to better support operational motivation objectives, i.e., decrease the number of counterfeit devices installed in the field and increase the pool of certified devices that may be introduced into the supply chain.
To improve %CDR, alternate DNA features were considered here using the experimentally collected WirelessHART device responses from [19]. It was observed that non-coherent phase transitions occur as the SFM source stimulus transitions between frequency steps. These transitions appear as time domain transients in the device output responses. Thus, two-dimensional DNA (2D-DNA) features are considered here using a Gabor Transform (GT), given its prior success in [13] for signals exhibiting similar transient effects, and to obtain signal information based on the local distribution of signal energy as a function of frequency. Demonstrations are performed here in light of targeting small-scale (e.g., five sensors) IoT/IIoT and IR 4.0 implementations. Relative to results in [19], demonstrations here are based on (1) a total of 8-choose-5 = 56 different 5-class MDA models—factor-of-2 increase in the number of models, with (2) three held-out counterfeit devices per model used to perform 56 × 5 × 3 = 840 individual counterfeit device ID verification assessments—factor-of-2.5 increase in the number of counterfeit assessments.

2. Demonstration Methodology

This section presents details for the sequence of steps taken to conduct experimental demonstrations and generate the results presented in Section 3. Summary details are provided for each step in the indicated subsection. These steps include:
  • Response Collection and Processing in Section 2.1—this includes Device Under Test (DUT) emplacement, DUT stimulation, DUT stimulated output collection, and pre-fingerprint generation signal processing (filtering and decimation) to reduce computational complexity and improve discriminability;
  • 1D Time Domain DNA (TdDna) Fingerprint Generation in Section 2.2—this includes generation of device TdDna fingerprints used to provide a performance baseline representing the pre-existing 1D-DNA fingerprinting capability;
  • 2D Gabor Transform DNA (GtDna) Fingerprint Generation in Section 2.3—this includes generation of device GtDna fingerprints used to demonstrate performance benefits of 2D-DNA fingerprinting considered herein;
  • Multiple Discriminant Analysis (MDA) in Section 2.4—this includes cross-validated training of the MDA models required for device discrimination assessments;
  • Device Discrimination in Section 2.5—this includes implementation of multi-model MDA device classification as a necessary precursor to implementing the device ID verification process to perform counterfeit detection and estimate %CDR;
  • Dimensional Reduction Analysis (DRA) in Section 2.5—this includes final actions taken to reduce the number of fingerprint features required to achieve a given level of discrimination performance while improving computational efficiency;

2.1. Response Collection and Processing

The experimentally collected signals used here for demonstration were originally collected in support of work reported in [19] using the setup shown in Figure 1. The basic setup was adopted from integrated circuit anti-counterfeiting work originally developed and demonstrated in [21] and includes three main hardware elements: (1) a Keysight N5222B PNA microwave network analyzer [27] used to provide the DUT input stimulus s IN t , (2) a LeCroy WaveMaster 825Zi-A oscilloscope [28] used to collect the DUT output response s OUT t , and (3) a WirelessHART adapter serving as the DUT. The N5222B source parameters were set to produce the SFM stimulus signal s SFM t which was power divided and input as s IN t to 1-of-5 available DUT wires denoted as W IN j for j∈{1, 2, …, 5}.
Given a goal of maximizing the SFM source-to-DUT electromagnetic interaction and increasing DUT discrimination, the SFM stimulus parameters were empirically determined based on discrete settings available on the N5222B analyzer. The resultant SFM parameters used for experimental collection included [19] (1) a total of NSFM = 9 frequency steps spanning a total frequency range of approximately 400 MHz < f < 450 MHz, with (2) the duration of each frequency step being TΔ = 0.125 ms for a total SFM pulse duration of TSFM = 1.125 ms. The DUT s OUT t response was collected by the 825Zi-A oscilloscope, digitized, and stored for subsequent post-collection processing.
For the SFM stimulus applied to a given W IN j for j∈{1, 2, …, 5}, the DUT output response s OUT t was collected from one of the remaining wires, denoted as W OUT k  k ∈ {1, 2, …, 5} for kj. Given there are five connection wires on the WirelessHART adapters, there are 20 order-matters permutations of W IN j : W OUT k (stimulus input versus output response) wire pairs that could be considered for active DNA fingerprinting. Demonstrations here are based on collections made in [19], with W IN j being the device Direct Power connecting wire and W OUT k being the device HART Signal connecting wire. It was empirically determined that this wire pair consistently yielded discernable device responses and consistent fingerprint features across all eight experimental devices.
To reduce computational complexity and improve device discrimination relative to results reported in [19], additional pre-fingerprint generation processing was introduced here. For each collected device response, the sequential pre-fingerprint generation processing included (a) BandPass (BP) filtering at the as-collected center frequency of fCol = 425 MHz using a 16th-order Butterworth filter having a passband of WBP = 50 MHz, (b) Down-Conversion (D/C) to near-BaseBand (BB) using a frequency of fD/C = 375 MHz, (c) post-D/C BP filtering at a center frequency of fCtr = 50 MHz using a 16th-order Butterworth filter having a passband of WBP = 50 MHz, and (d) sample decimation by a factor of five. Thus, each of the as-received WirelessHART responses at a sample rate of fS = 1 GSps (1,115,000 samples per pulse) were converted to an fS = 200 MSps rate (230,000 samples per pulse) prior to fingerprint generation.
The impact of pre-fingerprint generation processing for a representative WirelessHART response is shown in Figure 2. This processing was performed for collections from each of the NDev = 8 WirelessHART adapters (D1, D2, …, D8) listed in Table 1. Although the Siemens AW210 [25] and Pepperl+Fuchs Bullet [26] device labels make it appear that the devices are from two different manufacturers, it was determined that these devices are actually from the same manufacturer and were distributed under two different labels with dissimilar serial number sequencing—this difference is a result of company ownership transition to Pepperl+Fuchs. Thus, the discrimination conditions being considered represent the most challenging, like-model intra-manufacturer conditions using identical hardware devices that vary by serial number.
Table 1 shows that the devices were received with two different versions of operating firmware—the firmware version number is available when the devices are connected to the gateway. In an attempt to remove firmware as an experimental variable, the manufacturer was contacted about having the D1 and D3 devices reprogrammed with the version 200 firmware. Researchers were told that this was not a customer support option and thus firmware remained as an uncontrolled experimental variable. As supported by results presented here in Section 3 and previous related work [4,14,19] using these same devices, there is no correlation between firmware version and device discriminability.

2.2. 1D Time Domain DNA (TdDna) Fingerprint Generation

The time domain DNA (TdDna) fingerprint generation process used here has evolved over time and has been predominantly used in wireless passive DNA fingerprinting applications [4,13,14,15,18]. Selected process details are presented here for completeness, and the reader is referred to [4] for more details. Statistical fingerprint features are calculated from instantaneous responses of the down-converted, bandpass filtered, near-baseband pulses such as illustrated in Figure 3. Denoting the real-valued sample sequence as {sOut(n)}, the fingerprinted responses include instantaneous (1) magnitude calculated as M(n) = |sOut(n)}|, (2) phase calculated as Θ(n) = tan−1[HQ(n)/HR(n)] where HR(n) and HQ(n) are real and imaginary components of the Hilbert Transform [29] denoted by Hilbert[sOut(n)], and (3) frequency calculated as Φ(n) = gradient [Θ(n)].
A Region Of Interest (ROI) within {sOut(n)} is selected and the corresponding instantaneous {Μ(n)}, {Θ(n)} and {Φ(n)} response sequences are divided into NRgn subregions for statistical feature calculation. This is illustrated in Figure 3, which shows the {Μ(n)} magnitude response for the representative pulse shown in Figure 2. Considering the calculation of three statistics (variance, skewness and kurtosis [30]) using samples within each of the NRgn = 12 indicated subregions, and across all samples within entire ROI as well, the time domain fingerprints include a total of NFD,TD = (12+ 1) × 3 × 3 = 117 features when accounting for the three instantaneous {Μ(n)}, {Θ(n)} and {Φ(n)} response sequences.

2.3. 2D Gabor Transform DNA (GtDna) Fingerprint Generation

Consideration of Gabor transform features is motivated by related historical work [12,13,31,32,33] that considered detection, characterization and exploitation of transient and nonlinear effects in time varying signals. These effects are manifest in the Gabor transform space as localized signal energy distributions that vary as a function of frequency. The N5222B source inherently produces an SFM pulse having a non-uniform amplitude response across time. As evident in Figure 3, this variation is generally preserved in the device output response although altered by the SFM-to-DUT interaction. Empirical analysis showed that response amplitude transitions at all step boundaries corresponded to instantaneous phase transients that randomly varied at each boundary on a pulse-by-pulse basis for all devices. Thus, device responses here possess transient characteristics consistent with those considered in [12,13,31,32], and the Gabor transform was deemed to be a reasonable first choice for 2D DNA fingerprint generation.
The mkth Gabor transform coefficient for sampled signal s(n) is given by [12,13,31,33]
G m k = n = 1 M T N Δ s n W n m N Δ e j 2 π k n / M F ,
where W(n) is a given synthesis window of width Wτ, * denotes complex conjugate, m = 1, 2, …, MT, where MT is the total number of time index shifts, k = 0, 1, …, MF – 1, where MF is the total number of frequency index shifts, and NΔ is the number of samples shifted between transformations. Additional parameter constraints for the Gabor transform given by (1) include MFNΔ and mod(MT × NΔ, MF) = 0. Consistent with [12,13,31], an energy normalized Hamming window was used for W(n).
The utility of Gabor transformation for highlighting transients and localizing energy concentration is evident by comparing the conventional spectrogram response in Figure 4 with the Gabor transform response in Figure 5. These responses are shown using the same color bar scale (dB) and were generated using the representative SFM response used for Figure 3. The Gabor transform response was generated using a window width of Wτ = 1 × 10−3, NΔ = 460, MT = 500, and MF = 500 for a resultant (MT × NΔ)/MF = 460. As implemented here, the transformation for MF = 500 > NΔ = 460 represents oversampling conditions, which are generally desirable when processing noisy data [31,33].
Relative to the spectrogram response, it is evident in Figure 5 that the highest degree of Gabor localization occurs around 3-of-9 frequency step regions (k = 55, 82, 109) and a moderate degree of localization occurs around extreme frequency step regions (k = 28, 136). The Gabor localization effects are also evident by comparing low energy time-frequency regions where the Gabor response exhibits a more distinct structure. Representative examples of this include the response regions bounded within (1) k ∈ {20, 21, …, 58} and m ∈ {175, 176, …, 500}, and (2) k ∈ {70, 71, …, 116} and m ∈ {1, 2, …, 175}.
The Gabor transform features are generated from a given 2D Gabor transform response, e.g., the normalized magnitude response such as shown in Figure 5. As shown in Figure 6, a fingerprinting ROI is identified and divided into NTim × NFrq 2D subregions (patches), where NTim and NFrq are the number of time dimension and frequency dimension indices defining each patch. Fingerprint generation indexing is set such that the ROI is uniformly divided into NTimBlcks and NFrqBlcks along the time and frequency dimensions. Figure 6 shows the overlay of ROI patches used here for generating demonstration results.
The ROI overlay in Figure 6 was generated using NTimBlcks = 8 and NFrqBlcks = 9, with a total number of NTim = 55 and NFrq = 12 indices defining the block time-frequency extent. Thus, there are a total of NPtch = 8 × 9 = 72 total time-frequency patches with each patch containing a total of 55 × 12 = 660 elements. Fingerprint statistics (variance, skewness, and kurtosis statistics [30]) were calculated using elements within each of the NPtch = 72 patches, and across all ROI elements as well, such that the resultant Gabor transform DNA fingerprints included a total of NFD-GT = (72 + 1) × 3 = 219 statistical features.

2.4. Multiple Discriminant Analysis (MDA)

The MDA-based discrimination methodology used here was primarily adopted from related work in [4] that exploited passive DNA features and work in [19] that exploited passive DNA features. Both of these earlier works considered the same NDev = 8 WirelessHART adapters shown in Table 1 and used here for demonstration. The fundamental differences (processing, objectives, etc.) between active and passive DNA fingerprinting preclude direct comparison of results presented here with those presented in [4]. While providing a motivational basis for the active DNA fingerprinting work undertaken here, care is taken in making comparisons of results in [19] with those provided here—this is addressed further in Section 3 results. However, the MDA processing is fundamentally the same and limited details for the device discrimination process are presented here for completeness. The reader is referred to [4] for a more detailed description and development of MDA-based device classification and device ID verification.
The NDev = 8 WirelessHART devices in Table 1 were used to perform discrimination assessments for the NMdl = 56 (8-choose-5) model conditions shown in Table 2. As indicated, each model included NCls = 5 classes with each class represented by one of the designated Authentic (A) devices. The remaining three held-out devices were introduced as Counterfeit (C) devices for each of the modeled devices. As adopted from [4], the trained MDA model elements are denoted by (W, μF, σF, μn, Σn) where (1) W is the MDA projection matrix (dimension NFeat × NCls − 1), (2) μF is the input fingerprint mean normalization factor (dimension 1 × NFeat), (3) σF is the input fingerprint standard deviation normalization factor (dimension 1 × NFeat), (4) μn is the projected training class means (dimension 1 × NCLS − 1), and (5) Σn is the training class covariance matrix (dimension NCls − 1 × NCls − 1).

2.5. Device Discrimination

The trained MDA models were used for both Device Classification and Device ID Verification assessments. Given a trained ( W , μF, σF, μn, Σn) MDA model, a fingerprint from an unknown device, denoted as FUnk (dimension 1xNFeat), is projected into the MDA decision space using p Unk = F Unk μ F σ F 1 W (dimension 1 × NDev − 1) [4]. The resultant p Unk is used with a given measure of similarity to generate a test statistic (ZUnk) that is used for making device classification and device ID verification decision. Test statistic ZUnk is a real number that is used to estimate (1) which of the NCls modeled devices the unknown FUnk most closely represents—the fundamental device classification process, and (2) how much the unknown FUnk looks like fingerprints from 1-of-NCls specified devices—the fundamental device ID verification process used for estimating %CDR. Consistent with results in [4,19], the ZUnk test statistics used here were generated from distance-based Euclidean and probability-based Multi-Variate Normal (MVN) measures of similarity. The reader is referred to [4] for a detailed description of test statistic ZUnk generation and its use in making classification and ID verification decisions.
Results for the looks-most-like device classification process are summarized in a confusion matrix format [34]. A representative confusion matrix is shown in Table 3 for MDA classifier testing of an NCls = 5 model using NTst = 565 unknown testing fingerprints per class. The average cross-class percent correct classification (%C) is calculated as the sum of diagonal elements divided by the total number of estimates represented in the matrix (NTst × NCls). The Table 3 results yield an overall %C = [2438/(565 × 5)] × 100 ≈ 86.3%, with individual per-class performances ranging from %CCls = (403/565) × 100 ≈ 71.3% for Class 5 to %CCls = (526/565) × 100 ≈ 93.1% for Class 1.
Table 3 shows that a majority of the classification error is attributable to mutual “confusion” between Class 2 (D2) and Class 5 (D5). The classification estimates in Table 3 effectively represent Monte Carlo trials and 95% Confidence Interval (CI95%) analysis [35] is used throughout the paper when making comparisons and drawing conclusions. The ±CI95% intervals for per-class %CCls are presented in Table 3 by way of example and are based on NTst = 565 Monte Carlo trials.
MDA-based confusion matrix results such as presented in Table 3 were generated for all NMdl = 56 MDA models in Table 2. The presentation of a large number of resultant confusion matrices is avoided in the interest of brevity. As an alternative, the average per-model performances are presented in a %C versus Model ID format to enable comparison across models—an overall cross-model average %C is calculated and presented as well. Results are also presented for average per-device performance in a %C versus Device ID format to enable comparison across devices—an overall cross-device average %C is calculated and presented as well. These results are obtained by considering diagonal confusion matrix entries (correct estimates) for a given device in all models where that device is serving in an authentic role. Thus, the per-device %C averages are based on diagonal per-class entries in 35-of-56 models.
Results for the looks-how-much like device ID verification (counterfeit detection) process were generated using the same NMdl = 56 MDA models used for device classification. Counterfeit device ID verification assessments for a given model are denoted by D i U : D k (counterfeit:authentic), where D i U is one of three non-modeled counterfeit devices and D k is each of the modeled authentic devices. Considering model M54 in Table 2 as an example, there are a total of 3 × 5 = 15 D i U : D k counterfeit assessments performed for all i∈{1, 2, 5} and all k∈{3, 4, 6, 7, 8}. Accounting for all NMdl = 56 models, there were a total of 15 × 56 = 840 individual counterfeit detection assessments completed for estimating %CDR.
The D i U : D k counterfeit ID verification assessments and %CDR estimation are based on a binary accept/reject declaration process. The accept/reject decisions are made on a fingerprint-by-fingerprint basis using the following:
  • Generating a p Unk i = F Unk i μ F σ F 1 W fingerprint projection for each of the NTst fingerprints from the counterfeit D i U device;
  • Calculating the test statistic Z Unk i , k associated with the claimed authentic D k device using each of the counterfeit p Unk i projections;
  • Performing a Z Unk i , k t k V threshold comparison, where t k V is the device-dependent ID verification for the claimed authentic device D k and the ⪋ inequality condition is set as (a) greater than (>) for a higher-is-better match statistic (e.g., MVN probability), or (b) less than (<) for a lower-is-better match statistic (e.g., Euclidean Distance);
  • Making a binary accept/reject declaration based on threshold criteria with (a) an accept (false positive) being an undesirable outcome—counterfeit not detected, and (b) a reject (true negative) being a desirable outcome—counterfeit detected;
  • Calculating %CDR = [(NTstNRej)/NTst] × 100 as an estimate of counterfeit detectability, where NRej is the total number of binary reject decisions.
As with device classification, counterfeit device ID verification results are presented in two formats. The first is a per-model %CDR versus Model ID format. In this case, the %CDR average is based on averaging binary reject decisions for three different counterfeit devices (e.g., D1, D2 and D5 in M54) being compared against all five authentic devices (e.g., D3, D4, D6, D7, D8 in M54)—the per-model %CDR averages are based on a total of 3 × 5 = 15 counterfeit ID verification assessments. The second presentation format includes per-device %CDR versus Device ID. In this case, the presented %CDR average is based on averaging reject decisions for a given device when serving as a counterfeit in 21-of-56 models (e.g., D1 in M36–M56) and being compared against all other devices—the per-device %CDR averages are based on a total of 21 × 5 = 105 counterfeit ID verification assessments.

2.6. Dimensional Reduction Analysis (DRA)

Processing efficiency improvement can be obtained using fingerprints having a reduced number of DRA selected (NDRA) features. The DRA features are selected as a proper subset of the Full-Dimensional (FD) feature set containing NFD features with a goal of minimizing the impact (degradation) in classifier %C performance. DRA feature selection was performed here using classification results of (1) an Artificial Neural Network (ANN) based Learning Vector Quantization (LVQ) process adopted from [12], and (2) ensemble based Random Forest (RndF) process adopted from [14]. While there are certainly other feature selection methods that could be considered, the FD-vs-DRA classification results in Section 3 show that both methods are sufficiently robust for demonstration purposes. Unlike MDA classification, these classifiers provide a direct indication of feature relevance (importance) on the final classification decision.
Relative to MDA, there are increased computation costs with implementing ANN-based and ensemble-based classifiers. Thus, their envisioned use is limited to pre-deployment training and feature selection, with a goal of identifying DRA subsets of sufficiently relevant features that can be used with the more computationally efficient MDA-based discrimination processes described in Section 2.4. The DRA subsets are referred to herein as RndF-selected and LVQ-selected subsets and include a total of NDRA < NFD features. The total percentage of DRA reduction is calculated as [(NFDNDRA)/NFD)/] × 100.
The DRA feature selection process is illustrated in the stem plots provided in Figure 7 and Figure 8. These plots show post-classification RndF and LVQ relevance metrics for the NFD = 219 full-dimensional feature set at SNR = −20 dB. The SNR conditions are noted given that feature relevance and DRA selection are generally SNR dependent, with a greater number of features becoming increasingly relevant as SNR decreases. The post-classification NDRA = 43 selected features are denoted by blue asterisk () markers and the remaining least relevant 176-of-219 features are denoted by the red dot () markers. The 43-of-219 selection represents an approximate 80% reduction.
The top plots in Figure 7 and Figure 8 show the sorted (highest-to-lowest relevance) rank-ordered features with a majority of the NFD = 219 total features having zero to near-zero relevance and contributing minimally to the final classification decision. The bottom plot in each figure shows the unsorted feature relevance and the actual fingerprint feature indices for the most relevant features. The green triangle () markers in each of the bottom plots denote 18 common features identified by both the RndF and LVQ classifiers.
The common features in Figure 7 and Figure 8 were used as an additional jointly selected feature set containing NJoint = 18 features, i.e., the joint set was the intersection of the NDRA,RndF = 43 RndF-selected and NDRA,LVQ = 43 LVQ-selected feature sets. The impact of DRA on classification is determined by rerunning the classifiers and comparing the resultant %CDRA using the NDRA selected subsets with the original %CFD performance. Though not the main emphasis of this section, it is interesting to note that the NDRA-GT = 18 jointly selected feature set identified in Figure 7 and Figure 8 yielded statistically equivalent %CFD and %CDRA performance of %C ≈ 91% for both the RndF and LVQ classifiers—nearly a 92% reduction in the required number of features with no sacrifice in %C performance. What remains to be shown in the Results section is how the MDA classifier performs using the reduced dimensional NDRA,RndF and NDRA,LVQ selected feature sets.
For final discussion on DRA feature selection, it is insightful to consider where the jointly selected RndF features in Figure 7 and LVQ features in Figure 8 were generated from within the Gabor transform domain. This is illustrated in Figure 9, which shows (1) the Gabor transform response and overlaid ROI patches in Figure 6, and (2) numerical values in specific patches to indicate the number of jointly selected features generated from elements within those patches—the vertical side note indicates that 2-of-18 features were generated using all ROI elements. As indicated, all but one of the NJoint = 18 features were generated along diagonal patches where the Gabor transform produced the maximum element-to-element energy concentration changes.
The one off-diagonal feature in Figure 9 (patch centered at m ≈ 50 and k ≈ 65) corresponds to fingerprint feature index number 10 in Figure 7 and Figure 8 and has near-zero relevance for either the RndF or LVQ classifier. A cursory analysis of device discrimination performance with fingerprint feature number 10 removed from the jointly selected NJoint = 18 feature set proved to be inconsequential, i.e., there was no statistically significant change in either device classification or device ID verification performance.

3. Device Discrimination Results

Performance of MDA models representing all NCls = 8 devices is first considered in Section 3.1. These results are provided to (1) highlight benefits for transitioning from 1D time domain to 2D Gabor transform fingerprint features, and (2) demonstrate the effectiveness of RndF and LVQ DRA feature selection for reducing the number of fingerprint features while maintaining acceptable discrimination performance. The benefits of 2D Gabor-based fingerprints and DRA feature selection are carried over into multi-model assessments in Section 3.2. These results include NCls = 5 device multi-model classification and counterfeit ID verification for the NMdl = 56 models in Table 2.

3.1. 1D vs. 2D Classification Performance

The first step toward characterizing discrimination performance and processing efficiency improvement included generating the full-dimensional (FD) TdDna baseline classification results for comparing with GtDna classification results. The TdDna baseline was generated using NFD,TD = 117 TdDna features generated per Section 2.2 and the MDA/ML discrimination process in Section 2.4 for the NCls = 8 WirelessHART devices. The baseline TdDna FD (117) results are presented in Figure 10 and exhibit the typical trends in %C performance for variation in SNR. This includes a range of %C ≈ 100% at higher SNR representing collected response conditions and %C ≈ 1/NCls ≈ 12.5% at the lowest SNR representing random guessing.
The overlaid GtDna comparison results in Figure 10 were generated using the same device SFM responses. Note that each plotted data point in this figure includes the CI95% intervals, and all intervals effectively span the vertical extent of the data markers. These results were generated using NFD-GT = 219 full-dimensional and NDRA ∈ {18, 43, 117} dimensionally reduced feature sets selected using both the RndF and LVQ relevance rank-ordering process detailed in Section 2.6. The additional “Joint” NDRA = 18 feature set was generated as the intersection of RndF and LVQ selected feature sets. Collectively considering all plotted data points in Figure 10 and their corresponding CI95% intervals, GtDna fingerprinting is superior for all but one of the SNR conditions. The one exception occurs under the highest SNR = −6 dB condition, where TdDna and GtDna performances are statistically equivalent for all fingerprint sets considered. Most notably, GtDna DRA performance is statistically equivalent to GtDna FD performance for all DRA subsets under SNR ≥ −20 dB conditions. Benefits for using the NDRA = 18 feature set include (1) a GtDna vs. TdDna performance gain of %CΔ = %CGT − %CTD ≈ 51%, and (2) an approximate 92% reduction in the number of required GtDna features (18 vs. 291) with no trade-off penalty incurred in average cross-class %C performance.

3.2. Multi-Model Discrimination

The benefits of 2D GtDna fingerprinting highlighted in Section 3.1 for the NCls = 8 device models were likewise observed for all NMdl = 56 models in Table 2. That is, statistically equivalent discrimination was achieved using the NFD-GT = 219 and NDRA-GT = 18 feature sets. Thus, representative results for SNR = −20 dB conditions are presented and provide the basis for detailed discussion of multi-model classification and counterfeit ID verification performance. As noted in Section 2.5, both per-model and per-device %C classification and %CDR counterfeit ID verification performances are presented.
The per-device and per-model %C classification performances using the NFD-GT = 219 feature set are shown in Figure 11 and Figure 12, respectively. These include overlaid results for both ED () and MVN () measures of similarity. The corresponding per-device and per-model %C performances for the NDRA-GT = 18 feature set are shown in Figure 13 and Figure 14, respectively. Comparison of Figure 11 through Figure 14 shows that (1) the individual per-device and per-model results for MVN are statistically equivalent to, or better than, ED results for a majority of the individual assessments, with (2) the corresponding cross-device and cross-model averages (dashed lines) indicating that MVN provides overall marginally better performance (0.66% < %CΔ = %CMVN − %CED < 1.28%). The better MVN versus ED performance here is consistent with findings in [4,19].
The NMdl = 56 MDA models used for classification results in Figure 11 through Figure 14 were next used for counterfeit device ID verification assessments. The per-device and per-model performances for the NFD-GT = 219 feature set are shown in Figure 15 and Figure 16, respectively, with overlaid results for both the ED () and MVN () measures. The corresponding per-device and per-model counterfeit ID verification performances for the NDRA-GT = 18 feature set are shown in Figure 17 and Figure 18, respectively. These results collectively embody performance of 56 × 5 × 3 = 840 individual counterfeit device ID verification assessments. In comparing Figure 15 through Figure 18 results, it is evident that (1) the per-device and per-model MVN results are statistically equivalent to, or better than, ED results for a majority of the individual assessments, (2) the cross-device and cross-model %CDR averages (dashed lines) show that the MVN measure is once again marginally superior (%CDRΔ = %CDRMVN − %CDRED ≈ 1.6%), and (3) the per-device performance for the more efficient NDRA-GT = 18 feature set includes 87.0% < %CDR < 99.7% and an overall cross-device average of %CDR ≈ 91.4%. This represents a 92% reduction in the required number of fingerprint features with no trade-off penalty incurred in average %CDR performance.

4. Summary

This work was motivated by the need for providing reliable communications in IoT/IIoT and IR 4.0 systems that are becoming increasingly reliant on automation. WirelessHART communications is one of the key technologies for achieving desired automation objectives and its operational integrity must be ensured. This is addressed using stimulated responses from eight commercial WirelessHART adapters and active 2D Distinct Native Attribute (2D-DNA) fingerprinting. The 2D-DNA fingerprints are generated from Gabor transformed responses and used to train 56 Multiple Discriminant Analysis (MDA) models. Each 5-class model represents five authentic network devices and the three non-modeled devices are introduced as counterfeits to complete 5 × 3 × 56 = 840 individual authentic (modeled) versus counterfeit (non-modeled) ID verification assessments.
Counterfeit Detection Rate (%CDR) is estimated using an MDA-based ID verification process and is the primary metric for characterizing counterfeit detectability. ID verification is performed for the 840 authentic:counterfeit assessments using binary accept/reject threshold testing. The desired outcome is a reject decision (true negative) when a counterfeit device is presented for ID verification. Relative to motivational work in [19], the statistical significance of estimated %CDR is increased here given (1) a factor-of-2 increase in the number of MDA models considered (56 vs. 28), and (2) a factor-of-2.5 increase in the number of counterfeit detection assessments performed (840 vs. 336). Processing efficiency improvement is also achieved using Dimensional Reduction Analysis (DRA) to perform feature down-selection with Random Forest (RndF) [14] and Learning Vector Quantization (LVQ) [12] classifiers. The jointly selected DRA feature set contains only 18-of-291 full-dimensional features (an approximate 92% reduction) and is an important step toward achieving computational efficiency objectives.
The %CDR estimates are based on fingerprint-by-fingerprint ID verification assessments with the best case per-device %CDR of 87.6% < %CDRFD < 99.9% in Figure 15 obtained using the probability-based MVN measure of similarity. Considering all eight devices, this corresponds to average cross-device %CDRFD ≈ 92.5% for the full-dimensional fingerprints. This performance was echoed by the NJoint = 18 DRA performance in Figure 17 that includes per-device 87.0% < %CDRDRA < 99.7% and average cross-device %CDRDRA ≈ 91.4% using only 18-of-291 features. This represents a marginal sacrifice in %CDR performance (%CDRFD − %CDRDRA ≈ 1.1%) with considerable reduction in the number of required fingerprint features (18 vs. 291) and a corresponding boost in computational efficiency.

5. Conclusions

The %CDR > 90% here under small-scale network constraints are believed to be sufficient for motivating supply chain participants (manufacturers, distributors, customers) to consider using active 2D-DNA fingerprint features to certify field kits containing “matched” devices. The certification process could include generation of digital 2D-DNA credentials (e.g., model parameters, device fingerprint features, etc.) for certified package contents that are passed point-to-point as the package traverses the supply chain. The envisioned kit-based protection would require access to active 2D-DNA processing capability (stimulus generator, response collector, fingerprint generator, classifier) at each ID verification check point. The required processing would ideally be hosted in a relatively low cost, small form factor unit such as a Software Defined Radio (SDR) hosting an analog source generator and sufficient Field Programmable Gate Array (FPGA) processing capability to perform required post-collection fingerprint generation and discrimination.
Obtaining computational efficiency amidst limited resource constraints is a common challenge when bringing an experimental method to fruition for operational deployment. Of particular relevance to enhancing the experimental-to-operational transition potential of active 2D-DNA fingerprinting is that the %CDR > 90% here was achieved using 92% fewer fingerprint features (18-of-291) relative to what was used in prior motivational work [19]—this effectively reduces the required fingerprint generation, storage, transfer, and computation requirements. Processing improvements such as this are important, and work continues to further enhance computational efficiency. As a next step, work is underway to address digitization requirements for SDR-FPGA implementation. This specifically includes considering 2D surface quantization that must occur prior to fingerprint generation and the effect of bit depth on device discrimination.

Author Contributions

Conceptualization, W.H.M. and M.A.T.; Data curation, W.H.M.; Formal analysis, W.H.M. and M.A.T.; Investigation, W.H.M.; Methodology, W.H.M., M.A.T. and R.F.M.; Project administration, M.A.T. and R.F.M.; Resources, M.A.T. and R.F.M.; Supervision, M.A.T.; Validation, W.H.M. and M.A.T.; Graphic Visualization, W.H.M. and M.A.T.; Writing—original draft, M.A.T.; Writing—review and editing, W.H.M., M.A.T., and R.F.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by support funding received from the Spectrum Warfare Division, Sensors Directorate, U.S. Air Force Research Laboratory, Wright-Patterson AFB, Dayton OH, during U.S. Government Fiscal Years 2019–2022.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The experimentally collected WirelessHART data supporting the reported results were not approved for public release at the time of paper submission. Requests for release of these data to a third party should be directed to the corresponding author. Data distribution to a third party will be made on a request-by-request basis and are subject to public affairs approval.

Acknowledgments

The views and conclusions contained in this document are those of the authors and should not be interpreted as representing the official policies, either expressed or implied, of the United States Air Force or the U.S. Government. This paper is approved for public release, Case Number 88ABW-2022-0478.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

Abbreviations

The following abbreviations are used throughout the manuscript:
%CAverage Cross-Class Percent Correct Classification
ANNArtificial Neural Network
%CDRCounterfeit Detection Rate Percentage
CI95%95% Confidence Interval
EDEuclidean Distance
FDFull Dimensional
DNADistinct Native Attribute
DRADimensional Reduction Analysis
FPGAField Programmable Gate Array
GTGabor Transform
GtDnaGabor Transform DNA
GSpsGiga-Samples Per Second
IDIdentity/Identification
IoTInternet of Things
IIoTIndustrial Internet of Things
IR 4.0Industrial Revolution 4.0
LVQLearning Vector Quantized
MDAMultiple Discriminant Analysis
MHzMegahertz
MSpsMega-Samples Per Second
MVNMultivariate Normal
RFIDRadio Frequency Identification
RndFRandom Forest
SDRSoftware Defined Radio
SFMStepped Frequency Modulated
TDTime Domain
TdDnaTime Domain DNA
HARTHighway Addressable Remote Transducer

References

  1. FieldComm Group. WirelessHART: Proven and Growing Technology with a Promising Future; Global Control; FieldComm Group: Austin, TX, USA, 2018; Available online: https://tinyurl.com/fcgwirelesshartglobalcontrol (accessed on 22 June 2022).
  2. Devan, P.A.M.; Hussin, F.A.; Ibrahim, R.; Bingi, K.; Khanday, F.A. A Survey on the Application of WirelessHART for Industrial Process Monitoring and Control. Sensors 2021, 21, 4951. [Google Scholar] [CrossRef] [PubMed]
  3. Majid, M.; Habib, S.; Javed, A.R.; Rizwan, M.; Srivastava, G.; Gadekallu, T.R.; Lin, J.C.W. Applications of Wireless Sensor Networks and Internet of Things Frameworks in Industry Revolution 4.0: A Systematic Literature Review. Sensors 2022, 22, 2087. [Google Scholar] [CrossRef] [PubMed]
  4. Rondeau, C.M.; Temple, M.A.; Betances, J.A.; Schubert Kabban, C.M. Extending Critical Infrastructure Element Longevity Using Constellation-Based ID Verification. J. Comput. Secur. 2020, 100, 102073. [Google Scholar] [CrossRef]
  5. Cyber Security and Infrastructure Agency (CISA). Cybersecurity and Physical Security Convergence. 2021. Available online: https://www.cisa.gov/cybersecurity-and-physical-security-convergence (accessed on 22 June 2022).
  6. Soltanieh, N.; Norouzi, Y.; Yang, Y.; Karmakar, N.C. A Review of Radio Frequency Fingerprinting Techniques. IEEE J. Radio Freq. Identif. 2020, 4, 222–233. [Google Scholar] [CrossRef]
  7. Suski, W.C., II; Temple, M.A.; Mendenhall, M.J.; Mills, R.F. Using Spectral Fingerprints to Improve Wireless Network Security. In Proceedings of the IEEE Global Telecommunications Conference (GLOBECOM), New Orleans, LA, USA, 30 November–4 December 2008. [Google Scholar] [CrossRef]
  8. Reising, D.R.; Temple, M.A.; Mendenhall, M.J. Improved Wireless Security for GMSK-based Devices Using RF Fingerprinting. Int. J. Electron. Security Digit. Forensics 2010, 3, 41–59. [Google Scholar] [CrossRef]
  9. Reising, D.R.; Temple, M.A.; Mendenhall, M.J. Improving Intracellular Security Using Air Monitoring with RF Fingerprints. In Proceedings of the IEEE Wireless Communication and Networking Conference (WCNC), Sydney, Australia, 18–21 April 2010. [Google Scholar] [CrossRef] [Green Version]
  10. Williams, M.D.; Temple, M.A.; Reising, D.R. Augmenting Bit Level Network Security Using Physical Layer RF-DNA Fingerprinting. In Proceedings of the IEEE Global Telecommunications Conference (GLOBECOM), Miami, FL, USA, 6–10 December 2010. [Google Scholar] [CrossRef] [Green Version]
  11. Lopez, J.; Liefer, N.C.; Busho, C.R.; Temple, M.A. Enhancing Critical Infrastructure and Key Resources (CIKR) Level-0 Physical Process Security Using Field Device Distinct Native Attribute Features. IEEE Trans. Inf. Forensics Secur. 2018, 13, 1215–1229. [Google Scholar] [CrossRef]
  12. Reising, D.R.; Temple, M.A.; Jackson, J.A. Authorized and Rogue Device Discrimination Using Dimensionally Reduced RF-DNA Fingerprints. IEEE Trans. Inf. Forensics Secur. 2015, 10, 1180–1192. [Google Scholar] [CrossRef]
  13. Reising, D.R.; Temple, M.A. WiMAX Mobile Subscriber Verification Using Gabor-Based RF-DNA Fingerprints. In Proceedings of the IEEE International Conference on Communications (ICC), Ottawa, ON, Canada, 10–15 June 2012. [Google Scholar] [CrossRef] [Green Version]
  14. Rondeau, C.M.; Temple, M.A.; Schubert Kabban, C.M. TD-DNA Feature Selection for Discriminating WirelessHART IIoT Devices. In Proceedings of the 53rd Hawaii International Conference on System Sciences (HICSS), Maui, HI, USA, 7–10 January 2020; Available online: https://scholarspace.manoa.hawaii.edu/bitstreams/35252979-27c2-4ae0-b8fb-35529f731e5a/download (accessed on 22 June 2022).
  15. Talbot, C.M.; Temple, M.A.; Carbino, T.J.; Betances, J.A. Detecting Rogue Attacks on Commercial Wireless Insteon Home Automation Systems. J. Comput. Secur. 2018, 74, 296–307. [Google Scholar] [CrossRef]
  16. Stone, S.J.; Temple, M.A. RF-Based Anomaly Detection for PLCs in Critical Infrastructure Apps. Int. J. Crit. Infrastruct. Prot. 2012, 5, 66–73. [Google Scholar] [CrossRef]
  17. Stone, S.J.; Temple, M.A.; Baldwin, R.O. Detecting Anomalous PLC Behavior Using RF-Based Hilbert Transform Features and a Correlation-Based Verification Process. Int. J. Crit. Infrastruct. Prot. 2015, 9, 41–51. [Google Scholar] [CrossRef]
  18. Lukacs, M.W.; Zeqolari, A.J.; Collins, P.J.; Temple, M.A. RF-DNA Fingerprinting for Antenna Classification. IEEE Antennas Wirel. Propag. Lett. 2015, 14, 1455–1458. [Google Scholar] [CrossRef]
  19. Maier, M.J.; Hayden, H.S.; Temple, M.A.; Betances, J.A.; Fickus, M.C. Improving WirelessHART Trust in Critical Infrastructure Applications. In Critical Infrastructure Protection XVI; Springer: Cham, Switzerland, 2020. [Google Scholar]
  20. Paul, A.J.; Collins, P.J.; Temple, M.A. Enhancing Microwave System Health Assessment Using Artificial Neural Networks. IEEE Antennas Wirel. Propag. Lett. 2019, 18, 2230–2234. [Google Scholar] [CrossRef]
  21. Voetberg, B.; Carbino, T.; Temple, M.; Buskohl, P.; Denault, J.; Glavin, N. Evolution of DNA Fingerprinting for Discriminating Conductive Ink Specimens. In Proceedings of the Government Microcircuit Applications & Critical Technology Conference (GOMACTech), Albuquerque, NM, USA, 25–28 March 2019. [Google Scholar]
  22. FieldComm Group. WirelessHART User Case Studies; Technical Report; FieldComm Group: Austin, TX, USA, 2019; Available online: https://tinyurl.com/fcgwirelesscs (accessed on 22 June 2022).
  23. Yang, K.; Forte, D.; Tehranipoor, M.M. CDTA: A Comprehensive Solution for Counterfeit Detection, Traceability, and Authentication in the IoT Supply Chain. ACM Trans. Des. Autom. Electron. Syst. 2017, 22, 1–31. [Google Scholar] [CrossRef]
  24. Raut, R.D.; Gotmare, A.; Narkhede, B.E.; Govindarajan, U.H.; Bokade, S.U. Enabling Technologies for Industry 4.0 Manufacturing and Supply Chain: Concepts, Current Status, and Adoption Challenges. IEEE Eng. Manag. Rev. 2020, 48, 83–102. [Google Scholar] [CrossRef]
  25. Siemens. WirelessHART Adapter, SITRANS AW210, 7MP3111, User Manual; Siemens: Munich, Germany, 2012; Available online: https://tinyurl.com/yyjbgybm (accessed on 22 June 2022).
  26. Pepperl+Fuchs. WHA-BLT-F9D0-N-A0-*, WirelessHART Adapter, Manual. Available online: https://tinyurl.com/pepplusfucwirelesshart (accessed on 22 June 2022).
  27. Keysight Technologies. PNA Family Microwave Network Analyzer (N522x/3x/4xB), Configuration Guide, Doc ID: 5992-1465EN. 10 September 2021. Available online: https://www.keysight.com/us/en/assets/7018-05185/configuration-guides/5992-1465.pdf (accessed on 22 June 2022).
  28. LeCroy. WaveMaster® 8 Zi-A Series: 4 GHz-45GHz Doc ID: WM8Zi-A-DS-09May11. 2011. Available online: https://docs.rs-online.com/035e/0900766b8127e31c.pdf (accessed on 22 June 2022).
  29. Granados-Lieberman, D.; Valtierra-Rodriguez, M.; Morales-Hernandez, L.A.; Romero-Troncoso, R.J.; Roque, A. Osornio-Rios, R.A. A Hilbert Transform-Based Smart Sensor for Detection, Classification, and Quantification of Power Quality Disturbances. Sensors 2013, 13, 5507–5527. [Google Scholar] [CrossRef] [Green Version]
  30. Soberon, A.; Stute, W. Assessing Skewness, Kurtosis and Normality in Linear Mixed Models. J. Multivar. Anal. 2017, 161, 123–140. [Google Scholar] [CrossRef] [Green Version]
  31. Cho, S.; Jang, G.; Kwon, S. Time-Frequency Analysis for Power-Quality Disturbances via the Gabor-Wigner Transform. IEEE Trans. Power Deliv. 2010, 25, 494–499. [Google Scholar] [CrossRef]
  32. Zibulski, M.; Zeevi, Y. Oversampling in the Gabor Scheme. IEEE Trans. Signal Process. 1993, 41, 2679–2687. [Google Scholar] [CrossRef]
  33. Bastiaans, M.J.; Geilen, M.C.W. On the Discrete Gabor Transform and The Discrete Zak Transform. Signal Process. 1996, 49, 151–166. [Google Scholar] [CrossRef]
  34. Tharwat, A. Classification Assessment Methods. Appl. Comput. Inform. 2020, 17, 168–192. [Google Scholar] [CrossRef]
  35. Park, H.; Leemis, L.M. Ensemble Confidence Intervals for Binomial Proportions. Stat. Med. 2019, 38, 3460–3475. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Experimental setup similar to that used in [19] for collecting WirelessHART DUT responses used for active DNA fingerprint generation.
Figure 1. Experimental setup similar to that used in [19] for collecting WirelessHART DUT responses used for active DNA fingerprint generation.
Sensors 22 04906 g001
Figure 2. Illustration of pre-fingerprint generation processing effects on a representative WirelessHART response. The final down-converted bandpass filtered signal is used for subsequent time domain DNA (TdDna) and Gabor transform DNA (GtDna) generation.
Figure 2. Illustration of pre-fingerprint generation processing effects on a representative WirelessHART response. The final down-converted bandpass filtered signal is used for subsequent time domain DNA (TdDna) and Gabor transform DNA (GtDna) generation.
Sensors 22 04906 g002
Figure 3. Normalized time domain magnitude of a representative DUT stimulated output showing the full ROI (bounded by black dashed lines) and the NRgn = 12 subregion boundaries used for generating statistical time domain DNA fingerprint features.
Figure 3. Normalized time domain magnitude of a representative DUT stimulated output showing the full ROI (bounded by black dashed lines) and the NRgn = 12 subregion boundaries used for generating statistical time domain DNA fingerprint features.
Sensors 22 04906 g003
Figure 4. Normalized time-frequency spectrogram of the same representative DUT stimulated output used to generate the Figure 3 time domain magnitude response.
Figure 4. Normalized time-frequency spectrogram of the same representative DUT stimulated output used to generate the Figure 3 time domain magnitude response.
Sensors 22 04906 g004
Figure 5. Normalized time-frequency Gabor transform response of the same representative DUT stimulated output used to generate the Figure 3 time domain magnitude response.
Figure 5. Normalized time-frequency Gabor transform response of the same representative DUT stimulated output used to generate the Figure 3 time domain magnitude response.
Sensors 22 04906 g005
Figure 6. Overlay of the Gabor transform magnitude response from Figure 5 with the selected 2D fingerprinting ROI patch boundaries overlaid.
Figure 6. Overlay of the Gabor transform magnitude response from Figure 5 with the selected 2D fingerprinting ROI patch boundaries overlaid.
Sensors 22 04906 g006
Figure 7. Stem plots illustrating DRA feature selection for RndF classification using the NFD = 219 full-dimensional feature set. The blue asterisk () markers denote the most relevant NDRA,RndF = 43 features, the red dot () markers denote the least relevant 176-of-219 features, and the green triangle () markers identify selected features shared in common with LVQ in Figure 8.
Figure 7. Stem plots illustrating DRA feature selection for RndF classification using the NFD = 219 full-dimensional feature set. The blue asterisk () markers denote the most relevant NDRA,RndF = 43 features, the red dot () markers denote the least relevant 176-of-219 features, and the green triangle () markers identify selected features shared in common with LVQ in Figure 8.
Sensors 22 04906 g007
Figure 8. Stem plots illustrating DRA feature selection for LVQ classification using the NFD = 219 full-dimensional feature set. The blue asterisk () markers denote the most relevant NDRA,LVQ = 43 features, the red dot () markers denote the least relevant 176-of-219 features, and the green triangle () markers identify selected features shared in common with RndF in Figure 7.
Figure 8. Stem plots illustrating DRA feature selection for LVQ classification using the NFD = 219 full-dimensional feature set. The blue asterisk () markers denote the most relevant NDRA,LVQ = 43 features, the red dot () markers denote the least relevant 176-of-219 features, and the green triangle () markers identify selected features shared in common with RndF in Figure 7.
Sensors 22 04906 g008
Figure 9. Gabor transform and ROI patch overlay from Figure 6 with numerical values added to indicate the number of jointly selected RndF and LVQ features generated within the patches. The numbers identify 1-of-18, 2-of-18, or 3-of-18 features from among the NJoint = 18 features identified by the green triangle () markers in Figure 7 and Figure 8.
Figure 9. Gabor transform and ROI patch overlay from Figure 6 with numerical values added to indicate the number of jointly selected RndF and LVQ features generated within the patches. The numbers identify 1-of-18, 2-of-18, or 3-of-18 features from among the NJoint = 18 features identified by the green triangle () markers in Figure 7 and Figure 8.
Sensors 22 04906 g009
Figure 10. Composite MDA/ML classification performance for NCls = 8 devices using indicated feature sets. GtDna feature improvement is indicated by an SNRΔ = SNRGT − SNRTD ≈ 7 dB “gain” at %C ≈ 90% and %CΔ = %CGT − %CTD ≈ 51% improvement at SNR = −20 dB.
Figure 10. Composite MDA/ML classification performance for NCls = 8 devices using indicated feature sets. GtDna feature improvement is indicated by an SNRΔ = SNRGT − SNRTD ≈ 7 dB “gain” at %C ≈ 90% and %CΔ = %CGT − %CTD ≈ 51% improvement at SNR = −20 dB.
Sensors 22 04906 g010
Figure 11. Per-device MDA classification performance for the NMdl = 56 models in Table 2. Results are for the full-dimensional NFD-GT = 219 feature set generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 11. Per-device MDA classification performance for the NMdl = 56 models in Table 2. Results are for the full-dimensional NFD-GT = 219 feature set generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g011
Figure 12. Per-model MDA classification performance for the NMdl = 56 models in Table 2. Results are for full-dimensional NFD-GT = 219 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 12. Per-model MDA classification performance for the NMdl = 56 models in Table 2. Results are for full-dimensional NFD-GT = 219 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g012
Figure 13. Per-device MDA classification performance for the NMdl = 56 models in Table 2. Results are for jointly selected NDRA-GT = 18 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 13. Per-device MDA classification performance for the NMdl = 56 models in Table 2. Results are for jointly selected NDRA-GT = 18 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g013
Figure 14. Per-model MDA classification performance for the NMdl = 56 models in Table 2. Results are for jointly selected NDRA-GT = 18 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 14. Per-model MDA classification performance for the NMdl = 56 models in Table 2. Results are for jointly selected NDRA-GT = 18 features generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g014
Figure 15. Per-device ID verification performance for full-dimensional NFD-GT = 219 features using the NMdl = 56 MDA models used for generating %C results in Figure 11. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 15. Per-device ID verification performance for full-dimensional NFD-GT = 219 features using the NMdl = 56 MDA models used for generating %C results in Figure 11. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g015
Figure 16. Per-model ID verification performance for full-dimensional NFD-GT = 219 features using the NMdl = 56 MDA models used for generating %C results in Figure 12. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 16. Per-model ID verification performance for full-dimensional NFD-GT = 219 features using the NMdl = 56 MDA models used for generating %C results in Figure 12. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g016
Figure 17. Per-device ID verification performance for jointly selected NDRA-GT = 18 features using the NMdl = 56 MDA models used for generating %C results in Figure 13. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 17. Per-device ID verification performance for jointly selected NDRA-GT = 18 features using the NMdl = 56 MDA models used for generating %C results in Figure 13. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g017
Figure 18. Per-model ID verification performance for jointly selected NDRA-GT = 18 features using the NMdl = 56 MDA models used for generating %C results in Figure 14. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Figure 18. Per-model ID verification performance for jointly selected NDRA-GT = 18 features using the NMdl = 56 MDA models used for generating %C results in Figure 14. Results generated under SNR = −20 dB conditions using ED () and MVN () measures of similarity.
Sensors 22 04906 g018
Table 1. Selected details for NDev = 8 WirelessHART adapters used for demonstration.
Table 1. Selected details for NDev = 8 WirelessHART adapters used for demonstration.
Device IDDevice LabelFirmwareSerial Number
D1Siemens AW210198003095
D2Siemens AW210200003159
D3Siemens AW210198003097
D4Siemens AW210200003150
D5Pepperl+Fuchs Bullet2001A32DA
D6Pepperl+Fuchs Bullet2001A32B3
D7Pepperl+Fuchs Bullet2001A3226
D7Pepperl+Fuchs Bullet2001A32A4
Table 2. Device assignments for the NMdl = 56 models the NDev = 8 adapters serving in the indicated Authentic (A) or Counterfeit (C) roles.
Table 2. Device assignments for the NMdl = 56 models the NDev = 8 adapters serving in the indicated Authentic (A) or Counterfeit (C) roles.
Model IDD1D2D3D4D5D6D7D8
M1AAAAACCC
M2AAAACACC
M3AAAACCAC
M54CCAACAAA
M55CCACAAAA
M56CCCAAAAA
Table 3. Classification confusion matrix for model M1 (D1, D2, D3, D4, D5) in Table 2 showing NCls = 5 discrimination performance for SNR = −20 dB conditions.
Table 3. Classification confusion matrix for model M1 (D1, D2, D3, D4, D5) in Table 2 showing NCls = 5 discrimination performance for SNR = −20 dB conditions.
Called Class
Input ClassClass 1Class 2Class 3Class 4Class 5
Class 1526010290
Class 20438270100
Class 351753904
Class 432015320
Class 50152100403
%CCls93.1%77.5%95.4%94.2%71.3%
±CI95%2.1%3.4%1.7%1.9%3.7%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Mims, W.H.; Temple, M.A.; Mills, R.F. Active 2D-DNA Fingerprinting of WirelessHART Adapters to Ensure Operational Integrity in Industrial Systems. Sensors 2022, 22, 4906. https://doi.org/10.3390/s22134906

AMA Style

Mims WH, Temple MA, Mills RF. Active 2D-DNA Fingerprinting of WirelessHART Adapters to Ensure Operational Integrity in Industrial Systems. Sensors. 2022; 22(13):4906. https://doi.org/10.3390/s22134906

Chicago/Turabian Style

Mims, Willie H., Michael A. Temple, and Robert F. Mills. 2022. "Active 2D-DNA Fingerprinting of WirelessHART Adapters to Ensure Operational Integrity in Industrial Systems" Sensors 22, no. 13: 4906. https://doi.org/10.3390/s22134906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop