Next Article in Journal
Spatio-Trajectorial Optical Flow for Higher-Order Deformation Analysis in Solid Experimental Mechanics
Next Article in Special Issue
Sequential Fusion Filter for State Estimation of Nonlinear Multi-Sensor Systems with Cross-Correlated Noise and Packet Dropout Compensation
Previous Article in Journal
A Hybrid Trust Model against Insider Packet Drop Attacks in Wireless Sensor Networks
Previous Article in Special Issue
A Hybrid Indoor Altimetry Based on Barometer and UWB
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

extendGAN+: Transferable Data Augmentation Framework Using WGAN-GP for Data-Driven Indoor Localisation Model

1
Singtel Cognitive and Artificial Intelligence Lab (SCALE@NTU), Nanyang Technological University, 50 Nanyang Avenue, Singapore 639798, Singapore
2
School of Computer Science and Engineering, Nanyang Technological University, 50 Nanyang Avenue, Singapore 639798, Singapore
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(9), 4402; https://doi.org/10.3390/s23094402
Submission received: 30 January 2023 / Revised: 11 April 2023 / Accepted: 21 April 2023 / Published: 30 April 2023
(This article belongs to the Special Issue Multi‐Sensors for Indoor Localization and Tracking)

Abstract

:
For indoor localisation, a challenge in data-driven localisation is to ensure sufficient data to train the prediction model to produce a good accuracy. However, for WiFi-based data collection, human effort is still required to capture a large amount of data as the representation Received Signal Strength (RSS) could easily be affected by obstacles and other factors. In this paper, we propose an extendGAN+ pipeline that leverages up-sampling with the Dirichlet distribution to improve location prediction accuracy with small sample sizes, applies transferred WGAN-GP for synthetic data generation, and ensures data quality with a filtering module. The results highlight the effectiveness of the proposed data augmentation method not only by localisation performance but also showcase the variety of RSS patterns it could produce. Benchmarking against the baseline methods such as fingerprint, random forest, and its base dataset with localisation models, extendGAN+ shows improvements of up to 23.47%, 25.35%, and 18.88% respectively. Furthermore, compared to existing GAN+ methods, it reduces training time by a factor of four due to transfer learning and improves performance by 10.13%.

1. Introduction

With the rise of multi-floor developments, urbanisation, and smart cities increasing the living space and building complexity, indoor localisation and navigation has played a significant role in assisting users to their destination indoors. Its application has been deployed in commercial buildings [1,2], carparks [3], the health service [4], public transport [5,6], etc.
Indoor localisation has remained an active research area over the past few years. There has been a shift from using extra hardware to the hybrid techniques of the existing wireless technologies and data-driven machine learning, in which the fingerprint approach, with the WiFi-received signal strength (RSS), has high adoption rate for the indoor positioning system because of its effectiveness, simplicity, and minimal requirements for additional setup or hardware as it utilises the available pre-installed WiFi access points (APs). However, the RSS varies due to obstacles, multi-path phenomena, and fading effects, resulting in arbitrary fluctuations. Hence, to achieve the desired localisation accuracy, the machine learning approach [1,7,8] and deep learning approach [9] were explored. As a result, the deep learning approach has shown potential in learning the complex features from the training data, requires less effort in feature engineering, and achieves higher accuracy [10,11]. In summary, the desired positioning accuracy could be achieved using the deep learning model when the RSS shadowing and fading challenges, that affect the correction of RSS value and its node location, are addressed [12].
Although adopting the neural network prediction model means benefiting from its ability to learn without a well-constructed features/domain knowledge and to tolerate fault or corruption in the network, selecting the optimal architecture plays an important role improving performance of the model. For indoor application, the RSS-based model architecture started with the multi-layer perceptron [9] and rapidly adopted targeted architectures to extract relevant features by the feature reconstruction autoencoder [13] and/or spatially-aware convolution neural network (CNN) [14]. The CNN model has been shown to be an effective model for the RSS-based localisation application since it leverages the relationship of the wireless heatmap and the corresponding location. The study [15] took a step forward to incorporate time-dependent factors by inputting time-series RSS to the CNN model in order to address the randomness and noise caused by the RSS fluctuation issue. However, the CNN architecture was suggested based on empirical considerations. Thereupon, there is research interest to develop strategies for the model selection. It is seen in the adoption of well-performed CNN architecture. For instance, Ref. [16] introduces a VGG-block-building framework via heuristic hyperparameter search to optimise the the performance using CVTuner. In this study, various techniques ranging from the traditional machine learning method to the use of a CNN-based model for feature extraction to benchmark the performance.
A known drawback of training deep learning models is the prerequisite of sufficient good quality data, the pivotal component. For indoor localisation using RSS data, the site visit and data collection is essential. In other words, the larger the site, the more effort required to collect data for the premise, which is time consuming and labor-intensive [17,18,19]. For instance, Rashmi et al. reported having to collect at least 15 hours’ worth of data over a span of 7 days for the training of a CNN model [20]. To mitigate the data challenge, synthetic data are being generated to train data. Data augmentation methods for indoor localisation range from sampling with a defined distribution [2,20] to training a data-driven Generative Adversarial Network [21]. Although data aggregation using a defined distribution evidently improves the performance of the prediction model, there is a question of which distribution models the fluctuating nature of RSS and its environment. Therefore, more focus has been on the data-driven approach. In particular, a conditional GAN was introduced to perform data augmentation the targeted area with the received signal strength and its cell-ID location as input [22]. Even though the one-for-all augmenter learns the complex structure by using one-hot encoded location as the GAN’s auxiliary data, the training could become a complex one as the reference point grows from the reported 25 reference points. Wafa et al. took similar approach to generate data for the entire out (a one-for-all augmenter) and proposed the Selective Generative Adversarial Network with DNN model (Selective-SS-GAN) to predict pseudo-labelling at the unseen location. Meanwhile, AF-DCGAN trained a GANs per location, aiming to learn the data distribution for each unique location [23]. However, instead of RSS data, the method is used to generate additional amplitude feature maps of the Channel State Information (CSI) which requires additional hardware modification. Nonetheless, the method proposed for the CSI amplitude may not be applicable for RSS data as RSS characterises only the coarse value at the receiver and displays high variability. In summary, the exploration of data augmentation is crucial to the localisation performance where data-driven approach GANs show potential in augmenting better quality data. On the other hand, more generated data do not always improve the performance. The localisation’s performance in [22] stalemates when the generated data exceed an optimal point. Albeit proposing an augmentation method, the recommended amount of synthetic data were not addressed.
Addressing the aforementioned challenges in constructing the RSS fingerprint with synthetic data for the localisation model, the main contributions of this paper are the following:
  • Introduce an end-to-end recommendation to improve the data-driven localisation model’s performance by proposing a data augmentation pipeline and residual-network adaptation for the feature extraction of the localisation model.
  • Propose an extendGAN+ pipeline to generate synthetic data for the localisation model even with an extremely small training dataset. The approach leverage on the combination of up-sampling with the Dirichlet distribution at the location with below-threshold data points, transferred WGAN-GP and a filtering module for quality control. The use of transfer learning was to reduce the training time and computational resources while the WGAN-GP model was organically trained once at the location with maximum data point. In addition to the proposed method, we provide a practical recommendation to set the amount of the augmented data as more generated data do not always improve the performance.
  • Conduct experiments with the publicly accessible dataset (UJIndoorLoc [24]) and self-collected data at the building complex to evaluate the proposed method to provide better insights of the difference in data quality and localisation performance. The state-of-the-art network architectures for indoor localisation were used for the additional performance evaluation.

2. Methodology

2.1. Overview

In this section, we propose the transferable workflow that includes data augmentation and localisation models (Figure 1).
The extendGAN+ framework leverage on the WGAN-GP to create synthetic RSS data as a WGAN-GP model is trained as per unique location. Firstly, the training dataset is sorted for the unique location with most data points ( l o c m a x where d a t a ( R P m a x ) = m a x ( d a t a ( R P i ) ) ) and set data threshold ( d m a x d a t a ( R P m a x ) ). Upon selecting l o c m a x , the WGAN-GP model is trained from scratch to produce augmented data. The work of  [22] has shown that the amount of generated data has an impact on the localisation accuracy, peaked and saturated. The 1:1 ratio is chosen according to our empirical study on the impact of generated data; hence, we are generating RSS in d m a x amount using the trained WGAN-GP. For another unique location ( l o c i ) with data point ( d i ), where l o c i l o c m a x , the Dirichlet data aggregation method is used to up-sample d i to the size of d m a x . Subsequently, WGAN-GP model at l o c m a x is transferred to l o c i to generate augmented data of the above mentioned 1:1 ratio. In summary, at l o c i , we obtain the total data composition of original data ( d i ), Dirichlet up-sampling data ( d i d i r i c h l e t = d m a x = d i ) and WGAN-GP-generated data ( d i W G A N G P = d m a x ). It is to be noted that the data used for augmentation are the training dataset, while keeping the test dataset untouched. Subsequently, the localisation model could be trained with the new training dataset.

2.2. Data Augmentation–extendGAN+

For data-driven indoor localisation, the RSS fingerprint is crucial in determining the location. However, the fingerprint varies due to disruptions such as the non-line-of-sight (NLOS) effect. More data collection sessions are needed at different times of days or weeks, in addition to being taking in various changing environment scenarios such as crowd size and environment. It mitigates the representation challenge, albeit at the cost of inefficiency in time and labour. Thus, data-augmentation methods were introduced.
In the previous study [25], we proposed GAN+ data augmentation using the GAN+ framework, which combines the Dirichlet and GAN augmentation techniques, to generate augmented RSS achieving improved performance even with a small training database. The augmentation was applied to one location at a time by aggregating data with Dirichlet distribution and train the GAN model. The synthetic data were then filtered to remove outliers. Although obtaining a variation of RSS representation, training a GAN model per location is time inefficient, especially when scaling up the testbed. Moreover, each location contains a varying number of data points. Thus, setting a static parameter does not address in-label imbalancing. Therefore, in this study, we propose an improved GAN+ (extendGAN+) framework with a localisation model.

2.3. GAN to WGAN-GP

Among the other generative models, Generating Adversarial Networks (also known as GANs) represent a state-of-the-art deep learning framework which does not involve the maximum likelihood estimation, and the generator is trained without having seen the real data. There have been a variety of applications using GANs and their variations in numerous fields such as the medical field [26], human face images [27], maps [28], etc. [29,30]. However, GANs suffer their own setbacks such as non-convergence and mode collapsing. The vanishing gradient, resulting in non-convergence, refers to the situation where the gradient update of the generator’s weight is close to zero, such that its weight would not be updated effectively, while mode collapsing focuses on the produced output to be of random yield rather than specific to the target output.
Mitigating the mentioned challenges, the Wasserstein Generative Adversarial Network with Gradient Penalty (WGAN-GP) presented an effective solution by incorporating Wasserstein distance to the critics loss function [31,32]. The Wasserstein distance measures the distance between two probability distributions by calculating the minimum cost of transporting mass in converting from one data distribution to another (Equation (1)). Furthermore, the Wasserstein distance is derived in Equation (2) where f satisfies constraint ( f ( x 1 ) f ( x 2 ) x 1 x 2 ) to be a 1-Lipschitz function and s u p refers to the supremum of a set. To enforce the 1-Lipschitz function, the gradient penalty was used such that gradient of the critics would have the unit norm (with λ E x ^ P x ^ [ ( x ^ f ( x ^ ) 2 1 ) 2 ] in Equation (3)).
W ( P r , P g ) = i n f γ Π ( P r , P g ) E ( x , y γ ) [ x y ]
W ( P r , P g ) = s u p f L 1 E x P r [ f ( x ) ] E x P g [ f ( x ) ]
L c r i t i c = E x ˜ P g [ f ( x ˜ ) ] E x P r [ f ( x ) ] + λ E x ^ P x ^ [ ( x ^ f ( x ^ ) 2 1 ) 2 ]
With the change of loss function to the continuous metric, WGAN-GP could quantify the performance better by focusing on the image quality rather than the binary classification for whether the generator network could fool the discriminator network with the Sigmoid function. As a result, WGAN-GP yields the training stability with no sign of mode-collapsing and the critics can still learn when the generator performs well by the change of loss function. Moreover, the gradient penalty reduces the efforts to perform hyper-parameter tuning, as compared to WGAN without gradient penalty having to set weight-clipping parameter c.

2.4. Upsampling with Dirichlet Data Aggregation

To compensate for the imbalance of training data towards the location of the majority of data points, upsampling is adopted to synthetically generate data and add bias into the dataset. As RSS variations in nature are highly affected by obstacles and the multipath and fading effects, it requires more data collection effort to capture the fingerprint representation of the location. Thus, upsampling RSS data is the right fit to prepare the data for the WGAN-GP model, compensating for the lack of data at certain location. In fact, the upsampling process for RSS data is commonly performed via permutation of the fingerprints by repetition or generating random numbers from a predefined distribution, such as uniform or normal distributions, using existing data.
In this study, we adopted the Dirichlet distribution to upsample data for the locations that lack data to train with WGAN-GP ( d i < d m a x ). The Dirichlet distribution is commonly used in the area of text mining and text network analysis to fit a topic model as it allows a mixture of topics and words to overlap (rather than being repeated in discrete group). Differing from uniform or normal distributions, the Dirichlet distribution provides versatility, which characterises the random variability, and is able to govern the shape of the distribution by its nature being a multivariate probability distribution. A Dirichlet distribution [33], represented by a vector α , has the probability density given by Equation (4).
ρ ( x ) = Z i x i α i 1
where random variable x i is distributed according to the Dirichlet distribution D i r ( α ) ( x i D i r ( α ) ) if density function ρ ( x ) holds. α i = { α 1 , α 2 , . . . α n } > 0 is a vector that holds the parameter of the distribution while Z is the generalised multinomial coefficient and n is the number of samples.
Therefore, in this study, instead of augmenting each fingerprint/record (by row) with random numbers generated from a pre-defined uniform or normal distribution, we use a Dirichlet-distributed random variable for each access point (by column). The Dirichlet augmentation scheme requires at least 2 records from a given location (let the number of records at the unique location be N) (shown in Algorithm 1). Additionally, let the total number of access points be t and the number of datapoints in a unique location be N. In this scheme, new RSS fingerprints are generated by finding the weighted sum of the RSSI based on the individual APs. In other words, assigning random weight ( W = [ W A P 1 , W A P 2 , . . . , W A P t ] ) to each of these N records consisting of t measurements, such that the weights sum up to 1. The reading of an access point in the new sample can be obtained by performing a simple weighted average on the respective access point reading across the N records. Table 1 depicts the example of original records ( # R S S I ) from 1 to n and generated data using Dirichlet data augmentation in A1 and A2 (shaded grey).
Algorithm 1 Data Augmentation using Dirichlet Distribution (DA-Dirichlet)
  • Initialise new_data = [0 for i in range of AP size ’t’]
  • Initialise weight ← Dirichlet distribution with α and output size = 1
  • for t = 1 to T do                {for each access_point ’t’}
  •      for n = 1 to N do             {for each reference_point ’n’}
  •          new_data[t] + = reference_point[n,t] × weight[n]
  • Result: new_data

2.5. Transfer Learning with WGAN-GP

Conventional machine learning and deep learning algorithms have been traditionally designed to work in isolation. These algorithms are trained to solve specific tasks. The models have to be rebuilt from scratch once the feature-space distribution changes. Hence, transfer learning by feature extraction or fine-tuning from a pre-trained network has been widely used in deep learning applications for adaption of transferable domains, especially for discriminative models. In particular, the WGAN-GP architecture of GANs’ variations has been experimentally demonstrated to be stable and robust for the domain transfer task in the field of computer vision [34].
Utilising the prior obtained knowledge from an organically trained model at the location with the most datapoints ( d m a x ), as well as shortening the convergence time, we propose to perform transfer learning from l o c m a x (source domain) to l o c i (target domain) in this framework using WGAN-GP architecture (explained in Section 2.3 and Table 3. In other words, the saved model state of the WGAN-GP’s generator and critic were used for the transfer learning task by fine-tuning. This method was beneficial for the unique coordinates that contained fewer samples, as it improved the generated image. Since the images were normalised before training, the generated images must be denormalised before saving. The experiment sets the generating image to be the perfect square of the AP numbers, where it requires only 1/4 of originally trained epoch (1000 epoch reduces to 250 epoch) to converge with the transfer learning.

2.6. Filtering Module

A filtering module is incorporated to discard any outliers of the generated data from the model. We propose a technique to identify and minimise such data points from being included in the augmented dataset. The proposed technique is robust, unique, simple to implement, and can easily be extended and applied on any dataset. The idea comes from the fact that the outputs generated by the GANs model should be within an allowable deviation range from the corresponding original data samples. Intentionally, the generated image quality assessment, such as Fréchet inception distance (FID) or Inception Score [35], was not used as the measuring matrices in this proposal because they compute the differences between the original and generated data distribution. Because there is a lack of data for certain locations to begin with, the data distribution could not be fairly observed. Instead, in our method, we measure against the maximum allowable threshold Θ .
Given each unique location l o c ( l a t , l o n ) , there exists a sample ( R S S i , R S S j ) combination where i j . Each sample refers to the received signal strength R S S i = { R S S i , 1 , R S S i , 2 , . . . , R S S i , A P } where A P is the total number of detected access points. The differences between the two vectors ( R S S i , R S S j ) can be computed by summation of absolute differences of the two vectors at each access point a p . Empirically, we found that the L1 norm, the sum of the absolute values of the vector, and the L2 norm yield comparable results in this case; thus, the L1 norm was used to compute the difference. The maximum allowable threshold ( Θ ) in Equation (5) is the average of maximum of all location’s L1-norm summation of the per-location original data pairs over n u m _ l o c , the total number of unique locations present in the dataset. In order to ensure the quality of the generated dataset, the output generated by WGAN-GP at l o c ( R S S G i l o c ) is validated such that the minimum absolute difference (L1-norm) between the generated data and all original data of the location l o c is within the allowance of ( Θ ). Equation (6) depicts the dissimilarity that assesses how identical the augmented fingerprint is to the original one. The lower the score, the more identical the augmented data are to the original ones. It is to be noted that R S S O i and R S S G i are the R S S i of the Original and Generated dataset, respectively. In summary, the filtering condition is to be confirmed in Equation (7).
Θ = max ( O i , O j ) l o c a p A P R S S O i , a p R S S O j , a p n u m _ l o c
d i s s i m i l a r i t y ( G i l o c ) = min O l o c a p A P R S S G , a p R S S O , a p
d i s s i m i l a r i t y ( G i l o c ) Θ

3. Experiment Design and Data Preparation

The experiments are designed to address two main discussion points:
E1 
To verify the effectiveness of the proposed framework (extendGAN+). It is to have in-depth discussion on the generated data and the framework’s performance in comparison with the previous study GAN+ [25].
E2 
To evaluate the localisation performance with extendGAN+ datasets. The augmented data is inputted to train the localisation models and compared against baseline methods. It is to address the usability of the proposed augmentation method with the off-the-shelf deep learning methods on two case studies: the widely used public dataset UJILoc and a newly collected building complex dataset (explained in Section 3.1).

3.1. Base Dataset: Data Preparation

Two datasets are chosen as the base datasets: the data collection UJIndoorLoc [24] and the newly collected building complex dataset. A base dataset contains training (80% of the original training data), validation (20% of the original training data), and test sets. The input data are the RSS which have gone through data cleaning where undetected RSS are being replaced by 110 , thereafter normalised to [ 0 , 1 ] .
The UJIndoorLoc (referred as “UJI”) was collected at the Universitat Jaume I, Spain, covering three buildings with four to five floors per building (almost 110.00 m 2 ). It was collected by more than 20 different users and 25 Android devices. The building complex (referred as “BC”) was self-collected at the two-building complex (total area of 46,369 m 2 ) in Singapore. The data collection covered three floors with five participants walk-through on the marked route on two separate occasions/days using Google Pixel 4A, with an in-house data collection application. The details of data collection and preprocessing are presented in Appendix A. Both datasets were selected since they present different challenges in terms of data collection process, data collection layout, and data distribution (Figure 2). While UJI has more data collectors and various devices, BC exhibits a longer stretch of the test bed. Table 2 further illustrates the datasets’ characteristics and parameters. For instance, both UJI and BC have comparable training sizes despite having different building layouts. The data points per location depict the min-max number of records per location to train the localisation model. The numbers of 2 and 4 of UJI and BC respectively are evidently insufficient and highlight the need for data augmentation.

3.2. Augmented Dataset: Augmentation Setup

As proposed in Section 2.2, an anchor location was selected for the unique location with the most data points. In this case, l o c m a x was selected with d m a x . The WGAN-GP model was trained with converted-image ( 23 × 23 for UJI and 19 for BC) input, with RSS of total APs and padded −110 for the rest. Reducing the training time, for other unique locations, the Dirichlet distribution was used to aggregate the original data points until its data points reach d m a x = 75 , and the WGAN-GP model was transferred and tuned. In summary, each unique location consists of 150 datapoints (75 real/Dirichlet-aggregated data and 75 WGAN-GP generated data). It is to be noted that the validation and testing sets are not being augmented (i.e., the original data).
Table 3 presents the hyperparameter setting for the experiment. The selection of the hyperparameters was carried out through an iterative process of trial and error, and empirical testing was conducted by observing the Wasserstein distance for the convergence of the model, evaluating the quality of the generated samples by computing the dissimilarity score and testing the efficacy by using it in training with the localisation model. As the transfer learning approach is adopted for the extendGAN+, the training time is reduced to a quarter of the total time as the epoch is reduced from 1000 to 250 for locations other than l o c m a x . The training of the localisation model was split between buildings and floors.

3.3. Experiment Setup

To study the effectiveness of the data augmentation, localisation methods are needed in order to benchmark against the ground truth. The performance is measured by estimating the root mean square errors (or distance error in meters). Another evaluation metric is measuring the improvement ( I m p r o v e m e n t % ) of the method ( R M S E a l g o ) against the baseline ( R M S E b l ) are calculated as shown in Equation (8).
I m p r o v e m e n t % = R M S E a l g o R M S E b l × 100
In this study, the localisation models used for comparison and its configuration (Table 4) are as the following:
  • Fingerprint (FP): a baseline representing a widely used method deployed in the real environment.
  • RandomForest (RF): a baseline for the conventional machine learning methods. It is well-suited for small datasets.
  • Deep Neural Network (DNN): a simple DNN model was trained to predict the user’s location. It is used as a baseline method where the input data are a single sample of RSS data.
  • Residual Network (ResNet) [36]: a state-of-the-art CNN-based model representation for comparison. In particular, ResNet18 was selected due to its lower number of parameters, as it would take a shorter time to train, and comparable performance with its peers such as MobileNet.
  • Residual Network + DNN (ResNetNN): a combination of ResNet and DNN such that three fully-connected layers were added to learn the location estimates.
  • Time-series-input CNN (tCNN) [15]: a CNN-based model that uses consecutive time-dependent RSS as input. Differing from the original paper which stated that the area was grouped into 3 m × 3 m grids in order to produce the [ t , A P ] input for the model where t = 10 , each unique location was used instead of having to group the area into the above mentioned grid as sufficient data are created using the augmentation method.
The various hyper-parameters across all models are kept constant. The learning rate of 0.0001 along with an ADAM optimizer was used. The model was trained for 500 epochs with early stopping condition. Preprocessing step converted all RSS values to be between [ 0 , 1 ] , where 1 indicates strong signal strength.

4. Experiment Results and Analysis

4.1. E1: Synthetic Data Quality and extendGAN+ Effectiveness

In order to study the effectiveness of the proposed method, extendGAN+ is being compared with the previously studied GAN+ [25] with two evaluation criterias: data quality and distance error (improvement). It is to be noted that other comparisons, such as the data aggregation methods, are not included due to the fact that GAN+ has been compared with, and concluded to outperform, the afore-mentioned cases.
Figure 3 shows that the differences between the data augmentation methods discussed in this paper. The case presented is randomly selected from the UJI dataset at Building0, Floor0, and Coordinate [−7637.2570, 4,864,949.8143] with nine original data points. Using the proposed method, we generated 66 Dirichlet, 75 GAN, and 75 WGAN-GP points.
Figure 3a illustrates the variety of the received signal strength (RSS) at a particular location for each AP, where its RSSs are represented in each cells (ranging from black −110 and white 0). Among the nine data points, there are about five unique patterns representing the [−7637.2570, 4,864,949.8143] coordinate. This scenario explains the fingeprint being collected at different occasions at the same place. Hence, the objective of the data augmentation is to be able to replicate this variety of RSS with reduced manual data collection effort. Observing the data produced by Dirichlet, it was as if the five unique AP patterns were combined into one, where the augmented data embodies the interpolation of existing RSS values. Likewise, the data generated by GAN shows similarity to the data of Dirichlet with a few additional patterns. WGAN-GP is observed to produce most versatile behaviour, which almost showcases the unique patterns shown by the original. Moreover, it illustrates new variations with detected AP involved. According to the aforementioned objective, the data from WGAN-GP manage to produce the data as intended.
To quantify the discussion of the synthetic data quality, a dissimilarity score is calculated and Figure 3b presents the probability density distribution of the dissimilarity score of the generated data to the original data. The score is estimated using the minimum absolute difference between the generated data to all original case (explained in Section 2.6 and Equation (6)). It assesses how identical the augmented fingerprint is to the original ones. The histogram represents the actual probability density distribution while the line graph depicts the kernel density estimate (KDE) plot visualising the continuous probability density of the dissimilarity score. The KDE plots shows that Dirichlet data aggregation has the most narrow spread. In fact, the dissimilarity score distribution of Dirichlet is between 233.8 and 428.78 as compared to [78.94, 434.82] and [174.27, 431.80] of GAN and WGAN-GP respectively. It conforms with the prior observation that Dirichlet replicates the data by assigning values to all known APs for the location, hence generating similar sets of data. The results of GAN and WGAN-GP differ in terms of the spread and peak of the dissimilarity score distribution. The dissimilarity score distribution for GAN has a wide range of values. A dissimilarity score of 0 indicates that the generated sample is identical to the original data in both pattern and value. While this suggests that GAN is capable of producing samples that are nearly identical to the original data, the goal of creating synthetic fingerprint data is to generate samples that are similar in pattern to the original data, but different enough to mimic real-world variations in RSS caused by environmental factors such as obstructions or changes in the number of people in the area. Therefore, we desire output that is more dissimilar to the original data, in order to better capture the fluctuations in RSS that occur in the real world. In contrast, WGAN-GP produced a left-skewed distribution with a moderate spread, falling somewhere between the distributions produced by Dirichlet and GAN. This distribution better aligns with the desired characteristics of similarity and dissimilarity of the RSS patterns for the generated data. In comparison to the other methods, WGAN-GP is more suited to generating samples that capture the fluctuation in RSS that occurs in the real-world scenario.
In another aspect, localisation performance could be used to study the effectiveness of its input data. Hence, the average i m p r o v e m e n t % over GAN+ (Table 5) was derived from the distance error reported in the Appendix B using Equation (8).
Comparing the datasets, it shows that the BC benefits from the extendGAN+ more than from the UJI by showing positive improvement across all localisation methods. The results align with their respective base data’s characteristics such that improvement is more apparent for the dataset that lacks versatility in its base data. In other words, the increase in BC performance is due to its lack of data collection, with only two occasions, five users, and one device type, while UJI may not require much versatility in its pattern as the data were already collected with more devices and users. Regarding the localisation models, ResNetNN with extendGAN+ improves its performance across the building samples, up to 10.13% of improvement, while other methods shows mixed results. The DNN, ResNet and ResNetNN results could be explained such that the deeper and wider the architecture of the model, the more data it may need. Having the deeper structure, ResNetNN benefits from more balanced and versatile data. However, DNN’s, ResNet’s, and ResNetNN’s overall performances have gained better results from the extendGAN+. The poor results of tCNN are not uniquely poor for extendGAN+ but overall for the UJI dataset (shown in Appendix B).
In summary, combining the case illustration and dissimilarity score distribution, WGAN-GP produces more versatile data than GAN and Dirichlet while staying on course with the original data presentation. Thus, the proposed method has shown its effectiveness in generating the variety of RSS cases. The improvement% over GAN+ has reaffirmed the case study by showing that the performance of extendGAN+ improves over GAN+ overall.

4.2. E2: Localisation Performance

In this section, the data augmentation, extendGAN+, is evaluated based on localisation performance benchmarks (DNN, ResNet, ResNetNN and tCNN) against baseline methods (FP and RF) and its own base dataset (base_data). The performance benchmark is measured in form of distance error (meter) (Appendix B) and derived fraction of improvement shown in Figure 4.
The overall trend shows the extendGAN+ with DNN, ResNet, and ResNetNN improves the performance over the baseline methods of FP and RF for both datasets, except tCNN for UJI dataset. Over baseline methods, the maximum improvements have been achieved by 23.47% over FP and 25.35% over RF on average. On the other hand, as mentioned in Section 4.1, tCNN performs worse than any others in all categories regardless of data augmentations.
Another important evaluation parameter is “over base_data” where the focus is whether extendGAN+ data augmentation again increases in performance in comparison to the base dataset using the same localisation method. It is not surprising that not all three localisation methods improve at the same rate with the data augmentation. The performance varies depending on the samples and model architecture. The maximum improvement is 18.88% on average. Figure 5 depicts the corner examples of the data augmentation using extendGAN+. The proposed method yields significant improvement at UJI_B2F4 and BC_F-1 in contrast to UJI_B2F3 and BC_F2. UJI_B2F4 does not have the training data covering the test dataset that causes the localisation model to be less accurate with the limited data. BC_F-1 brings up a comparable example to the urban building complex, e.g., a shopping mall, where the center of the area is hollow. In this case, the RSS of APs from other floors are affecting the collected data of the target floor. It causes the prediction to be inaccurate. The data augmentation could add to the fingerprint cases to improve the accuracy.
In this experiment, it could be concluded that the extendGAN+ data augmentation is indeed effective and usable for improving the localisation performance in comparison to the baseline methods and base datasets. It is observed that the extendGAN+ is the most effective when the original data is scarce, incomplete or affected by uncertainty. On the other hands, it is the least effective when the data is sufficient or the area are fully covered by the data collection.

5. Conclusions

In this study, we introduced the extendGAN+ data augmentation transferable framework to synthesise RSS data and improve the localisation performance. The recommended framework not only highlighted the effectiveness of using Wasserstein loss and gradient norm penalty to improve the data quality, but also provided a practical guideline for the amount of generated data. The performance of the data augmentation was discussed in two main discussion points, (1) evaluating the data quality and (2) its impact on the off-the-shelf localisation model in comparison to the baseline methods and base datasets. The experiment was carried out on public and self-collected data to address various challenges in data collection and localisation model deployment in the real-world environment. The data quality and dissimilarity score have shown that extendGAN+ is able to produce various RSS pattern and gain improvement up to 10.13% over the previous study GAN+ using ResNetNN model. For the localisation performance, extendGAN+ is is most effective when the orginal data are scarce, incomplete, or prone to mixed signals due to a hollow area in the buildings. It achieves the maximum improvement: up to 23.47% over Fingerprint, 25.35% over Random Forest, and 18.88% over its base dataset. Nonetheless, the limitation of this method boiled down to the combination of the augmented data and its localisation models, as it has been shown in the study that a deeper and wider model favors the data with augmentation as well as the need for versatility in the originally collected data. In future work, we aim to further study the combination of localisation models with the data augmentations, especially localisation models that adopt a federated learning framework against data augmentation.

Author Contributions

Conceptualization, S.Y. and W.G.; methodology, S.Y. and W.G.; software, W.G.; validation, S.Y. and W.G, B.-S.L. and H.L.O; formal analysis, S.Y. and W.G; investigation, S.Y., B.-S.L. and H.L.O.; resources, S.Y.; data curation, S.Y. and W.G; writing—original draft preparation, S.Y. and W.G; writing—review and editing, S.Y.; visualization, S.Y.; supervision, B.-S.L. and H.L.O; project administration, S.Y., B.-S.L. and H.L.O; funding acquisition, B.-S.L. and H.L.O. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by Singtel Cognitive and Artificial Intelligence Lab for Enterprises (SCALE@NTU), which is a collaboration between Singapore Telecommunications Limited (Singtel) and Nanyang Technological University (NTU) that is supported by A*STAR under its Industry Alignment Fund (LOA Award number: I1701E0013).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

This research was conducted at Singtel Cognitive and Artificial Intelligence Lab for Enterprises (SCALE@NTU), which is a collaboration between Singapore Telecommunications Limited (Singtel) and Nanyang Technological University (NTU) that is supported by A*STAR under its Industry Alignment Fund (LOA Award number: I1701E0013).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Building Complex (BC)—Data Collection and Pre-Processing

The data are collected at a two-building complex and the experiment is conducted on three floors, over an area of 46,369 square meters. The data collection was carried out by our in-house data collection application and five participants walked along the marked route on two separate occasions/days using Google Pixel 4A.
The raw data were collected using the self-implemented web server and mobile application (Figure A1). Before commencing collecting data on each route, a calibration step was taken place to estimate the user stride’s length, which would aid the ground truth generation.
Figure A1. Data Collection Application.
Figure A1. Data Collection Application.
Sensors 23 04402 g0a1
Subsequently, the data collected are pre-processed as follows:
  • Convert the GPS/EPSG:4326 coordinate system (native to the mobile application) to SVY21/EPSG:3414 (meter-based projected coordinate system)
  • Filter removable Access Point MAC address:
    -
    non-stationary MAC address: remove MAC address that are uncommon between the two different data collection session.
    -
    possible mobile hotspot: remove MAC addresses that appear strong in more than one cluster. It means removing any access-point with a size larger than 26 m-by-26 m.
  • Group raw reference points into 2 m-by-2 m grid. Due to individuals having different stride length, the raw data result in too many unique coordinates containing very few sample/data-points.

Appendix B. Localisation Performance—Distance Error

Table A1. Distance Error between Predicted Location and Ground Truth (meters).
Table A1. Distance Error between Predicted Location and Ground Truth (meters).
MethodInputsUJIBC
B0B1B2Building
Floor 0/Floor B1
FPbase6.70120.48113.0663.103
RFbase9.00815.63112.6072.302
DNNbase6.62412.82011.5972.406
GAN+6.92913.10610.8458.549
extendGAN+6.41812.75311.9096.396
ResNetbase6.35113.95910.6182.278
GAN+5.76010.5029.9692.203
extendGAN+6.21311.1019.4372.023
ResNetNNbase5.82911.78210.7752.565
GAN+6.53812.90411.6872.444
extendGAN+6.30112.16010.1872.318
tCNNbase9.04919.1609.5323.961
GAN+9.44318.2608.8723.534
extendGAN+9.62121.90411.0303.303
Floor 1
FPbase5.62313.6269.2603.919
RFbase10.17211.8379.4343.521
DNNbase5.9679.9627.4793.671
GAN+6.1469.4298.3046.711
extendGAN+5.6319.6757.6176.949
ResNetbase6.09518.5028.2613.072
GAN+4.73910.1386.5503.143
extendGAN+4.5567.9776.2643.104
ResNetNNbase5.32712.0526.4153.606
GAN+5.99315.4706.5423.550
extendGAN+5.83513.4966.3883.702
tCNNbase8.21717.3738.2764.642
GAN+7.75218.0629.2894.312
extendGAN+8.07718.0818.3174.304
Floor 2
FPbase5.59113.13812.8432.991
RFbase6.52412.77213.3872.598
DNNbase6.13311.3608.8982.627
GAN+5.99411.5179.7135.663
extendGAN+5.71312.0428.8535.351
ResNetbase8.61710.4069.1422.213
GAN+4.5089.6758.5392.210
extendGAN+4.88410.3529.4682.237
ResNetNNbase5.18810.5899.8155.767
GAN+5.05410.95712.0193.113
extendGAN+5.5459.8359.3802.834
tCNNbase6.38612.82112.6614.198
GAN+6.85512.42511.6853.657
extendGAN+6.48115.23113.5773.465
Floor 3
FPbase6.4999.50211.091-
RFbase7.50111.5518.750-
DNNbase7.0219.2069.040-
GAN+7.1728.5888.639-
extendGAN+7.2398.6308.964-
ResNetbase5.8578.60210.012-
GAN+5.8387.9249.287-
extendGAN+6.5187.68510.139-
ResNetNNbase5.8898.69710.293-
GAN+6.7648.90610.272-
extendGAN+6.1907.86011.509-
tCNNbase8.0089.52010.871-
GAN+8.22011.19510.929-
extendGAN+7.21913.50417.889-
Floor 4
FPbase--15.461-
RFbase--20.247-
DNNbase--15.626-
GAN+--16.149-
extendGAN+--14.037-
ResNetbase--14.702-
GAN+--15.618-
extendGAN+--15.845-
ResNetNNbase--13.959-
GAN+--16.583-
extendGAN+--14.554-
tCNNbase--19.683-
GAN+--19.362-
extendGAN+--18.029-

References

  1. Liu, H.; Darabi, H.; Banerjee, P.; Liu, J. Survey of wireless indoor positioning techniques and systems. IEEE Trans. Syst. Man Cybern. Part C Appl. Rev. 2007, 37, 1067–1080. [Google Scholar] [CrossRef]
  2. Rizk, H.; Shokry, A.; Youssef, M. Effectiveness of data augmentation in cellular-based localization using deep learning. In Proceedings of the 2019 IEEE Wireless Communications and Networking Conference (WCNC), Marrakesh, Morocco, 5–18 April 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1–6. [Google Scholar]
  3. Nguyen, D.V.; De Charette, R.; Nashashibi, F.; Dao, T.K.; Castelli, E. WiFi fingerprinting localization for intelligent vehicles in car park. In Proceedings of the 2018 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Nantes, France, 4–27 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 1–6. [Google Scholar]
  4. Van Haute, T.; De Poorter, E.; Crombez, P.; Lemic, F.; Handziski, V.; Wirström, N.; Wolisz, A.; Voigt, T.; Moerman, I. Performance analysis of multiple Indoor Positioning Systems in a healthcare environment. Int. J. Health Geogr. 2016, 15, 1–15. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Czogalla, O.; Naumann, S.; Schade, J.; Schönrock, R. Indoor Positioning and Navigation for Pedestrian Guidance in Public Transport Facilities. Technical Report. 2015. Available online: https://www.researchgate.net/profile/Olaf-Czogalla/publication/283045487_Indoor_Positioning_and_Navigation_for_Pedestrian_Guidance_in_Public_Transport_Facilities/links/58ac4fa492851cf0e3ccd544/Indoor-Positioning-and-Navigation-for-Pedestrian-Guidance-in-Public-Transport-Facilities.pdf (accessed on 8 June 2022).
  6. Retscher, G.; Bekenova, A. Urban Wi-Fi fingerprinting along a public transport route. J. Appl. Geod. 2020, 14, 379–392. [Google Scholar] [CrossRef]
  7. Dayekh, S.; Affes, S.; Kandil, N.; Nerguizian, C. Cooperative localization in mines using fingerprinting and neural networks. In Proceedings of the 2010 IEEE Wireless Communication and Networking Conference, Sydney, NSW, Australia, 18–21 April 2010; IEEE: Piscataway, NJ, USA, 2010; pp. 1–6. [Google Scholar]
  8. Olejniczak, A.; Blaszkiewicz, O.; Cwalina, K.K.; Rajchowski, P.; Sadowski, J. Deep Learning Approach for LOS and NLOS Identification in the Indoor Environment. In Proceedings of the 2020 Baltic URSI Symposium (URSI), Warsaw, Poland, 5–8 October 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 104–107. [Google Scholar]
  9. Félix, G.; Siller, M.; Alvarez, E.N. A fingerprinting indoor localization algorithm based deep learning. In Proceedings of the 2016 Eighth International Conference on Ubiquitous and Future Networks (ICUFN), Vienna, Austria, 5–8 July 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1006–1011. [Google Scholar]
  10. Kim, K.S.; Lee, S.; Huang, K. A scalable deep neural network architecture for multi-building and multi-floor indoor localization based on Wi-Fi fingerprinting. Big Data Anal. 2018, 3, 1–17. [Google Scholar] [CrossRef] [Green Version]
  11. Nowicki, M.; Wietrzykowski, J. Low-effort place recognition with WiFi fingerprints using deep learning. In Proceedings of the International Conference Automation; Springer: Berlin/Heidelberg, Germany, 2017; pp. 575–584. [Google Scholar]
  12. Fan, S.; Wu, Y.; Han, C.; Wang, X. A structured bidirectional LSTM deep learning method for 3D terahertz indoor localization. In Proceedings of the IEEE INFOCOM 2020-IEEE Conference on Computer Communications, Toronto, ON, Canada, 6–9 July 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 2381–2390. [Google Scholar]
  13. Khatab, Z.E.; Hajihoseini, A.; Ghorashi, S.A. A fingerprint method for indoor localization using autoencoder based deep extreme learning machine. IEEE Sens. Lett. 2017, 2, 1–4. [Google Scholar] [CrossRef]
  14. Song, X.; Fan, X.; He, X.; Xiang, C.; Ye, Q.; Huang, X.; Fang, G.; Chen, L.L.; Qin, J.; Wang, Z. Cnnloc: Deep-learning based indoor localization with wifi fingerprinting. In Proceedings of the 2019 IEEE SmartWorld, Ubiquitous Intelligence & Computing, Advanced & Trusted Computing, Scalable Computing & Communications, Cloud & Big Data Computing, Internet of People and Smart City Innovation (SmartWorld/SCALCOM/UIC/ATC/CBDCom/IOP/SCI), Leicester, UK, 19–23 August 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 589–595. [Google Scholar]
  15. Ibrahim, M.; Torki, M.; ElNainay, M. CNN based indoor localization using RSS time-series. In Proceedings of the 2018 IEEE Symposium on Computers and Communications (ISCC), Natal, Brazil, 25–28 June 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 01044–01049. [Google Scholar]
  16. Sinha, S.; Le, D.V. Completely Automated CNN Architecture Design Based on VGG Blocks for Fingerprinting Localisation. In Proceedings of the 2021 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Lloret de Mar, Spain, 29 November–2 December 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 1–8. [Google Scholar]
  17. Liu, X.Y.; Aeron, S.; Aggarwal, V.; Wang, X.; Wu, M.Y. Adaptive sampling of RF fingerprints for fine-grained indoor localization. IEEE Trans. Mob. Comput. 2015, 15, 2411–2423. [Google Scholar] [CrossRef] [Green Version]
  18. Yang, Z.; Wu, C.; Liu, Y. Locating in fingerprint space: Wireless indoor localization with little human intervention. In Proceedings of the 18th annual INTERNATIONAL Conference on Mobile Computing and Networking, Istanbul, Turkey, 22–26 August 2012; pp. 269–280. [Google Scholar]
  19. He, S.; Chan, S.H.G. Wi-Fi fingerprint-based indoor positioning: Recent advances and comparisons. IEEE Commun. Surv. Tutor. 2015, 18, 466–490. [Google Scholar] [CrossRef]
  20. Sinha, R.S.; Lee, S.M.; Rim, M.; Hwang, S.H. Data augmentation schemes for deep learning in an indoor positioning application. Electronics 2019, 8, 554. [Google Scholar] [CrossRef] [Green Version]
  21. Njima, W.; Chafii, M.; Chorti, A.; Shubair, R.M.; Poor, H.V. Indoor localization using data augmentation via selective generative adversarial networks. IEEE Access 2021, 9, 98337–98347. [Google Scholar] [CrossRef]
  22. Boulis, J.; Hemdan, M.; Shokry, A.; Youssef, M. Data Augmentation using GANs for Deep Learning-based Localization Systems. In Proceedings of the 29th International Conference on Advances in Geographic Information Systems, Beijing, China, 2–5 November 2021; pp. 672–673. [Google Scholar]
  23. Li, Q.; Qu, H.; Liu, Z.; Zhou, N.; Sun, W.; Sigg, S.; Li, J. Af-dcgan: Amplitude feature deep convolutional gan for fingerprint construction in indoor localization systems. IEEE Trans. Emerg. Top. Comput. Intell. 2019, 5, 468–480. [Google Scholar] [CrossRef] [Green Version]
  24. Torres-Sospedra, J.; Montoliu, R.; Martínez-Usó, A.; Avariento, J.P.; Arnau, T.J.; Benedito-Bordonau, M.; Huerta, J. UJIIndoorLoc: A new multi-building and multi-floor database for WLAN fingerprint-based indoor localization problems. In Proceedings of the 2014 International Conference on Indoor Positioning and Indoor Navigation (IPIN), Busan, Korea, 27–30 October 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 261–270. [Google Scholar]
  25. Yean, S.; Somani, P.; Lee, B.S.; Oh, H.L. GAN+: Data Augmentation Method using Generative Adversarial Networks and Dirichlet for Indoor. 2021. Available online: http://ceur-ws.org/Vol-3097/paper8.pdf (accessed on 20 December 2021).
  26. Kadurin, A.; Aliper, A.; Kazennov, A.; Mamoshina, P.; Vanhaelen, Q.; Khrabrov, K.; Zhavoronkov, A. The cornucopia of meaningful leads: Applying deep adversarial autoencoders for new molecule development in oncology. Oncotarget 2017, 8, 10883. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  27. Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and improving the image quality of stylegan. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 5 August 2020; pp. 8110–8119. [Google Scholar]
  28. Li, J.; Chen, Z.; Zhao, X.; Shao, L. MapGAN: An intelligent generation model for network tile maps. Sensors 2020, 20, 3119. [Google Scholar] [CrossRef]
  29. Dan, Y.; Zhao, Y.; Li, X.; Li, S.; Hu, M.; Hu, J. Generative adversarial networks (GAN) based efficient sampling of chemical composition space for inverse design of inorganic materials. Npj Comput. Mater. 2020, 6, 1–7. [Google Scholar] [CrossRef]
  30. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 9 November 2017; pp. 4681–4690. [Google Scholar]
  31. Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A.C. Improved Training of Wasserstein GANs. In Proceedings of the Advances in Neural Information Processing Systems; Guyon, I., Luxburg, U.V., Bengio, S., Wallach, H., Fergus, R., Vishwanathan, S., Garnett, R., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2017; Volume 30. [Google Scholar]
  32. Lucic, M.; Kurach, K.; Michalski, M.; Gelly, S.; Bousquet, O. Are gans created equal? a large-scale study. Adv. Neural Inf. Process. Syst. 2018, 31, 31. [Google Scholar]
  33. Nguyen, V.A.; Boyd-Graber, J.; Altschul, S.F. Dirichlet mixtures, the Dirichlet process, and the structure of protein space. J. Comput. Biol. 2013, 20, 1–18. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  34. Wang, Y.; Wu, C.; Herranz, L.; van de Weijer, J.; Gonzalez-Garcia, A.; Raducanu, B. Transferring gans: Generating images from limited data. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 6 October 2018; pp. 218–234. [Google Scholar]
  35. Heusel, M.; Ramsauer, H.; Unterthiner, T.; Nessler, B.; Hochreiter, S. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  36. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 12 December 2016; pp. 770–778. [Google Scholar]
Figure 1. extendGAN+ Overview.
Figure 1. extendGAN+ Overview.
Sensors 23 04402 g001
Figure 2. Base Datasets—Data Point Illustration. (a) UJI-Building0. (b) BC-Building. (c) UJI-Building1. (d) UJI-Building2.
Figure 2. Base Datasets—Data Point Illustration. (a) UJI-Building0. (b) BC-Building. (c) UJI-Building1. (d) UJI-Building2.
Sensors 23 04402 g002
Figure 3. Data Augmentation Quality Example (B0F0, −7637.2570, 4,864,949.8143). (a) Data Quality Visualisation—RSS Patterns. (b) Dissimilarity Score Density.
Figure 3. Data Augmentation Quality Example (B0F0, −7637.2570, 4,864,949.8143). (a) Data Quality Visualisation—RSS Patterns. (b) Dissimilarity Score Density.
Sensors 23 04402 g003
Figure 4. Localisation Performance—extendGAN+ Improvement (improvement%/100).
Figure 4. Localisation Performance—extendGAN+ Improvement (improvement%/100).
Sensors 23 04402 g004
Figure 5. Case study—effectiveness of extendGAN+.
Figure 5. Case study—effectiveness of extendGAN+.
Sensors 23 04402 g005
Table 1. Example Augmented Data of DA-Dirichlet.
Table 1. Example Augmented Data of DA-Dirichlet.
RSSIWAP_1WAP_2WAP_3WAP_4WAP_5WAP_6WAP_T
1−110.00−110.00−110.00−110.00−110.00−87.57−110.00
2−110.00−110.00−110.00−110.00−110.00−88.00−110.00
3−110.00−56.08−110.00−110.00−110.00−95.33−75.92
4−110.00−61.25−110.00−110.00−110.00−110.00−78.94
.
N−110.00−110.00−110.00−110.00−88.00−87.57−110.00
Augmented Data
A1−110.00−77.65−110.00−110.00−106.98−97.92−90.16
A2−110.00−84.12−110.00−110.00−106.12−94.90−93.61
Table 2. Base Datasets Description.
Table 2. Base Datasets Description.
InformationUJI DescriptionBC Description
Area of data collection3 buildings,
4–5 floors/building
1 building,
3 floors
Input data520 access points,
RSS range = [−104, 0] or
−110 if undetected
345 access points,
RSS range = [−110, 0] or
−110 if undetected
Label dataLongitude, Latitude
(EPSG:3857)
Longitude, Latitude
(EPSG:3414)
Training/testing size19,937/111116,157/6465
Per-location data points[2, 80][4, 104]
Table 3. extendGAN+ Hyperparameters.
Table 3. extendGAN+ Hyperparameters.
HyperparametersGAN+extendGAN+
UJIBCUJIBC
Data
(per unique location)
RSS (520)RSS (345)RSS image (23 × 23)RSS image (19 × 19)
Generator10 (input—latent noise),
128, 256, 512, 1024,
520 (output)
10 (input—latent noise),
128, 256, 512,
345 (output)
100 (input—latent noise),
3 × 3 × 512,
6 × 6 × 256,
12 × 12 × 128,
23 × 23 × 1 (output)
100 (input—latent noise),
3 × 3 × 512,
5 × 5 × 256,
10 × 10 × 128,
19 × 19 × 1 (output)
Discriminator/Critics520 (input),
512, 256, 128,
1 (output—real/fake)
345 (input),
256, 128,
1 (output—real/fake)
23 × 23 × 1 (input),
12 × 12 × 128,
6 × 6 × 256,
3 × 3 × 512,
1 (output—score)
19 × 19 × 1 (input),
10 × 10 × 128,
5 × 5 × 256,
3 × 3 × 512,
1 (output—score)
Critic Iteration--55
Gradient Penalty--1010
OutputRSS (520)RSS (345)RSS (23 × 23)RSS (19 × 19)
Batchsize4444
Epoch100010001000 (for l o c m a x )
250 (for l o c i )
1000 (for l o c m a x )
250 (for l o c i )
Learning Rate0.0010.0010.0010.001
Table 4. Experiment Cases.
Table 4. Experiment Cases.
Localisation MethodInputsConvolutional LayersHidden Layers
UJIBC
FPbase---
RFbase---
DNNbase-500, 500, 500300, 300, 300
GAN+
extendGAN+
ResNetbaseResNet18--
GAN+
extendGAN+
ResNetNNbaseResNet18500, 500, 500300, 300, 300
GAN+
extendGAN+
tCNNbaseLayer1: 8 out channels and 10 × 3 kernel
Layer2: 4 out channels and 5 × 3 kernel
Pooling: 2 × 2 (stride = 2)
128, 128, 128128, 128, 128
GAN+
extendGAN+
Table 5. Improvement% over GAN+.
Table 5. Improvement% over GAN+.
Localisation
Methods
UJI
B0B1B2BC
DNN4.87%−1.24%3.32%9.04%
ResNet−6.00%2.91%−2.36%2.74%
ResNetNN1.26%10.13%6.13%3.29%
tCNN2.89%−15.82%−15.63%4.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yean, S.; Goh, W.; Lee, B.-S.; Oh, H.L. extendGAN+: Transferable Data Augmentation Framework Using WGAN-GP for Data-Driven Indoor Localisation Model. Sensors 2023, 23, 4402. https://doi.org/10.3390/s23094402

AMA Style

Yean S, Goh W, Lee B-S, Oh HL. extendGAN+: Transferable Data Augmentation Framework Using WGAN-GP for Data-Driven Indoor Localisation Model. Sensors. 2023; 23(9):4402. https://doi.org/10.3390/s23094402

Chicago/Turabian Style

Yean, Seanglidet, Wayne Goh, Bu-Sung Lee, and Hong Lye Oh. 2023. "extendGAN+: Transferable Data Augmentation Framework Using WGAN-GP for Data-Driven Indoor Localisation Model" Sensors 23, no. 9: 4402. https://doi.org/10.3390/s23094402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop