Next Article in Journal
Method of Evaluating Multiple Scenarios in a Single Simulation Run for Automated Vehicle Assessment
Previous Article in Journal
Comparative Analysis of Predictive Interstitial Glucose Level Classification Models
Previous Article in Special Issue
Research on Path Planning and Path Tracking Control of Autonomous Vehicles Based on Improved APF and SMC
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Using the LSTM Neural Network and the UWB Positioning System to Predict the Position of Low and High Speed Moving Objects

by
Krzysztof Paszek
1,* and
Damian Grzechca
2
1
Department of Telecommunications and Teleinformatics, Faculty of Automatic Control, Electronics and Computer Science, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland
2
Department of Electronics, Electrical Engineering and Microelectronics, Faculty of Automatic Control, Electronics and Computer Science, Silesian University of Technology, Akademicka 16, 44-100 Gliwice, Poland
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8270; https://doi.org/10.3390/s23198270
Submission received: 18 July 2023 / Revised: 25 September 2023 / Accepted: 2 October 2023 / Published: 6 October 2023
(This article belongs to the Special Issue Automated Guided Vehicle Integrated with Collaborative Robot)

Abstract

:
Automation of transportation will play a crucial role in the future when people driving vehicles will be replaced by autonomous systems. Currently, the positioning systems are not used alone but are combined in order to create cooperative positioning systems. The ultra-wideband (UWB) system is an excellent alternative to the global positioning system (GPS) in a limited area but has some drawbacks. Despite many advantages of various object positioning systems, none is free from the problem of object displacement during measurement (data acquisition), which affects positioning accuracy. In addition, temporarily missing data from the absolute positioning system can lead to dangerous situations. Moreover, data pre-processing is unavoidable and takes some time, affecting additionally the object’s displacement in relation to its previous position and its starting point of the new positioning process. So, the prediction of the position of an object is necessary to minimize the time when the position is unknown or out of date, especially when the object is moving at high speed and the position update rate is low. This article proposes using the long short-term memory (LSTM) artificial neural network to predict objects’ positions based on historical data from the UWB system and inertial navigation. The proposed solution creates a reliable positioning system that predicts 10 positions of low and high-speed moving objects with an error below 10 cm. Position prediction allows detection of possible collisions—the intersection of the trajectories of moving objects.

1. Introduction

Developing economy, and thus industry, is primarily based on the automation of processes. A good example of this is the automation of the movement of objects (driving a vehicle), which is noticeable not only in industrial halls but also in road traffic. Both scientists and companies are genuinely interested in automated warehouses [1,2]. It can already be seen in such places where goods are transported by autonomous robots—automated guided vehicles (AGVs) [3,4]. This usually happens along strictly defined routes, but humans are not involved in the transport process.
Automation of transport is a crucial element of the future, where it will support (it has already been partially achieved) or replace people in the process of driving a vehicle. Companies already offer public transportation using autonomous vehicles [5]. The future of autonomous vehicles is also confirmed by works carried out not only in the field related to systems and positioning technologies but also to aspects such as timetable planning [6], ethical and financial considerations [7], and accessibility [8].
Many technologies provide additional information to drivers, allowing them to see things normally invisible to the driver, such as a vehicle in the blind spot. Radars, lidars [9,10,11,12,13,14], video analysis [15,16,17,18,19,20], ultrasonic and hall systems [21,22] observe and scan the vehicle’s surroundings and can provide information about the impending danger (i.e., the position of the object relative to an obstacle or another traffic participant is determined). They can also decide and act without the driver’s intervention, e.g., brake in an emergency or park the vehicle in the indicated place (active parking assist) [23]. These systems also allow the monitoring of the distance to the car in front, warning the driver about not keeping sufficient distance or automatically adjusting the speed to the current traffic on the road [24]. Nevertheless, these systems must “see” other traffic participants or an obstacle so that the superior system can react—thus, the object must be within the device’s field of view (FOV), which is usually limited vertically and horizontally. The information about the obstacle’s or moving object’s position must be known in advance so the system has time to make a decision (e.g., warn the driver or stop the vehicle). This is particularly important in highly urbanized areas or warehouse halls, i.e., in places where visibility for both drivers/operators and the systems scanning the environment (radars, lidars and vision systems) is limited and does not allow to protect moving vehicles against possible dangerous situations. Additionally, information about dangers has to be propagate and therefore vehicles must communicate with the environment and each other [25,26]. To ensure the most recent data, the system should be a real-time system in order to perform actions and respond to events within predictable and specific time constraints [27]. In positioning systems, defined time constraints allow to determine the possible displacement of the object from the position in which the sensors’ data were obtained to the place where the information about the position is available. Time constraints and the possible displacement associated with them are used as one of the parameters determining the minimum distance from objects to other objects or obstacles—the position uncertainty. Various rules are used to ensure a safe distance between vehicles by a driver. One of them—the three-second rule, says that the driver should maintain a minimum distance from the vehicle ahead, which is equivalent to 3 s of reaching this object at a given speed. This time may seem long, but it should be considered that the driver (human) needs time to react—the time (and therefore the distance) can be shortened for computer systems where the information processing time is shorter. However, the approaching object must still be visible.
As can be seen, systems capable of detecting a possibly dangerous situation will not always be able to detect and predict collision, and their most significant limitation is the requirement for visibility. Thus, for this task, another technology is required to determine the object’s absolute position in the local or global frame of reference so that it is possible to determine the trajectories of moving objects and their potential collisions. In cooperation with inertial navigation, the GPS is the most commonly used system to determine the absolute position [28,29,30]. However, the GPS signal is often distorted, reflected, attenuated, or utterly unavailable in urbanized areas, and inertial navigation system errors accumulate over time [31].
Ultra-wideband (UWB) is a rapidly developing technology increasingly used to position objects indoors. The UWB positioning system provides information about distances between network nodes—tags associated with the positioned object and the anchors (reference points). Then, the obtained distances are used to calculate the position in the trilateration process [32,33]. UWB technology has become a beneficial alternative to other systems (especially GPS) used in the automotive industry (in limited areas) due to its low energy demand, low costs, additional data transferability, and high communication security [34,35,36,37,38]. Due to many possible applications, the UWB system is not only an alternative but is also increasingly used with other systems that determine the position of an object (as a supplementary system), creating a cooperative positioning system [39]. The advantages of the UWB system, such as the accurate positioning or the possibility to communicate, make this system widely researched [40,41,42]. Still, one should always remember the system’s limitations, such as the frequency of acquiring positions (and connected with its displacement), as well as the possibility of communication errors or loss. Yet, this system is mainly studied for the positioning of slowly moving or static (at least at the moment of measurement) objects, so an indoor positioning system (IPS) is used. However, with the increase in the speed of the object and the decrease in the frequency of obtaining data from the system, the displacement becomes greater; therefore, position prediction becomes crucial and is even required within the context of real-time locating systems (RTLS). Additionally, the maximum distance between nodes is limited and depends on system configuration. For example, the range of DW1000 modules working in operational mode 2 is limited to 100 m; mode 5 is limited to 50 m (on the market, there are also modules with an amplifier that allow ranging over 150 m). In practice, the range is lower because of signal attenuation and disruption. For this reason, the usage of the UWB system in larger areas requires more UWB nodes. This, in turn, introduces new problems in scheduling communications to avoid collisions and to ensure the highest frequency of position acquisition. In the studies presented in this article, the movement of the object with an initial speed of up to 20 m/s is considered, which covers the use of the UWB not only as an IPS but also as an outdoor positioning system (OPS). The problem of displacement (resulting from delays in the positioning system mentioned above) can be solved by predicting the object’s position based on historical data from the positioning system or systems. In addition, position prediction allows continued positioning in situations where a temporary signal loss from the UWB system occurs (no data packet with position). Then, without prediction, it is impossible to determine the object’s absolute position in the local frame of reference, and thus, it is impossible to avoid dangerous situations. It may seem that at low speeds with which AGVs move, position prediction does not matter. However, the inertia of a vehicle carrying a heavy load (several tons), for which the braking distance will be long, should be taken into account.
Consider a situation where a vehicle moves along a narrow road surrounded by buildings (or a warehouse that is densely packed with racks). When approaching an intersection, the driver or autonomous system should ensure it is safe to cross the intersection. However, in highly limited visibility conditions, a driver or an assistance system may need a significant speed reduction or even stopping the vehicle (regardless of road markings) to ensure safety. The ability to predict objects’ positions and communication between vehicles is required to provide the biggest possible capacity of a road (or the fastest transportation of goods in warehouses) and to eliminate possible dangerous situations in this type of scenario. Knowing the future positions of the vehicles, it is possible to determine whether vehicles’ trajectories intersect and, if so, when and whether a dangerous situation may occur.
Prediction of the position of moving objects is, in other words, forecasting an event or feature value of a variable based on information from the past. Various methods are used to forecast the value of a variable, e.g., Kalman filter (KF), autoregressive integrated moving average (ARIMA) [43], or recurrent neural networks (e.g., LSTM) [44,45]. KF is widely used to estimate the position of an object. Its application involves the preparation of a mathematical model, which also allows for data fusion in multi-sensor systems. With the increase in data sources, the mathematical model becomes complex and often requires certain assumptions to simplify the model. Finding a direct relationship between the data is sometimes complicated or even impossible. The big challenge is to tune the noise covariance matrix, which tells how relevant the model is, i.e., what is the difference between the model and the actual motion of the object. ARIMA models are used for short-period prediction based on past data. This model is widely used for forecasting univariate problems—cannot capture dependencies between different data sources (variables). The vector autoregressive (VAR) is designed for multivariate problems—so it can consider multiple data sources and find the relation between them [46]. This model can be extended with exogenous variables (unmodeled inputs to the model) that are introduced to explain outcomes in the model (VARX—vector autoregression with exogenous series). The neural network-based approaches can find dependencies between a large number of variables without a complex mathematical model—so the way of data fusion is learned during training. The commonly used recurrent neural network for forecasting problems is LSTM. In the position prediction, the dynamic model can be unknown for LSTM in contrast to KF, where the model needs to be specified.
Undoubtedly, the use of the UWB system will make it possible to “see” other traffic participants, and the prediction of the object’s position using the LSTM network may lead to the determination of the trajectory of the objects and their possible intersection—a dangerous situation, solving the problem of limited frequency of acquiring positions and displacement during data acquisition and processing (which is especially important when the subject is moving in a dense environment and/or at high speed).
In positioning systems, the KF, which is often used to estimate (filter) positions based on various data from various technologies, comes in different versions. Data from inertial sensors and the UWB system are commonly fused in multiple ways using KF. In [47], the authors presented an interesting method that combines UWB and pedestrian dead reckoning (PDR) in different conditions of the density of UWB anchors and different visibility of positioning pedestrians. They prepared two Kalman filters for orientation and position, using dynamic and fixed variance estimation. The authors also emphasize that incorrect error quantification may affect the system’s filtering capacity (position estimation) for a longer time. The average position error of their method under line-of-sight (LOS) conditions with a high density of anchors is about 0.5 m. The authors of [48] presented the weaknesses of the Kalman filter and its various modifications (extended, unscented, and cubature Kalman filters) both in simulation and real tests. These include, among others, the lack of gross error resistance, dependency on the initial values, and problems with the linearization and stability of the algorithms. The RMSE of the position in simulated static measurements ranges from 42–191 cm, depending on the Kalman filter variant used. The authors proposed the use of a robust particle filter for which the RMSE is 40 cm in the same test scenarios, but according to the authors, the computation is complex and time-consuming.
Apart from traditional methods, machine learning methods are also used to increase the accuracy of determined positions. Their use is not always directly related to the object position, e.g., fingerprinting [49,50], but they have an impact on it, e.g., by detecting LOS/non-line-of-sight (NLOS) conditions [51,52,53] or even LOS/NLOS/multi-path (MP) conditions [54] or system error prediction [55,56]. In [57], the authors proposed the use of an LSTM network to determine the position of an object based on distances to 3 reference nodes. The distances used in the research were also obtained from the simulation environment of the UWB system, as well as the ground truth, which is difficult to obtain in other than simulation conditions. The accuracy of the proposed network was compared to other methods based on least square estimation, fingerprint estimation, maximum likelihood estimation and weighted centroid estimation algorithms, back propagation neural network, extreme learning machine, and recurrent neural networks. The authors indicated that their LSTM network approach allowed them to locate the object with a mean position accuracy of 7 cm, which was the best result compared to other tested methods. The authors of [58] proposed the use of the LSTM network for positioning objects with a fingerprinting method based on channel state information (CSI). CSI provides more information than the receiving signal strength indicator (RSSI) commonly used in Wi-Fi fingerprinting. The authors mapped high-dimensional amplitude and phase information to low-dimensional space using principal component analysis to reduce the input data size to the LSTM network, which predicts the position. The model the authors presented allows us to determine the object’s position with an average error of about 1 m (depending on, among others, the environment in which the research was carried out).
As presented, LSTM networks are widely used at different stages of object positioning. The research shows that they often give better results than other commonly used algorithms. No solution was found in which the LSTM network (which is commonly used for forecasting problems) was used for the position prediction of the object using displacement prediction.
The approach proposed in this article allows us to determine the moving object’s position with the use of an LSTM neural network based on data from UWB and IMU. There are the following contributions to this paper:
  • The training neural network database was prepared, including movements in different directions and with different speeds and accelerations.
  • The LSTM network architecture was introduced for object positioning based on UWB position (displacement) and IMU sensors.
  • The prepared model was analyzed, and the accuracy of the position prediction was evaluated.
  • The execution time of the prepared model was analyzed in the context of the displacement of the considered types of vehicles.
The paper is organized as follows. Section 2 describes the source of position errors and how much the displacement impacts it. Section 3 presents the system architecture and test scenarios prepared using the UWB simulator. Section 4 describes the structure of the LSTM network and pre-processing of the data. Section 5 describes the accuracy of the prepared model. Section 6 concludes the article.

2. Positioning of Moving Objects

Data pre-processing, correction, and filtration allow the removal of random system errors or measurement inaccuracies of the device or sensors used. The sensor measurements are affected by many factors, ranging from the manufacturing factors of the sensing device to the environmental factors present at the time of the measurement process (e.g., LOS and NLOS conditions).
Concentrating on the data from the UWB positioning system, the accuracy of the distances determined (collected in the first phase of the positioning) is affected by environmental conditions that can cause signal reflections or attenuation (e.g., a passing person or a passing vehicle between the UWB system nodes) or errors in distance measurement (e.g., rounding error which depends on the used microcontroller or clocks instability). The time of the whole positioning process should be known (starting from the ranging process, going through data acquisition, and ending with position calculation) to estimate the possible displacement of the moving object. It should be noted that in order to determine the position of the object on the plane (2D—two-dimensional), it is necessary to obtain at least three distances from the tag (a mobile UWB node) to the chosen reference points (stationary UWB nodes). However, processing the acquired distances is relatively small in relation to the time needed for data acquisition from the UWB system.
So, the most significant impact on the accuracy of the determined positions of the moving object is the time delay resulting from the ranging process between the tag placed on the moving object and the reference points. As the speed of the object increases and the acquisition time increases, the object’s displacement also increases—see Figure 1, according to Formula (1):
d = v · t p
where d —displacement of the moving object; v —speed of the moving object; t p —time needed to collect the required distances to determine the object’s position.
Therefore, the distances obtained from the system refer to positions from the past. On the market, various ready-to-use solutions for positioning objects using UWB technology can be found, which are characterized by different positioning frequency (which often depends on the size of the system—number of nodes) and accuracy. However, the accuracy of positioning systems often does not take the aspect of object movement into account (when the objects are constantly in motion—do not stop during the positioning process), and the accuracy of the system provided by the manufacturer applies, in fact, to stationary measurements.
An acquisition time of 100 ms results in a displacement of 1.4 m for a car moving at 14 m/s, see Figure 1. In such a situation, regardless of the accuracy of the UWB positioning system or the applied filtration on data, the moving object can lead to a dangerous situation (collisions with other traffic participants or elements of infrastructure). Therefore, it is necessary to predict the position of the object so that it is possible to detect dangerous situations. Additionally, predicting the position allows for determining the object’s position as accurately as possible (the calculated prediction of position corresponds to the current real position due to the fact that the object is in motion—the distances received are used to calculate the location at a given moment refer to past positions of the object).

3. System Description

Positioning of objects plays a vital role on a global scale (in the global frame of reference—the globe) and locally (in the local frame of reference covering a limited area, e.g., manoeuvering yards, intersections, warehouse halls, production halls). Information about the position in the local reference frame can be converted and used, for example, in a building diagram or a map of a town, which allows the location of objects in a broader sense. Systems and technologies that determine the object’s position differ from each other. They have their advantages and disadvantages, where the disadvantages of one technology should be outbalanced by the advantages of another or by appropriate data analysis.
In this article, a positioning system consisting of a UWB and attitude and heading reference system (AHRS) that allows for positioning moving objects in a limited area is considered—an example system scenario is presented in Figure 2. The movement of AGVs (e.g., forklifts), which usually move at a speed of up to 2 m/s and acceleration of up to 2 m/s2, and motor vehicles (e.g., cars, motorbikes) moving in urban traffic at a speed of up to 14 m/s and acceleration of up to 10 m/s2 have been studied.
This article proposes using the long short-term memory (LSTM) neural network to predict the object’s position based on data from the UWB (position in a 2D plane) and AHRS (acceleration, angular speed, and direction). Obtaining the ground truth (the reference position) for the UWB positioning system is complicated and often challenging if the movement of motor vehicles is considered. Therefore, the data set for artificial neural network training and testing was prepared in a simulation in the MATLAB computing platform. Data from the UWB system were generated using the authors’ simulator that reflects the operation of a real UWB system based on DW1000 (manufactured by Decawave—now part of Qorvo) modules (with a positioning time of 12 ms), which provide ground truth for further analysis [59]. On the other hand, the inertial unit was simulated using the simulator, which is built into the MATLAB computing platform, using the parameters of the Xsense MTI-670-DK inertial unit. The simulation environment with a sample path is presented in Figure 3. Four reference UWB nodes were arranged on a square plan of 10 m (no obstacles in the test stand) with the following coordinates (x [m], y [m]): anchor 1 (0, 0), anchor 2 (0, 10), anchor 3 (10, 0), anchor 4 (10, 10).
Simulation allows the preparation of a data set that is large enough to train the artificial neural network correctly and provide the reference positions. Moreover, the simulation allows for the preparation of test scenarios that are difficult to obtain in real conditions (e.g., enormous acceleration or speed) or the execution cost is very high.
Data were prepared for two variants of the object’s motion in a straight line:
  • Variant A—uniform motion with speeds in the range of 0.25 m/s to 20 m/s, and
  • Variant B—uniformly accelerated motion with an initial speed of 0 m/s and an acceleration in the range of 0.1 m/s2 to 10 m/s2.
The movement paths were carried out in such a way that they could reflect changes not only in the displacement value but also in terms of the direction of the object’s movement in the local frame of reference. The total number of test scenarios is 1716. The test parameters scenarios (variants A and B) are summarized in Table 1.

4. LSTM Network

The LSTM network is a type of recurrent neural network that uses whole data sequences. A cell state contains information learned in the previous and current time steps. This type of network cell comprises three gates: input gate, output gate, and forget gate, which at every discrete time step add information to the memory cell or remove information from it [60,61,62,63]. The structure of the single memory cell is presented in Figure 4.
The h t vector is a hidden state of the cell (also called output) which corresponds to the short-term state (2), while the c t vector is a cell state corresponding to the long-term state (3). States of the cell are controlled using gates: the forget gate f t , as the first, decides which long-term information will continue to pass the network—it decides which information stays and which gets deleted (4). Candidates to be added to the long-term state are selected using the cell candidate g t (5). New information is added by the input gate i t to the long-term state (6). In the end, a short-term state is created using the output gate o t (7):
h t = o t σ c c t
c t = f t c t 1 + i t g t
f t = σ g W f V t + R f h t 1 + b f
g t = σ c W g V t + R g h t 1 + b g
i t = σ g W i V t + R i h t 1 + b i
o t = σ g W o V t + R o h t 1 + b o
where W f ,   W g ,   W i ,   W o —input weights matrices of, respectively, forget gate, cell candidate, input gate, and output gate; R f ,   R g ,   R i ,   R o —recurrent weights matrices of, respectively, forget gate, cell candidate, input gate, and output gate; b f ,     b g ,   b i ,   b o —bias vector of, respectively forget gate, cell candidate, input gate, and output gate; V t —input vector of the LSTM unit; σ g —the gate activation function; σ c —the state activation function; —denotes Hadamard product.
Input weights— W f ,   W g ,   W i ,   W o , recurrent weights— R f ,   R g ,     R i ,   R o , and biases— b f ,   b g ,   b i ,   b o , are learned (tuned) during the training process of the network.
The activation function used to update the cell state and output (hidden state) is a hyperbolic tangent function (8), but the activation function applied to the gates is a sigmoid function (9) given as:
σ c = e 2 z 1 e 2 z + 1
σ g = 1 1 + e z
where z —the summed weighted input (of the cell, gate).
The input data to the LSTM network—discussed in this article, are the displacements of the object determined using historical data from the UWB system—where the displacement of the object to the current position from the previous position is defined as displacement on the plane (X and Y axis) with the appropriate sign (10); and data obtained from the AHRS: accelerometer, gyroscope and magnetometer in X, Y and Z axes, at constant intervals. The data from the systems listed above form a vector of parameters— V (11), which is used in the neural network training process and then in the prediction of the position:
Δ P = Δ x , Δ y = x i x i 1 , y i y i 1
V = v 1 , v 2 , , v 11 = Δ x , Δ y , a x , a y , a z , g x , g y , g z , m x , m y , m z
where v j —element of the neural network input, j 1 , 2 , , 11 ; Δ x —displacement in the X-axis [m]; Δ y —displacement in the Y-axis [m]; a x ,   a y ,   a z —acceleration from the accelerometer, respectively, in the X, Y, and Z-axes [m/s2]; g x ,   g y ,   g z —angular velocity from the gyroscope, respectively, in the X, Y, and Z-axes [rad/s]; and m x ,   m y ,   m z —magnetic field from the magnetometer, respectively in the X, Y, and Z-axes [µT].

4.1. Pre-Processing of Data

Machine learning algorithms (especially those that use gradient descent methods for optimization) are, in most cases, sensitive to data (features) of different orders and ranges of values or units (12). Therefore, the data should be scaled before the training stage—in the pre-processing phase, and an appropriate scaling technique should be used, e.g., data normalization or standardization.
Θ l + 1 = Θ l α E Θ l
where Θ —network parameters vector (weights and biases); α —learning rate; E Θ —loss function, which is used to evaluate a machine learning algorithm.
The article used min-max normalization to scale the displacement (data obtained using UWB technology) and the data from inertial sensors (13). This method linearly converts the data to the desired interval m i n x , m a x x . If the range from 0 to 1 is chosen, Equation (13) simplifies to (14):
v j = v j m i n v j m a x v j m i n v j · m a x v j m i n v j + m i n v j
v j = v j m i n v j m a x v j m i n v j
where m i n v j —the minimum value of the feature v j ; m a x v j —the maximum value of the feature v j ; m i n v j —the desired minimum value of the feature v j ; m a x v j —the desired maximum value of the feature v j .
When the data used in the training of the artificial neural network are normalized, the values of the object displacement prediction (output of the network) are also normalized. Therefore, the output data from the LSTM network should be denormalized before further analysis using the previously adopted extreme values (15):
v j = v j · m a x v j m i n v j + m i n v j
where m i n v j —the minimum value of the feature v j ; m a x v j —the maximum value of the feature v j .
An overestimation or underestimation of the distances very often characterizes the UWB system. There are many reasons for this phenomenon, including the desynchronization of clocks between nodes, different accuracy of the clocks used, or the system configuration (delays). The correction mechanism is used in the pre-processing phase of data analysis to minimize the impact of the ranging error. The correction mechanism is based on mean bias error, expressed as a distance function using a second-degree polynomial. After receiving distances from the UWB system, the correction values are determined by which the distances obtained from the system should be reduced. Then, the corrected distances are used in the trilateration process (the trilateration method based on the simplex method is used in this research). The correction mechanism is described in detail in [64].
The estimation of the next position involves the use of data that could not describe similar cases but a whole set of different paths that should cover, to some extent, the possible situations (paths of movement) that may occur in the local frame of reference. It should be noted that in the proposed approach, successive displacements Δ P in the local coordinate system are used as the network’s input so that the network learns based on the displacement and not on absolute positions in the local frame of reference. So, the sequences (paths) are somewhat similar to each other (for a given speed and acceleration). This approach makes the system more universal and independent of the target coordinate system, which makes prediction possible on any section of the path along which the object is moving, regardless of its absolute position. It should also be mentioned that a finite number of straight lines can approximate an arc (bend).

4.2. Truncate Sequences

Due to a large number of measurement series (the average length of the series is 707 positions on the path) and the limited amount of computer memory, the data has to be divided into batches (a limited set of measurement series of the same length), which are sequentially put to the LSTM network. This solution updates the network parameters after each batch, and the data length (length of the path) between successive batches may vary. The data series in a single batch must be the same length as they all pass the network. For this purpose, after sorting the test scenarios and analyzing the length of all measurement series, it was decided that the best batch size is 37 series. Such a size of the batches does not require much interference in the series length (about 4% of all positions on the paths). It also allows for learning with the available operational memory. Batches consisting of 37 test scenarios are truncated to the shortest series in a given batch (red part of the graph in Figure 5). Another way is to extend the shorter series to the longest one. In this solution, data from the series are not lost; however, this procedure may introduce noise into the network, negatively affecting the predictive properties of the network.

4.3. Network Structure

The research was carried out using the LSTM neural network with a different number of memory cells (the general structure of the network is shown in Figure 6). A network with one hidden layer was chosen because, in the problem of position prediction, an increase in error was noticed with the number of hidden layers while significantly extending the network learning time [57].
The input layer consists of 11 features (11), which are normalized using min-max normalization (see Section 4.1)—vector V is obtained (16):
V = Δ x ,   Δ y , a x , a y , a z , g x , g y , g z , m x , m y , m z
where Δ x —normalized displacement in the X-axis; Δ y —normalized displacement in the Y-axis; a x , a y , a z —normalized acceleration from the accelerometer, respectively, in the X, Y, and Z-axes; g x , g y , g z —normalized angular velocity from the gyroscope, respectively, in the X, Y, and Z-axes; and m x , m y , m z —normalized magnetic field from the magnetometer, respectively, in the X, Y, and Z-axes.
A hidden layer consisting of LSTM cells (networks with 20, 70, 100, and 150 memory elements in the hidden layer are analyzed), a fully connected layer, and an output layer fully connected with the last hidden layer (regression layer on which predicted displacement—output of the network, are obtained (17)).
Δ P = Δ x p ,   Δ y p
where Δ x p —normalized predicted displacement in the X-axis; Δ y p —normalized predicted displacement in the Y-axis.
The learning process is based on the ADAM (adaptive moment estimation) optimizer.

5. Results

Prediction of the first position is performed on data known from the UWB system and inertial sensors (accelerometer, gyroscope, and magnetometer). However, each subsequent prediction is based on data from inertial sensors (available data from the system) and previous predictions (no data from the UWB system). Data associated with 20 previous positions were used to initialize the state of the trained network.
The performances of the trained neural networks were evaluated using the root mean square error (RMSE)—test set of data is a re-simulated data with the parameters described in Table 1. The RMSE of the prediction of ten positions depending on the number of memory cells broken down into the motion variant (A and B) is presented in Table 2. The starting position of the prediction is placed in 30%, 50%, and 80% of the movement of the twenty-meter path, which allows for the analysis of position accuracy at different stages of the object’s movement, covering various ranges of distances between the UWB anchors and tag (which affects the distance correction mechanism and thus the trilateration result). The memory cells of 150 reduce the error by about 2% for an increase in the number of memory cells by 50 compared to a network of 100 memory cells. Using a network with more than 100 memory cells does not bring many benefits but only increases the time needed to train the network and increases memory usage. For this reason, further results presented in this article should use the LSTM network, whose hidden layer consists of 100 memory cells. Such a network requires 44,600 weights (input weights: 4 (gates) × 100 (memory cells) × 11 (input features); recurrent weights: 4 × 100 × 100; output weights 2 (output of the network—displacements) × 100) and 402 biases (in the LSTM layer: 4 (gates) × 100 (memory cells); and in the fully connected layer: 2 (displacements)) to be tuned.
The accuracy of the prediction depending on the number of predicted positions is presented in Table 3. For randomly selected test scenarios from variants A and B, the prediction of the 1, 5, 10, 20, 40, 60, 80 and 100 positions were made. With the increase in the number of predicted positions, the prediction error increases from 6.53 cm for the first prediction to 94.25 cm for the 100th prediction of the position.
As mentioned above, the time of data processing influences the object’s displacement from the position where the first frame was exchanged to the place where the new position is available. On a personal computer with a six-core processor with a 3.6 GHz clock speed, the neural network initialization time is 1.7 ms, and the time for a single prediction is 1.9 ms (averaged values of all predicted positions). Network initialization should be performed when data from the UWB system are acquired (ranging process). Therefore, this time does not affect the displacement of the object. However, the prediction time affects the object’s displacement, which for a motor vehicle (14 m/s) results in a displacement of 2.7 cm, and for an AGV (2 m/s), a displacement of 0.4 cm. The displacement resulting from the prediction using the LSTM neural network is over six times smaller than the displacement resulting from the time of data acquisition from the UWB system (12 ms).
When the position prediction is carried out, the lack of data from the UWB system is simulated. As the LSTM neural network output gives the object displacement from the previous position to the next position, the last known position from the UWB system from which the next position can be calculated is required to determine the predicted position. Thus, the accuracy of the prediction also depends on the accuracy of the last known position received from the UWB system, from which subsequent positions are determined using the predicted displacement. So, pre-processing and data correction (of the distances from the UWB system) described in Section 4.1 are necessary and unavoidable.
An accurate comparison of the object position prediction error in variants A (divided into velocities) and B (divided into accelerations) was carried out for 10 position predictions (which corresponds to 120 ms when the acquisition time from the UWB system is 12 ms). The example of the prediction of 10 object positions is presented in Figure 7—the ground truth is the reference path of the movement (blue dots), the emulated path is the path determined using the UWB positioning system simulator (red dots), the prediction is the path obtained using presented data analysis and the prepared LSTM model (green line), the blue and red lines are the reference and simulated path, respectively, which correspond to the part of the movement which is predicted. Part (a) presents a case where there is no update from the UWB system over 10 positioning periods (no data from the UWB system over 120 ms), and part (b) presents a case where the network is updated every positioning period (data from the UWB system are available).
First, the accuracy of the UWB system for uniform motion is analyzed. The cumulative RMSE for ten object positions obtained from the UWB system in relation to the reference position is presented in Table 4. An increase in the average RMSE can be observed with each successive prediction, remaining below 5 cm until the seventh position prediction (which for a motor vehicle (14 m/s) results in a displacement of 1.18 m, and for an AGV (2 m/s) a displacement of 17 cm). It should be noted that for low speeds with a position acquisition frequency of 12 ms, the error is most affected by the inaccuracy of the UWB system, not the displacement of the object—for the speed of 0.5 m/s, the object will move only by 6 mm, where the accuracy of the presented system can be assumed at 10 cm.
Next, the accuracy of the UWB system for uniformly accelerated motion is analyzed. The cumulative RMSE for ten object positions obtained from the UWB system in relation to the reference position (obtained from the UWB system simulator) is presented in Table 5. Since the accelerated motion started from v0 = 0 [m/s], the displacements at the beginning of the movement between successive positions are small. The most significant impact on the RMSE is the inaccuracy resulting from the trilateration algorithm, not the object’s displacement. The system’s accuracy ranges from 4.5 cm to 5.3 cm. There is a visible trend towards increasing RMSE with increasing acceleration, which is associated with greater dynamics of changes in displacement. However, the average RMSE value is about 5 cm, which is comparable to the results obtained for uniform motion.
The obtained accuracy of the predicted positions for both the AGV and the motor vehicle is sufficient to allow uninterrupted positioning of the object during a short period (up to 120 ms) of signal loss from the UWB system and to predict the next position of the object. For AGVs, which usually move at a speed of up to 2 m/s and acceleration up to 2 m/s2, the RMSE value of the prediction of a path consisting of 10 positions is 4.9 cm for uniform motion, and 4.3 cm uniformly accelerated motion (according to extreme values). For motor vehicles moving in urban traffic at speeds up to 14 m/s and accelerations up to 10 m/s2, the RMSE value for the prediction of a path consisting of 10 positions is 7.3 cm for uniform motion and 6.8 cm for uniformly accelerated motion (according to extreme values).
The histogram showing the RMSE of each (independently) prediction of 10 object positions is shown in Figure 8. The RMSE of each subsequent predicted object position (up to the 10th prediction) is presented using a boxplot in Figure 9. The 75th percentile is 7 cm (10th prediction of the position).
The prepared model was also tested in different test scenarios. As previously mentioned, four reference UWB nodes were arranged on a square plan of 10 m (no obstacles in the test stand). The movements of the object were simulated (using the same UWB simulator as in scenarios A and B) not only in straight lines but also in arcs in two variants:
  • Variant C—uniform motion with speeds in the range of 0.25 m/s to 20 m/s, and
  • Variant D—uniformly accelerated motion with initial speeds in the range of 0 m/s to 18 m/s and an acceleration in the range of 0.1 m/s2 to 10 m/s2.
The movement paths reflect the changes in the displacement value and the direction of the object’s movement for both lines and arcs (in two turning variants: clockwise and counterclockwise). The total number of test scenarios is 23,572.
The influence of a larger (30, 40, 50, 80) and smaller (10) number of V vectors used to initialize the LSTM network was checked (in the previous considerations, 20 vectors V were used). The RMSE of the prediction of ten positions depending on the number of previous positions used to initialize the state of the trained network for the motion variants C and D is presented in Table 6 and Table 7, respectively. The first row shows the RMSE of the raw data—the average RMSE of RAW data for variant C is 10.4 cm and 9.6 for variant D. For variants C and D, initialization with 10 previous positions gives a 6% greater error in relation to raw data; while in relation to initialization with 20 previous positions, the error is greater by 70% for variant C and 31% for variant D. For variant C, 20 previous positions allow to achieve error less than 10 cm up to 10 position predictions. However, for variant D to achieve an error of less than 10 cm up to 10 position predictions, the number of previous positions has to be doubled (up to 40).
As can be seen, the increase in the number of positions used for network initialization positively affects prediction accuracy. However, the greatest increase in accuracy is visible for 40 initialization vectors. The prepared object position prediction system allows us to effectively predict 10 object positions in scenarios that did not occur in the training set. In addition, a position prediction error is often lower (starting from 20 initialization vectors—starting from the third row in Table 6 and Table 7) than a position calculated directly from UWB data (the first row in Table 6 and Table 7—RAW)—error less than 10 cm up to 10 position predictions.
For randomly selected test scenarios from variants C and D, 10 positions were predicted using the VARX model with a lag of 3 and the presented LSTM network—initialization with 20 previous positions. Figure 10 shows the RMSE for simulated data (circles), position predictions using LSTM (crosses), and the VARX model (asterisks).
On average, the RMSE of the LSTM network is 3% lower than for the VARX model. It should be noted that the time needed to determine a single prediction in the VARX model is 8.7 ms, 4.5 times greater than in the case of the LSTM network (which will result in 4.5 times greater displacement of the object during the prediction).

6. Conclusions

The accuracy of the position determination depends on the accuracy of the determined distances between the nodes in the network and the appropriate pre-processing. The computing units used in the vehicles (e.g., cars, AGVs) are computationally efficient, so the accuracy of the determined position is mainly affected by the delay connected with the time needed to measure the distance between the UWB nodes in the network (ranging process) and to make the position available for other purposes. Reduction in data processing and acquisition time is not possible, but it is possible to determine the position prediction based on historical data using the LSTM neural network. Using filters with a moving window for real-time systems is not recommended, considering the continuous movement of the objects.
The conducted experiments and comparisons prove that the prediction of the object’s position is possible using the prepared LSTM neural network with only one hidden layer with high accuracy—it enables the determination of the object’s position with an accuracy of less than 10 cm up to 10 positions (for both dataset of test scenarios) which correspond to 120 ms of continuous positioning of a moving object (which gives 24 cm of the driven distance for AGVs and 1,68 m for motor vehicles). The number of predicted positions can be increased, considering the simultaneous increase in the prediction error. It should be borne in mind that the accuracy of a larger number of predictions can be affected by a rapid change in the object’s motion parameters (direction, acceleration). Therefore, to maintain the system’s high accuracy, the network requires periodic initialization (e.g., with each subsequent data packet from the UWB system).

Author Contributions

Conceptualization, K.P. and D.G.; methodology, K.P. and D.G.; software, K.P.; validation, K.P. and D.G.; formal analysis, K.P. and D.G.; investigation, K.P.; resources, K.P.; data curation, K.P.; writing—original draft preparation, K.P.; writing—review and editing, D.G.; visualization, K.P.; supervision, D.G.; project administration, D.G.; funding acquisition, K.P. and D.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partially funded by the European Social Funds, project no POWR.03.02.00-00-I007/17-00 “CyPhiS—the program of modern PhD studies in the field of Cyber-Physical Systems”; the Norway Grant 2014-2021 contract no.: NOR/POLNOR/CoBotAGV/0027/2019-00 “Automated Guided Vehicles integrated with Collaborative Robots for Smart Industry Perspective”; the Polish Ministry of Education and Science partially by Statutory Research, and partially by Young Researchers funds of Department of Electronics, Electrical Engineering and Microelectronics and the Department of Telecommunications and Teleinformatics, Faculty of Automatic Control, Electronics and Computer Science, Silesian University of Technology.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

d Displacement of the moving object
v Speed of the moving object
t p Time needed to collect the required distances to determine the object’s position
h t Hidden state of the cell
c t Long-term state
f t Forget gate
g t Cell candidate
W f , W g , W i , W o Input weights of, respectively, forget gate, cell candidate, input gate, and output gate
R f , R g , R i , R o Recurrent weights of, respectively, forget gate, cell candidate, input gate, and output gate
b f , b g , b i , b o Bias of, respectively, forget gate, cell candidate, input gate, and output gate
V t Input vector of the LSTM unit
σ g Gate activation function
σ c The state activation function
z The summed weighted input of the cell
v j Element of the neural network input
Δ x Displacement in the X-axis
Δ y Displacement in the Y-axis
a x , a y , a z Acceleration from the accelerometer, respectively, in the X, Y, and Z-axes
g x , g y , g z Angular velocity from the gyroscope, respectively, in the X, Y, and Z-axes
m x , m y , m z Magnetic field from the magnetometer, respectively, in the X, Y, and Z-axes
Θ Network parameters vector
α Learning rate
E Θ Loss function

References

  1. Sullivan, R.; Lert, J.; Toebes, S.C.; Ulrich, N.; Hinshaw, F. Autonomous Transports for Storage and Retrieval Systems 2019. U.S. Patent US8425173B2, 28 September 2023. [Google Scholar]
  2. Raizer, L. Automatic Warehouse System 2019. U.S. Patent 10,259,649, 16 April 2019. [Google Scholar]
  3. Ziebinski, A.; Mrozek, D.; Cupek, R.; Grzechca, D.; Fojcik, M.; Drewniak, M.; Kyrkjebø, E.; Lin, J.C.-W.; Øvsthus, K.; Biernacki, P. Challenges Associated with Sensors and Data Fusion for AGV-Driven Smart Manufacturing. In Computational Science—ICCS 2021; Paszynski, M., Kranzlmüller, D., Krzhizhanovskaya, V.V., Dongarra, J.J., Sloot, P.M.A., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2021; Volume 12745, pp. 595–608. ISBN 978-3-030-77969-6. [Google Scholar]
  4. Cupek, R.; Drewniak, M.; Fojcik, M.; Kyrkjebø, E.; Lin, J.C.-W.; Mrozek, D.; Øvsthus, K.; Ziebinski, A. Autonomous Guided Vehicles for Smart Industries—The State-of-the-Art and Research Challenges. In Computational Science—ICCS 2020; Krzhizhanovskaya, V.V., Závodszky, G., Lees, M.H., Dongarra, J.J., Sloot, P.M.A., Brissos, S., Teixeira, J., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2020; Volume 12141, pp. 330–343. ISBN 978-3-030-50425-0. [Google Scholar]
  5. TRANSDEV. Shared Autonomous Transport Services. Available online: https://www.transdev.com/en/our-innovations/shared-autonomous-mobility/ (accessed on 24 September 2020).
  6. Cao, Z.; Ceder, A.; Zhang, S. Real-Time Schedule Adjustments for Autonomous Public Transport Vehicles. Transp. Res. Part C Emerg. Technol. 2019, 109, 60–78. [Google Scholar] [CrossRef]
  7. Adnan, N.; Md Nordin, S.; bin Bahruddin, M.A.; Ali, M. How Trust Can Drive Forward the User Acceptance to the Technology? In-Vehicle Technology for Autonomous Vehicle. Transp. Res. Part A Policy Pract. 2018, 118, 819–836. [Google Scholar] [CrossRef]
  8. Meyer, J.; Becker, H.; Bösch, P.M.; Axhausen, K.W. Autonomous Vehicles: The next Jump in Accessibilities? Res. Transp. Econ. 2017, 62, 80–91. [Google Scholar] [CrossRef]
  9. Hata, A.Y.; Wolf, D.F. Feature Detection for Vehicle Localization in Urban Environments Using a Multilayer LIDAR. IEEE Trans. Intell. Transport. Syst. 2016, 17, 420–429. [Google Scholar] [CrossRef]
  10. Patole, S.M.; Torlak, M.; Wang, D.; Ali, M. Automotive Radars: A Review of Signal Processing Techniques. IEEE Signal Process. Mag. 2017, 34, 22–35. [Google Scholar] [CrossRef]
  11. Coelingh, E.; Eidehall, A.; Bengtsson, M. Collision Warning with Full Auto Brake and Pedestrian Detection—A Practical Example of Automatic Emergency Braking. In Proceedings of the 13th International IEEE Conference on Intelligent Transportation Systems, Funchal, Portugal, 19–22 September 2010; IEEE: Funchal, Portugal, 2010; pp. 155–160. [Google Scholar]
  12. Harfouch, Y.A.; Yuan, S.; Baldi, S. An Adaptive Switched Control Approach to Heterogeneous Platooning with Intervehicle Communication Losses. IEEE Trans. Control Netw. Syst. 2018, 5, 1434–1444. [Google Scholar] [CrossRef]
  13. Wang, Y.-T.; Peng, C.-C.; Ravankar, A.; Ravankar, A. A Single LiDAR-Based Feature Fusion Indoor Localization Algorithm. Sensors 2018, 18, 1294. [Google Scholar] [CrossRef]
  14. Grzechca, D.; Ziębiński, A.; Paszek, K.; Hanzel, K.; Giel, A.; Czerny, M.; Becker, A. How Accurate Can UWB and Dead Reckoning Positioning Systems Be? Comparison to SLAM Using the RPLidar System. Sensors 2020, 20, 3761. [Google Scholar] [CrossRef]
  15. Zhou, H.; Zou, D.; Pei, L.; Ying, R.; Liu, P.; Yu, W. StructSLAM: Visual SLAM with Building Structure Lines. IEEE Trans. Veh. Technol. 2015, 64, 1364–1375. [Google Scholar] [CrossRef]
  16. Friebe, M.; Petzold, J. Visualisation Functions in Advanced Camera-Based Surround View Systems. In Advanced Microsystems for Automotive Applications 2014; Fischer-Wolfarth, J., Meyer, G., Eds.; Lecture Notes in Mobility; Springer International Publishing: Cham, Switzerland, 2014; pp. 267–276. ISBN 978-3-319-08086-4. [Google Scholar]
  17. Liu, T.; Zhang, J.; Zhang, K.; Xu, J.; Wang, D.; Wang, X. Vehicle-Mounted Surround Vision Algorithm Based on Heterogeneous Architecture. Multimed. Tools Appl. 2020, 79, 24707–24730. [Google Scholar] [CrossRef]
  18. Kim, J.; Jun, H. Vision-Based Location Positioning Using Augmented Reality for Indoor Navigation. IEEE Trans. Consum. Electron. 2008, 54, 954–962. [Google Scholar] [CrossRef]
  19. Greenhalgh, J.; Mirmehdi, M. Real-Time Detection and Recognition of Road Traffic Signs. IEEE Trans. Intell. Transport. Syst. 2012, 13, 1498–1506. [Google Scholar] [CrossRef]
  20. Wang, C.; Zhang, H.; Yang, M.; Wang, X.; Ye, L.; Guo, C. Automatic Parking Based on a Bird’s Eye View Vision System. Adv. Mech. Eng. 2014, 6, 847406. [Google Scholar] [CrossRef]
  21. Ledvina, J.; Vykydal, L.; Horsky, P. Fast Automatic Tuning of a Synthetic Inductor for Automotive Transformer-Less Ultrasonic Sensor in Park Assist Systems. IEEE Sens. J. 2019, 19, 10568–10573. [Google Scholar] [CrossRef]
  22. Ziebinski, A.; Bregulla, M.; Fojcik, M.; Kłak, S. Monitoring and Controlling Speed for an Autonomous Mobile Platform Based on the Hall Sensor. In Computational Collective Intelligence; Nguyen, N.T., Papadopoulos, G.A., Jędrzejowicz, P., Trawiński, B., Vossen, G., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2017; Volume 10449, pp. 249–259. ISBN 978-3-319-67076-8. [Google Scholar]
  23. Stempnik, V.; Choi, B.; Crawford, M.; Shaffer, A.D.; Kane, M.D. Enhanced Alignment Method for Park Assist 2015. U.S. Patent 8957786B2, 17 February 2015. [Google Scholar]
  24. Xiao, L.; Gao, F. A Comprehensive Review of the Development of Adaptive Cruise Control Systems. Veh. Syst. Dyn. 2010, 48, 1167–1192. [Google Scholar] [CrossRef]
  25. Uhlemann, E. Time for Autonomous Vehicles to Connect [Connected Vehicles]. IEEE Veh. Technol. Mag. 2018, 13, 10–13. [Google Scholar] [CrossRef]
  26. Cupek, R.; Ziębiński, A.; Drewniak, M.; Fojcik, M. Application of OPC UA Protocol for the Internet of Vehicles. In Computational Collective Intelligence; Nguyen, N.T., Papadopoulos, G.A., Jędrzejowicz, P., Trawiński, B., Vossen, G., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2017; Volume 10449, pp. 272–281. ISBN 978-3-319-67076-8. [Google Scholar]
  27. Ziebinski, A.; Cupek, R.; Drewniak, M.; Wolny, B. Soft Real-Time Systems for Low-Cost Unmanned Ground Vehicle. In Computational Collective Intelligence; Nguyen, N.T., Chbeir, R., Exposito, E., Aniorté, P., Trawiński, B., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2019; Volume 11684, pp. 196–206. ISBN 978-3-030-28373-5. [Google Scholar]
  28. Hellmers, H.; Kasmi, Z.; Norrdine, A.; Eichhorn, A. Accurate 3D Positioning for a Mobile Platform in Non-Line-of-Sight Scenarios Based on IMU/Magnetometer Sensor Fusion. Sensors 2018, 18, 126. [Google Scholar] [CrossRef]
  29. Schleicher, D.; Bergasa, L.M.; Ocana, M.; Barea, R.; Lopez, M.E. Real-Time Hierarchical Outdoor SLAM Based on Stereovision and GPS Fusion. IEEE Trans. Intell. Transport. Syst. 2009, 10, 440–452. [Google Scholar] [CrossRef]
  30. Narkhede, P.; Poddar, S.; Walambe, R.; Ghinea, G.; Kotecha, K. Cascaded Complementary Filter Architecture for Sensor Fusion in Attitude Estimation. Sensors 2021, 21, 1937. [Google Scholar] [CrossRef]
  31. Grzechca, D.; Paszek, K. Short-Term Positioning Accuracy Based on Mems Sensors for Smart City Solutions. Metrol. Meas. Syst. 2019, 26, 95–107. [Google Scholar] [CrossRef]
  32. Li, J.; Yue, X.; Chen, J.; Deng, F. A Novel Robust Trilateration Method Applied to Ultra-Wide Bandwidth Location Systems. Sensors 2017, 17, 795. [Google Scholar] [CrossRef] [PubMed]
  33. Thomas, F.; Ros, L. Revisiting Trilateration for Robot Localization. IEEE Trans. Robot. 2005, 21, 93–101. [Google Scholar] [CrossRef]
  34. Torrent-Moreno, M.; Mittag, J.; Santi, P.; Hartenstein, H. Vehicle-to-Vehicle Communication: Fair Transmit Power Control for Safety-Critical Information. IEEE Trans. Veh. Technol. 2009, 58, 3684–3703. [Google Scholar] [CrossRef]
  35. Guo, L.; Chen, X.; Parini, C.G. Miniature Ultra-Wideband Antenna for Wireless Universal Serial Bus Dongle Applications. IET Microw. Antennas Propag. 2012, 6, 113. [Google Scholar] [CrossRef]
  36. Alarifi, A.; Al-Salman, A.; Alsaleh, M.; Alnafessah, A.; Al-Hadhrami, S.; Al-Ammar, M.; Al-Khalifa, H. Ultra Wideband Indoor Positioning Technologies: Analysis and Recent Advances. Sensors 2016, 16, 707. [Google Scholar] [CrossRef]
  37. Digital Key—The Future of Vehicle Access; Car Connectivity Consortium: Beaverton, OR, USA, 2020.
  38. Fontana, R.J. Recent System Applications of Short-Pulse Ultra-Wideband (UWB) Technology. IEEE Trans. Microw. Theory Techn. 2004, 52, 2087–2104. [Google Scholar] [CrossRef]
  39. Gabela, J.; Retscher, G.; Goel, S.; Perakis, H.; Masiero, A.; Toth, C.; Gikas, V.; Kealy, A.; Koppányi, Z.; Błaszczak-Bąk, W.; et al. Experimental Evaluation of a UWB-Based Cooperative Positioning System for Pedestrians in GNSS-Denied Environment. Sensors 2019, 19, 5274. [Google Scholar] [CrossRef]
  40. Jiménez, A.R.; Seco, F. Improving the Accuracy of Decawave’s UWB MDEK1001 Location System by Gaining Access to Multiple Ranges. Sensors 2021, 21, 1787. [Google Scholar] [CrossRef]
  41. López Medina, M.Á.; Espinilla, M.; Paggeti, C.; Medina Quero, J. Activity Recognition for IoT Devices Using Fuzzy Spatio-Temporal Features as Environmental Sensor Fusion. Sensors 2019, 19, 3512. [Google Scholar] [CrossRef]
  42. Monica, S.; Ferrari, G. Improving UWB-Based Localization in IoT Scenarios with Statistical Models of Distance Error. Sensors 2018, 18, 1592. [Google Scholar] [CrossRef]
  43. Hernandez-Matamoros, A.; Fujita, H.; Hayashi, T.; Perez-Meana, H. Forecasting of COVID19 per Regions Using ARIMA Models and Polynomial Functions. Appl. Soft Comput. 2020, 96, 106610. [Google Scholar] [CrossRef] [PubMed]
  44. Siami-Namini, S.; Tavakoli, N.; Namin, A.S. A Comparison of ARIMA and LSTM in Forecasting Time Series. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; IEEE: New York, NY, USA, 2018; pp. 1394–1401. [Google Scholar]
  45. Elsworth, S.; Güttel, S. Time Series Forecasting Using LSTM Networks: A Symbolic Approach. arXiv 2020, arXiv:2003.05672. [Google Scholar] [CrossRef]
  46. Dowell, J.; Pinson, P. Very-Short-Term Probabilistic Wind Power Forecasts by Sparse Vector Autoregression. IEEE Trans. Smart Grid 2015, 7, 763–770. [Google Scholar] [CrossRef]
  47. Hölzke, F.; Borstell, H.; Golatowski, F.; Haubelt, C. Pedestrian Localization with Stride-Wise Error Estimation and Compensation by Fusion of UWB and IMU Data. Sensors 2023, 23, 4744. [Google Scholar] [CrossRef]
  48. Li, Y.; Gao, Z.; Xu, Q.; Yang, C. Comprehensive Evaluations of NLOS and Linearization Errors on UWB Positioning. Appl. Sci. 2023, 13, 6187. [Google Scholar] [CrossRef]
  49. Wang, X.; Gao, L.; Mao, S.; Pandey, S. CSI-Based Fingerprinting for Indoor Localization: A Deep Learning Approach. IEEE Trans. Veh. Technol. 2016, 66, 763–776. [Google Scholar] [CrossRef]
  50. Wang, R.; Luo, H.; Wang, Q.; Li, Z.; Zhao, F.; Huang, J. A Spatial–Temporal Positioning Algorithm Using Residual Network and LSTM. IEEE Trans. Instrum. Meas. 2020, 69, 9251–9261. [Google Scholar] [CrossRef]
  51. Wang, F.; Tang, H.; Chen, J. Survey on NLOS Identification and Error Mitigation for UWB Indoor Positioning. Electronics 2023, 12, 1678. [Google Scholar] [CrossRef]
  52. Kim, D.-H.; Farhad, A.; Pyun, J.-Y. UWB Positioning System Based on LSTM Classification with Mitigated NLOS Effects. IEEE Internet Things J. 2023, 10, 1822–1835. [Google Scholar] [CrossRef]
  53. Zhao, Y.; Wang, M. The LOS/NLOS Classification Method Based on Deep Learning for the UWB Localization System in Coal Mines. Appl. Sci. 2022, 12, 6484. [Google Scholar] [CrossRef]
  54. Sang, C.L.; Steinhagen, B.; Homburg, J.D.; Adams, M.; Hesse, M.; Rückert, U. Identification of NLOS and Multi-Path Conditions in UWB Localization Using Machine Learning Methods. Appl. Sci. 2020, 10, 3980. [Google Scholar] [CrossRef]
  55. Yang, S.; Ben, T.-B.; Xiang, W. Build Up a Real-Time LSTM Positioning Error Prediction Model for GPS Sensors. In Proceedings of the 2019 IEEE 90th Vehicular Technology Conference (VTC2019-Fall), Honolulu, HI, USA, 22–25 September 2019; IEEE: Honolulu, HI, USA, 2019; pp. 1–5. [Google Scholar]
  56. Gao, D.; Zeng, X.; Wang, J.; Su, Y. Application of LSTM Network to Improve Indoor Positioning Accuracy. Sensors 2020, 20, 5824. [Google Scholar] [CrossRef]
  57. Poulose, A.; Han, D.S. UWB Indoor Localization Using Deep Learning LSTM Networks. Appl. Sci. 2020, 10, 6290. [Google Scholar] [CrossRef]
  58. Zhang, Y.; Qu, C.; Wang, Y. An Indoor Positioning Method Based on CSI by Using Features Optimization Mechanism with LSTM. IEEE Sens. J. 2020, 20, 4868–4878. [Google Scholar] [CrossRef]
  59. Paszek, K.; Grzechca, D.; Becker, A. Design of the UWB Positioning System Simulator for LOS/NLOS Environments. Sensors 2021, 21, 4757. [Google Scholar] [CrossRef] [PubMed]
  60. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
  61. Ordóñez, F.; Roggen, D. Deep Convolutional and LSTM Recurrent Neural Networks for Multimodal Wearable Activity Recognition. Sensors 2016, 16, 115. [Google Scholar] [CrossRef]
  62. Yu, H.; Wu, Z.; Wang, S.; Wang, Y.; Ma, X. Spatiotemporal Recurrent Convolutional Networks for Traffic Prediction in Transportation Networks. Sensors 2017, 17, 1501. [Google Scholar] [CrossRef]
  63. Steven Eyobu, O.; Han, D. Feature Representation and Data Augmentation for Human Activity Classification Based on Wearable IMU Sensor Data Using a Deep LSTM Neural Network. Sensors 2018, 18, 2892. [Google Scholar] [CrossRef]
  64. Paszek, K.; Grzechca, D.; Tomczyk, M.; Marciniak, A. UWB Positioning System with the Support of MEMS Sensors for Indoor and Outdoor Environment. J. Commun. 2020, 15, 511–518. [Google Scholar] [CrossRef]
Figure 1. Displacement of the object depending on object speed and data acquisition time.
Figure 1. Displacement of the object depending on object speed and data acquisition time.
Sensors 23 08270 g001
Figure 2. An example system scenario.
Figure 2. An example system scenario.
Sensors 23 08270 g002
Figure 3. The simulation environment with sample path: anchors—red dots; reference path—red line; and path determined using the simulated UWB system—blue line.
Figure 3. The simulation environment with sample path: anchors—red dots; reference path—red line; and path determined using the simulated UWB system—blue line.
Sensors 23 08270 g003
Figure 4. The memory cell structure of the LSTM neural network.
Figure 4. The memory cell structure of the LSTM neural network.
Sensors 23 08270 g004
Figure 5. Sorted and trimmed data series (test scenarios)—the red indicates the series’ length, and the blue indicates removed (truncated) data from the series in the batch.
Figure 5. Sorted and trimmed data series (test scenarios)—the red indicates the series’ length, and the blue indicates removed (truncated) data from the series in the batch.
Sensors 23 08270 g005
Figure 6. Structure of the LSTM neural network.
Figure 6. Structure of the LSTM neural network.
Sensors 23 08270 g006
Figure 7. The example of the neural network prediction, (a)—no update from the UWB system over 10 positioning periods, (b)—update from the UWB system every positioning period.
Figure 7. The example of the neural network prediction, (a)—no update from the UWB system over 10 positioning periods, (b)—update from the UWB system every positioning period.
Sensors 23 08270 g007
Figure 8. The RMSE histogram of the 10 position prediction.
Figure 8. The RMSE histogram of the 10 position prediction.
Sensors 23 08270 g008
Figure 9. The RMSE up to the 10th prediction.
Figure 9. The RMSE up to the 10th prediction.
Sensors 23 08270 g009
Figure 10. The RMSE up to the 10th prediction—VARX and LSTM.
Figure 10. The RMSE up to the 10th prediction—VARX and LSTM.
Sensors 23 08270 g010
Table 1. Test scenarios summary.
Table 1. Test scenarios summary.
Test ParameterValue
Acceleration [m/s2]0; 0.1; 0.2; 0.3; 0.4; 0.5; 0.6; 0.7; 0.8; 0.9; 1; 2; 3; 4; 5; 6; 7; 8; 9; 10
Initial speed [m/s]0; 0.25; 0.5; 0.75; 1; 1.25; 1.5; 1.75; 2; 2.5; 3; 3.5; 4; 5; 6; 7; 8; 9; 10; 11; 12; 13; 14; 15; 16; 17; 18; 19; 20
Direction of movement [rad] 0 ,   π , 2 π ± π 10 , π 9 , π 8 , π 7 , π 6 , π 5 , π 4 , π 3
Path length [m]20
Table 2. RMSE of the 10 prediction positions of the moving object at different stages of the path.
Table 2. RMSE of the 10 prediction positions of the moving object at different stages of the path.
Number of Memory CellsVariant of MovementRMSE [cm]
80%50%30%Avg
20A6.236.409.467.36
B5.766.499.817.36
70A5.765.228.366.45
B5.136.219.326.89
100A5.684.577.776.01
B4.495.759.296.51
150A5.784.377.315.82
B4.145.109.456.23
Table 3. Cumulative RMSE of the object position predictions.
Table 3. Cumulative RMSE of the object position predictions.
Number of Predictions
151020406080100
RMSE [cm]6.536.827.5410.4421.4939.1263.3194.25
Table 4. The RMSE of the position prediction for uniform motion.
Table 4. The RMSE of the position prediction for uniform motion.
v [m/s]RMSE [cm]
12345678910
0.254.904.985.065.145.245.355.475.605.745.89
0.53.713.763.813.873.934.014.104.204.324.44
0.753.403.453.483.543.603.693.793.904.024.17
12.632.692.742.792.872.953.053.163.293.42
1.252.963.003.043.083.133.203.273.353.443.54
1.53.203.213.243.283.333.393.463.543.633.74
1.753.333.383.463.573.723.894.094.324.574.85
23.223.303.423.573.723.904.114.344.584.86
2.52.592.672.742.842.963.123.303.513.743.98
32.562.602.672.752.863.003.173.393.633.89
3.52.992.972.973.023.113.233.383.563.763.99
43.133.133.183.283.433.643.884.164.474.82
53.062.982.952.993.113.293.533.834.164.54
63.193.223.313.493.744.064.424.835.265.73
73.333.343.423.623.914.294.715.195.706.25
83.303.193.233.463.774.204.685.235.816.45
93.993.773.703.773.944.224.585.005.476.01
104.364.093.963.984.154.444.815.285.806.38
114.674.314.124.114.264.564.965.436.006.62
125.905.404.994.734.634.714.935.295.746.27
136.295.895.555.385.355.465.686.016.406.87
146.385.805.375.145.115.275.616.066.637.29
156.796.205.755.475.415.535.826.246.767.35
167.547.046.636.376.246.266.416.667.017.44
178.397.817.276.856.596.506.556.767.067.47
189.408.758.237.777.417.177.077.127.307.60
1910.029.488.918.448.077.827.707.717.848.08
2010.089.418.958.458.037.687.417.247.177.21
Avg4.834.644.504.454.494.604.785.035.335.68
Table 5. The RMSE of the position prediction for uniformly accelerated motion.
Table 5. The RMSE of the position prediction for uniformly accelerated motion.
a [m/s2]RMSE [cm]
12345678910
0.12.742.762.772.792.842.913.003.123.253.39
0.22.952.902.872.832.822.832.872.922.993.08
0.32.862.862.923.023.163.333.543.754.014.27
0.43.933.923.923.913.913.913.933.964.004.08
0.53.203.123.042.972.922.902.892.932.993.08
0.63.673.573.493.443.403.373.363.383.423.49
0.73.623.493.393.293.203.123.063.033.013.01
0.83.743.703.633.563.513.493.483.513.553.62
0.93.973.863.763.693.623.583.563.583.643.72
14.214.103.993.903.823.773.753.753.793.85
25.195.004.824.674.554.444.374.314.294.30
35.795.605.405.225.054.914.814.744.724.75
47.066.956.856.726.606.486.366.256.166.08
56.446.215.945.695.435.215.024.884.784.76
67.236.936.636.336.095.855.675.545.465.44
77.617.276.946.626.285.975.705.475.315.22
88.338.077.747.437.146.846.596.386.246.16
98.768.498.167.837.487.156.856.586.386.24
109.499.218.888.588.257.927.617.317.056.83
Avg5.305.165.014.874.744.634.554.494.484.49
Table 6. The RMSE of the position prediction for variant C—lines and arcs.
Table 6. The RMSE of the position prediction for variant C—lines and arcs.
Number of V RMSE [cm]
12345678910
RAW10.3310.3510.3410.3510.3610.3810.3910.4310.4510.48
107.087.808.649.5010.4011.3312.2813.2714.2815.32
205.265.405.595.836.126.466.847.277.758.27
304.995.075.205.395.645.956.326.757.247.80
404.884.854.864.925.035.205.425.706.036.42
504.884.834.804.814.884.995.165.395.665.99
804.404.384.394.444.534.664.845.065.325.63
Table 7. The RMSE of the position prediction for variant D—lines and arcs.
Table 7. The RMSE of the position prediction for variant D—lines and arcs.
Number of V RMSE [cm]
12345678910
RAW9.669.649.629.609.599.609.619.629.659.67
106.366.897.628.439.3410.3211.3912.5313.7415.03
205.425.646.026.507.077.748.509.3510.2911.32
305.355.555.926.396.967.638.409.2710.2411.30
405.155.185.345.575.896.296.777.347.998.72
505.245.185.225.335.525.776.116.527.017.57
805.225.085.055.075.155.285.485.756.076.47
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Paszek, K.; Grzechca, D. Using the LSTM Neural Network and the UWB Positioning System to Predict the Position of Low and High Speed Moving Objects. Sensors 2023, 23, 8270. https://doi.org/10.3390/s23198270

AMA Style

Paszek K, Grzechca D. Using the LSTM Neural Network and the UWB Positioning System to Predict the Position of Low and High Speed Moving Objects. Sensors. 2023; 23(19):8270. https://doi.org/10.3390/s23198270

Chicago/Turabian Style

Paszek, Krzysztof, and Damian Grzechca. 2023. "Using the LSTM Neural Network and the UWB Positioning System to Predict the Position of Low and High Speed Moving Objects" Sensors 23, no. 19: 8270. https://doi.org/10.3390/s23198270

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop