Next Article in Journal
A Reliability-Oriented Framework for the Preservation of Historical Railway Assets Under Regulatory and Material Uncertainty
Previous Article in Journal
GaussianMix: Rethinking Receptive Field for Efficient Data Augmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Reconstruction Method for Irregularly Sampled Observation Sequences for Digital Twin

1
Faculty of Electronic and Information Engineering, Xi’an Jiaotong University, Xi’an 710049, China
2
Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing 100089, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(9), 4706; https://doi.org/10.3390/app15094706
Submission received: 21 March 2025 / Revised: 21 April 2025 / Accepted: 21 April 2025 / Published: 24 April 2025
(This article belongs to the Section Applied Thermal Engineering)

Abstract

:
Various uncertainties such as communication delay, packet loss and disconnection in the Industrial Internet, as well as the asynchronous sampling of sensors, can cause irregularity, sparsity, and misalignment of sampling sequences, and thereby seriously affect the training and prediction performance of a digital twin model. Sequence reconstruction is an effective way to deal with the above problems, but if the measurement data become sparse or contain significant noise due to packet loss and electromagnetic interference, existing methods struggle to achieve ideal results. Therefore, a novel variational autoencoder model based on a parallel reference network and neural controlled differential equation (PRN-NCDE) is proposed in this article to solve the problem of reconstructing irregular series under sparse measurements and high noise levels. First, a multi-channel self-attention module is established, which can not only analyze the position and feature information of the sampled data to improve the reconstruction accuracy under sparse measurements, but also effectively tackle the misalignment and irregularity of the observation sequence through multi-channel and mask mechanisms. Second, to improve the accuracy of sequence reconstruction under large noise levels, a PRN is established to obtain reference features, which are weighted and fused with the features of observed data. Third, we use the NCDE to construct a decoder that can combine the control input of the system to predict the output values to solve the problem of sequence reconstruction in a controlled system. Finally, a weighted loss function is constructed to better train the network parameters of the model. This article takes the furnace of the boiler system in a coal-fired power plant as the test object to verify the effectiveness and fitting accuracy of the proposed PRN-NCDE model compared to the existing methods for a controlled system under sparse measurements and large noise levels. Simulation results show that the proposed PRN-NCDE model can improve the estimation accuracy by more than 50% and 70% compared with the recurrent neural network-NCDE (RNN-NCDE) under different sampling numbers and noise levels, and by more than 80% and 60% compared with the recurrent neural network-NODE (RNN-NODE).

1. Introduction

As one of the key technologies for the digital transformation of traditional industries, digital twin can create virtual mappings of physical entities, which enables real-time reflection of the entire lifecycle of physical systems and facilitates their optimization and monitoring [1]. Maintaining consistency between the twin model and the physical system is crucial for the successful implementation of digital twins’ functionalities [2]. However, several factors can degrade the training and prediction performance of twin models. First, network communication often has irregularities such as delays and packet loss, leading to irregular sampling intervals [3]. Second, sensors in physical systems have varying sampling periods, causing misalignment and sparsity in multivariate sequences. These issues make the sampled sequences unsuitable for model training and produce negative impacts on prediction performance [4].
Sequence reconstruction is an effective means for these problems in digital twin systems. It involves rebuilding irregular observation sequences using function fitting or generative neural networks to obtain regular sequence data. Traditional methods include polynomial fitting [5] and maximum likelihood estimation [6]. Polynomial fitting approximates complex functions by learning suitable polynomial coefficients but requires predefining the polynomial degree [7,8]. A higher predefined polynomial degree may lead to overfitting and a lower one may have negative impact on the fitting performance. For complex and ever-changing observation sequences, it is often difficult to make an assumption on its degree. Moreover, polynomial fitting is very sensitive to noise, resulting in its inapplicability for noise-corrupted observation sequences [9]. Based on the curve fitting method in statistics, maximum likelihood estimation determines model parameters by calculating likelihood function values but shares similar limitations with polynomial fitting, such as narrow applicability and the need to assume a model form. Narrow applicability means this method can only produce satisfactory results under some certain distributions [10]. And an assumed model form will prevent the method from dealing well with complex sequences influenced by control inputs. Fixed-time discretization is another conventional approach that divides continuous-time observations into fixed-length, non-overlapping data windows [11]. Although this method is easy to follow and utilize, it may induce empty windows and data aggregation issues when window sizes are increased [12]. Therefore, window size is a key parameter for this method. However, determining an appropriate window size for complex sequences is often difficult.
Alternative solutions involve developing neural network models that can directly use irregular sequences as inputs, including interpolation-based and neural ordinary differential equation (NODE)-based methods. Interpolation-based methods use past and future observations for interpolation. Yoon et al. [13] studied a multi-directional recurrent neural network (RNN) that can use past and future observation data flows at a given time step to realize interpolation, thus improving estimation performance in missing measurement scenarios. Shukla et al. [14] developed an interpolation prediction network, which consists of several semi-parametric radial basis function (RBF) interpolation layers. This interpolation network can interpolate multivariate and irregularly time series against a set of reference time points. In [15,16,17], attention mechanism is used for time encoding and several multi-time attention-based RNNs are proposed for irregular sequence modeling. However, these interpolation-based methods face variable uncertainty when observation intervals change significantly [18]. Exponential decayed RNN autoencoder frameworks for sequence reconstruction were studied in [19,20], but Mozer et al. [21] found these networks have no obvious improvements in prediction accuracy compared to the standard RNN autoencoders. In 2018, Chen et al. introduced the NODE network [22], offering new solutions for irregular sequence reconstruction. Subsequently, Rubanova et al. [23] developed a latent NODE model, which combines RNN and NODE in a variational autoencoder (VAE) framework, outperforming the exponential decay RNNs. The GRU-ODE-Bayes model proposed by Brouwer et al. [24] achieved good results on sparse data. Huang et al. [25] presented LG-ODE, which is a latent ordinary differential equation VAE for multi-agent systems with known graph structure. LG-ODE can learn the embedding of high-dimensional trajectories and deduce the latent system dynamics simultaneously. In practical applications, most systems have state trajectories influenced not only by initial values but also by control inputs. NODEs cannot account for control inputs’ impact on output sequences, limiting their application in controlled system sequence reconstruction. Moreover, when the measurement data are sparse or noisy due to network issues and electromagnetic interference, existing methods struggle to handle digital twin sequence reconstruction tasks.
To address these challenges, this article proposes a VAE model based on a parallel reference network (PRN) and neural controlled differential equation (NCDE) [26] for noisy multivariate irregular sequence reconstruction in controlled systems. First, we establish a multi-channel self-attention (MCSA) module to analyze the position information and correlations of sampled data in the target sequence, improving reconstruction accuracy under sparse measurements while effectively handling misalignment and irregularity through multi-channel and masking mechanisms. Second, to enhance reconstruction accuracy under high noise levels, we construct a PRN to obtain reference features from the prediction results of digital twin model and fuse them with actual data features. We also calculate feature weights based on the noise level of the observation sequence to determine the proportion of actual and reference information in the fused features. Third, we use NCDE to build a decoder that can predict observations at any time by incorporating control inputs to solve the sequence reconstruction problem in controlled systems. Finally, we develop a weighted loss function based on feature weights to better train the model’s network parameters. Simulation experiments demonstrate the effectiveness and fitting accuracy of the proposed model for controlled systems under sparse measurements and high noise levels compared to the existing methods.
Overall, the main contributions of this article are threefold:
(1)
We propose a PRN-NCDE-based VAE model that improves sequence reconstruction accuracy for controlled systems under sparse measurements and high noise levels.
(2)
We develop an MCSA module that can not only analyze data position and correlations to enhance reconstruction performance under sparse measurements, but also effectively handle misaligned and irregular observation sequences.
(3)
To improve reconstruction accuracy under high noise levels, we establish a PRN to obtain reference features and calculate feature weights based on the noise level of observation data for weighted fusion of latent features.
The rest of this article is organized as follows. Section 2 formally describes the irregular sequence reconstruction problem. Section 3 presents the overall framework of the PRN-NCDE model and analyzes the MCSA module, PRN, NCDE network, and weighted loss function. Section 4 validates effectiveness of the proposed model compared to the existing methods through a boiler system under different sampling numbers and high noise levels. Conclusions and future work are discussed in Section 5.

2. Problem Formulation

In this section, we first give formal descriptions of multivariate regular and irregular observation sequences.
Definition 1.
Consider an observation dataset  𝒟 = y n , t n n = 0 N , where  y n d is the observation vector at sampling time  t n , and  N + 1 is the total number of samples. Let NaN denote the missing data in the observation vector. And when an observation vector has missing data, its dimension satisfies  dim y n < d . Then, if dataset 𝒟 meets the following two conditions:
(1)
For any two consecutive observation vectors y n and y n + 1 in 𝒟, the sampling interval Δ t n = t n + 1 t n is a constant C , i.e., t n + 1 t n C for n 0 , N .
(2)
No observation vector y n in 𝒟 contains missing data NaN, i.e., dim y n = d for n 0 , N .
Then, 𝒟 forms a regular sampled time series. Conversely, if 𝒟 does not satisfy both conditions, it forms an irregular sampled time series.
Condition (1) in Definition 1 describes the regularity of the sampling sequence distribution on the time dimension, that is, whether a fixed sampling interval is followed. Condition (2) describes whether the sampling time of every variable in the multivariate observation sequence is aligned. Only when 𝒟 satisfies both conditions, which means that the sampling times of all the variables are aligned and there is an unified and fixed sampling interval, it can be regarded as a regular multivariate observation sequence. If either of the conditions is not met, it is considered as an irregular multivariate observation sequence. Overall, a regular multivariate sampling sequence needs to meet both conditions: a fixed sampling interval and aligned sampling times.
Due to various uncertainties in communication networks and non-synchronous sampling of sensors, twin systems usually receive non-aligned, irregular sampling sequences. Consider the nonlinear discrete-time system corrupted by measurement noise as follows:
x n + 1 = f x n , u n y n + 1 = h x n + 1 , u n + v n
where x n s , u n c , and y n d are the state vector, control vector, and observation vector of the system at sampling time step t n , respectively. f : s × c s and h : s × c d are the nonlinear state transition function and measurement function, respectively. For system (1), we make the following assumptions:
Assumption 1.
The process noise of the system is ignored, and it is considered that system (1) only contains measurement noise  v n , which follows a Gaussian distribution with mean  0 and covariance  R .
Assumption 2.
Due to the uncertainties of network communication and non-synchronous sampling of sensors, the time series formed by the observations  y n of system (1) is no longer regular. Instead, it is a non-aligned irregular time series with non-aligned sampling times and non-fixed sampling intervals.
Based on the above descriptions, the main objective of this article is as follows: Given a set of noisy, non-aligned and irregular observation sequences y n i r r , t n n = 0 N from a controlled system, we want to develop a VAE network model that can maximize the evidence lower bound (ELBO) given by
max W e c , W d c 𝔼 x 0 p e x 0 | y n i r r , t n n = 0 N log p d y ^ 0 i r r , , y ^ N i r r | x 0 KL p e x 0 | y n i r r , t n n = 0 N | | p x 0
to train the encoder network weights W e c and decoder network weights W d c , as shown in Figure 1. y ^ n i r r is the estimated observation vector from the neural network model, corresponding to the original observation vector y n i r r ; p e is the conditional distribution that the encoder network needs to approximate; p d is the conditional distribution that the decoder needs to approximate; p x 0 is the prior distribution. Once the optimal sequence reconstruction model is trained, it can predict observations at any desired time and produce the regular sequence needed y n r , t n n = 0 N by the twin system. Here, y n r is the estimated observation vector at desired time t n ; N + 1 is the number of samples in the regular sequence.
The next section will focus on the reconstruction method for noisy irregular sequences in controlled systems and propose a VAE network model based on PRN-NCDE to obtain regular sequences with higher fitting accuracy.

3. Irregularly Sampled Observation Sequence Reconstruction Method Based on PRN-NCDE

This section constructs a deep generative network model based on PRN-NCDE within the VAE framework, which mainly consists of an encoder and a decoder. First, the overall network framework is established, and the workflow of the model and the functions of each module are described. Then, the establishment of the MCSA module in the encoder, the PRN and the calculation method of latent feature weights, as well as the construction method of the NCDE network in the decoder, are analyzed in detail, and the corresponding weighted loss function is established. Finally, the approximation ability of the reconstruction model for the observation sequence is discussed. We refer the readers to [27,28], and [26,29] for detailed theories and implementations of VAE, MCSA and NCDE.

3.1. Overall Network Framework of PRN-NCDE

The overall network framework of the sequence reconstruction model, as shown in Figure 2, consists primarily of an encoder and a decoder. The encoder comprises two identical network structures that map the actual observational sequence and the reference observational sequence into latent features, subsequently merging them through weighted fusion based on the noise level of the actual observational data. The decoder, consisting of an NCDE network and an output layer, utilizes these fused latent features to reconstruct the desired regular sequence.
Initially, the input of the encoder is made up by the actual irregular observation sequence and the reference observation sequence, each passing through the identical neural network architecture to produce the actual and reference latent features, respectively. In order to enhance reconstruction accuracy under sparse measurement conditions, a multi-head self-attention mechanism is adopted to process both sequences, focusing on the position information of the measurement data and capturing the correlations among them. The multi-channel self-attention mechanism specifically addresses the irregularity and non-alignment issues in actual observation sequences.
Taking the coal-fired power plant boiler system as an illustrative example, multivariate irregular observational sequences are first expanded to equal-length sequences based on desired sampling intervals to deal with irregular sampling intervals. Furthermore, discrepancies in sensor sampling times and frequencies lead to non-alignment among variables such as gas density, gas temperature, and gas oxygen content, causing inconsistent positions of missing values across different variables. Therefore, the expanded sequence is segmented along variable dimensions and input into corresponding masked multi-head self-attention modules, where masks indicate missing values, preventing negative impacts during network training and prediction. Subsequently, the output sequence from the self-attention module undergo group normalization and reversal before being fed into an LSTM network for reverse-time inference from time t N to t 0 , obtaining the hidden state at the initial time step t 0 . A fully connected layer maps this hidden state to the mean and variance of the latent feature distribution, and by combining with noise sampled from a standard Gaussian distribution can generate latent features.
Similarly, reference observation sequences generated by a digital twin model undergo processing through an identical PRN structure to obtain reference latent features. Incorporating PRN to acquire the reference latent features provides prior information, thereby enhancing the reconstruction model’s fitting performance under higher noise conditions. Higher noise levels would increase the noise information in the latent features from the actual observation sequences, significantly impairing the decoder’s effectiveness. Hence, weights assigned to latent features are set as a nonlinear function of observation sequence noise variance. With noise level increasing, we should reduce the weight of latent features of the actual observation sequence and rely more on the reference latent features. Noise variance is approximated from the difference between actual and reference sequences. The resulting weighted fusion feature x 0 f u s is subsequently used as the decoder input.
The decoder can be viewed as the inverse process of the encoder, reconstructing observation data from latent features through neural networks. Compared with NODE models that rely solely on the initial value x 0 f u s , the NCDE-based decoder, which takes the influence of the control inputs into consideration during sequence reconstruction, can deal with the sequence reconstruction of a controlled system better. The NCDE network includes multi-channel fully connected layers and activation function layers linked to corresponding control derivative channels to perform element-wise multiplication with control derivatives. Summation across channels produces the NCDE output. The number of channels in the NCDE network is determined by the dimensionality of control variables c . Ultimately, the NCDE-generated predictions are mapped via the output layer to yield the desired regular data sequence.

3.2. Multi-Channel Self-Attention Module

The multi-channel self-attention (MCSA) module comprises multiple masked multi-head self-attention modules designed to handle the misalignment and irregularities inherent in multivariate observation sequences. Initially, the observation sequence of each dimension within the coal-fired power plant is expanded to a fixed-length sequence at a desired sampling interval. Consequently, for the irregular observation vector y n i r r , certain dimensional observations will be missing. These missing observations can be filled using a constant value τ , represented as follows:
y n c = y n i r r m + τ m
where y n c d is the observation vector after filling. m 0 , 1 d is an indicator function taking values of either 0 or 1, with 1 denoting the presence of an observation and 0 denoting its absence. m is the vector formed by inverting elements of m , and represents element-wise multiplication.
Therefore, the irregular observation sequence can be expanded into a regular sequence with the desired length N + 1 , represented as follows:
Y = y 0 c , , y n c , , y N c
where Y d × N + 1 represents the expanded multivariate sequence containing the filled values τ .
Additionally, due to the non-alignment of multivariate observation sequence in the boiler system, the filled positions of missing values are not entirely consistent. Thus, the number of MCSA self-attention channels should correspond to the dimensionality of observational variable d , and the sequence Y must be divided according to variable dimensions (i.e., the rows of Y ), and input into their corresponding self-attention module channels. Taking observational variables in the boiler system as an example, division by variable dimension can be expressed as
ρ b , T g s , , O c p T = Block Row Y = y 0 c , , y n c , , y N c
where ρ b , T g s , O c p 1 × N + 1 represent the time series with length N + 1 of gas density, gas temperature, and gas oxygen content. Block Row represents row-wise partitioning of the matrix. Since the filled positions for each variable are inconsistent, the divided sequences are separately input into corresponding multi-head self-attention modules, represented as follows:
s 1 , s 2 , , s d T = MCSA M H 1 ρ b , M H 2 T g s , , M H d O c p T
with
M H i s i = W o i s a t t Q i , 1 , K i , 1 , V i , 1 , , s a t t Q i , H , K i , H , V i , H
h 1 , , H ,   Q i , h = W q i , h s i ,   K i , h = W k i , h s i ,   V i , h = W v i , h s i
where s i represents the time series of the i-th variable, s i is the output row vector of MCSA, M H i represents the multi-head attention operation of the i-th channel, W o i 1 × H v is the output projection matrix of the i-th channels and s a t t Q i , h , K i , h , V i , h is the self-attention operation of the h-th head of the i-th channel. W q i , h k × 1 , W k i , h k × 1 , and W v i , h v × 1 are the corresponding projection matrices.
The output results of each self-attention channel are concatenated along the dimension direction and written as Y = concat s 1 , s 2 , , s d , where Y d × N + 1 represents the concatenated matrix and concat represents the concatenation operation.

3.3. Weighted Fusion of Latent Features and Weight Calculation

Sensor data in twin systems can have significant measurement noise covariance due to electromagnetic interference. Existing methods struggle with reconstructing irregular observation sequences with high noise levels. Large measurement noise will heavily contaminate the actual observation data, which means that the real distribution of the measurement information cannot be well learned. An effective means for dealing with this issue is to incorporate reference latent features into the model training and prediction process, and fuse them with latent features from noisy actual observations, thus reducing the impact of noise on reconstruction results.
Therefore, we utilize the prediction results of a high-fidelity digital twin model as a reference observation sequence, and construct a PRN with the same network structure to obtain reference latent features. We can assume that the actual and reference observation sequences are mapped to latent states x 0 o b s and x 0 r e f , respectively, through identical neural network structures, with their posterior distributions following normal distributions given by
p e x 0 o b s | y n i r r , t n n = 0 N = 𝒩 x 0 o b s | μ 0 o b s , σ 0 2 , o b s p e x 0 r e f | y n r e f , t n n = 0 N = 𝒩 x 0 r e f | μ 0 r e f , σ 0 2 , r e f
with
μ 0 o b s = fnn ɀ 0 o b s σ 0 2 , o b s = exp fnn ɀ 0 o b s μ 0 r e f = fnn ɀ 0 r e f σ 0 2 , r e f = exp fnn ɀ 0 r e f
where μ 0 o b s and μ 0 r e f denote the mean values of x 0 o b s and x 0 r e f , σ 0 2 , o b s and σ 0 2 , r e f are the corresponding covariances. ɀ 0 o b s and ɀ 0 r e f are the initial hidden states from two LSTM networks. y n i r r , t n n = 0 N and y n r e f , t n n = 0 N represent the actual irregular observation sequence and the reference observation sequence generated by the model, respectively. fnn is the feedforward neural network transforming the initial hidden state ɀ 0 into μ 0 and σ 0 2 , and exp denotes the exponential operation which can ensure the covariance matrix is positive definite.
By using the reparameterization trick, x 0 o b s and x 0 r e f can be written as
x 0 o b s = μ 0 o b s + σ 0 o b s ς x 0 r e f = μ 0 r e f + σ 0 r e f ς
where ς 𝒩 0 , I is sampled from the standard normal distribution.
Then, we need to determine reasonable weight allocation criteria for the weighted fusion of latent features. By introducing reference latent features, the reconstruction performance under large noise conditions can be improved. Therefore, the noise level of the actual observation sequence can be used for weight allocation. When noise is small, the weight of the actual observation sequence should be increased to capture more of its feature information. Conversely, when noise is large, the weight of the reference latent features should be increased to avoid introducing excessive measurement noise. Assuming the measurement noise follows a Gaussian distribution with mean 0 and covariance R , we estimate the noise covariance using the reference observation sequence:
R ^ = 1 N n = 0 N y ˜ n i r r y ˜ n r e f y ˜ n i r r y ˜ n r e f T
where y ˜ n i r r and y ˜ n r e f are the normalized actual and reference observations.
Let R ^ max = max diag R ^ be the largest diagonal element of R ^ . The reference feature weight is then calculated by
Q = s 1 R ^ max Q = s 2 R ^ max + c Q = q 2 R ^ max < r 1 r 1 R ^ max r 2 R ^ max > r 2
where s 1 = q 1 / r 1 , s 2 = q 2 q 1 / r 2 r 1 , c = q 1 s 2 r 1 . r 1 , r 2 , q 1 and q 2 are the threshold parameters needed to be set properly. If the noise level is smaller than r 1 , there is no need to increase the weight of the reference observation feature and the model will primarily utilize the actual observation features to realize sequence reconstruction. While if the noise level is larger than r 1 , the weight of the reference feature should be increased so that the negative impact of noise on reconstruction accuracy could be reduced. In addition, in order to guarantee that the fused feature always contains some feature information of the actual data, the maximum reference feature weight is set to q 2 < 1 .
Finally, the fused feature can be expressed as follows:
x 0 f u s = 1 Q x 0 o b s + Q x 0 r e f

3.4. NCDE-Based Decoder

The decoder predicts the desired regular data sequence from the fused initial latent feature x 0 f u s . Generally, most real systems are controlled, e.g., the coal-fired power plants. However, the prediction results of the NODE method depend only on the initial hidden state without considering the impact of control inputs on state changes. The NCDE proposed by Kidger et al. [28] takes the impact of control inputs into consideration during the prediction process. The solution of the NCDE is defined as follows:
x t = x 0 + t 0 t f c d x w , w ; W c d d U w   = x 0 + t 0 t f c d x w , w ; W c d d U w w d w d w
where x t r is the hidden state at time t t 0 , t N , U w : t 0 , t N c is a thrice-spline curve based on the control sequence u n r n = 0 N over the interval t 0 , t N , and f c d x t , t ; W c d : r r × c is a neural network with learnable parameters W c d .
Consider a generative model defined by the NCDE network with the initial latent feature x 0 f u s and desired sampling times t n n = 0 N . The initial latent feature x 0 f u s follows a normal distribution given by
x 0 f u s p e x 0 f u s = 𝒩 x 0 f u s | μ 0 f u s , σ 0 2 , f u s
where
μ 0 f u s = 1 Q μ 0 o b s + Q μ 0 r e f σ 0 2 , f u s = 1 Q 2 σ 0 2 , o b s + Q 2 σ 0 2 , r e f
Using Formula (12), we can compute the latent features at all desired time points. The neural network f c d consists of multiple channels of multilayer feedforward networks, with the number of channels determined by the system’s control dimension c . Taking the boiler system as an example, since it has two control inputs (coal feed and secondary air flow), f c d has two channels as follows:
h 1 , h 2 = f c d x t , t ; W c d = mlfnn x t , x t
where mlfnn denotes a multilayer feedforward network cascaded with fully connected layers and LeakyReLU activation functions, and h 1 , h 2 r × 1 are the outputs of the mlfnn for each channel. Multiplying h 1 and h 2 with their respective control derivatives and summing the results yields the latent feature of the NCDE network at the desired time t n + 1 given by
x t t = t n + 1 = h 1 d U 1 t d t t = t n + h 2 d U 2 t d t t = t n
where U 1 and U 2 are the cubic spline curves formed by the sequences of control input 1 (coal feed) and control input 2 (secondary air flow), respectively.
After obtaining the latent features at all desired time points, the output layer generates the desired observation sequence as follows:
y 0 r , , y N r = fnn x 0 , , x N
where y n r p d y n r | x n ,   n = 0 , 1 , , N . y 0 r , , y N r is the desired regular observation sequence generated from the latent feature sequence x 0 , , x N at sampling time t 0 , , t N .

3.5. Weighted Loss Function of the PRN-NCDE Model

Through reparameterization, parameters can be learned by using the gradient descent method. According to [30], the objective loss function in Equation (2) can be simplified as
𝒥 W e c , W d c | y n i r r , t n n = 0 N = n = 0 N 1 2 y n i r r y ^ n i r r 2 1 2 λ t r σ 0 2 + μ 0 T μ 0 n log σ 0 2
where y n i r r and y ^ n i r r are the encoder input and decoder output at the corresponding time, λ is a hyperparameter controlling the variance and n is the dimension of the latent feature x 0 . μ 0 and σ 0 2 are the mean and covariance of x 0 , respectively. t r denotes the trace operator.
This objective function measures the reconstruction accuracy of the decoder output y ^ n i r r relative to the encoder input y n i r r . When the actual observation sequence y n i r r , t n n = 0 N is not corrupted by noise or the noise level is low, this function enables the model to learn the data distribution effectively. However, if the actual observational data contain high levels of noise, achieving the desired training performance using this objective function is difficult.
To address this challenge, we propose a weighted loss function for the PRN-NCDE model, which is defined as follows:
𝒥 W e c , W d c | y n i r r , t n n = 0 N , y n r e f , t n n = 0 N = n = 0 N 1 2 y n i r r y ^ n i r r 1 Q 2 1 2 y n r e f y ^ n i r r Q 2           λ 2 t r σ 0 2 , f u s + μ 0 f u s T μ 0 f u s n log σ 0 2 , f u s
where Q is the reference latent feature weight calculated in Section 3.3.
Equation (18) integrates data from the reference observation sequence. It uses the reference feature weight, which is calculated based on the noise level in the actual observation data, to determine the importance of the actual and reference data distributions for parameter training. When noise is low, the loss function will rely more on the actual observation data for training. Conversely, the reference observation data will be used to mitigate the negative impact of noise on model learning.

4. Simulation Experiments and Analysis

This section selects the furnace of the boiler system as the test object to evaluate the reconstruction performance of the developed model. We use the nominal data of a 600 MW coal-fired power plant under 100% rated conditions to train the sequence reconstruction model and validate its effectiveness. Under the rated conditions, it is required that the coal flow rate is about 74.5 kg/s and the excess air coefficient is 1.2; thus, the gas oxygen content in the boiler system can be maintained around 3.2% and the boiler system can operate stably. The reference observation sequence is generated by the digital twin model of the furnace, which is defined by the following equations [31,32,33]:
x k + 1 = A x k + B u k y k + 1 = C x k + 1
where x , u , and y represent the system’s state, input, and observation vectors, respectively, and they are defined as follows:
x = V g s ρ b T g s O c p T u = W c f V s a T y = V g s T g s P b O c p T
The matrices A , B , and C are given by:
A = a 11 0 0 0 0 a 22 0 0 0 0 a 33 0 a 41 0 0 a 44           B = 0 b 12 b 21 b 22 b 31 b 32 0 0             C = 1 0 0 0 0 0 1 0 0 c 32 0 0 0 0 0 1
where
a 11 = 1 T a a 22 = 1 V g s / V b a 33 = 1 V g s V b K s l T g s 4 T s v 4 C g s ρ b V b T g s a 41 = 21 V s a V 0 W c f / V s a V b a 44 = 1 V g s / V b             b 12 = T a b 21 = 1 / V b b 22 = ρ a / V b b 31 = Q n e t , a r / C g s ρ b V b b 32 = H a ρ a / C g s ρ b V b             c 32 = R g s T g s
The physical meanings of the parameters and the value ranges of the input and output quantities are shown in Table 1. To validate the model’s performance, we compare the proposed model with the RNN-NODE method from [22] and the improved RNN-NCDE method. The network architecture parameters for each model are listed in Table 2.
The model training process uses the following settings: number of iterations—200; initial learning rate— 2 × 10 3 ; minimum learning rate— 2 × 10 4 ; decay rate—0.999; hyperparameter— λ = 0.005 . The thresholds for the reference latent feature weights are set as r 1 = 5 × 10 4 , r 2 = 5 × 10 3 , q 1 = 5 × 10 6 , and q 2 = 0.9 . The desired regular sequence length is set to N = 500 , with a sampling interval of 1 s . The control input u is normalized to the range 0 , 1 , and the observation y is normalized to the range 1 , 1 . The experimental analysis focuses on two aspects: (1) performance under different sampling numbers; (2) performance under high noise levels.
The relevant information about the parameters of the experimental environment, including CPU, GPU, memory, Matlab and deep learning toolbox version, are detailed in Table 3.

4.1. Analysis of Model Reconstruction Results Under Different Sampling Numbers

The dataset for model training and validation consists of 150 irregular observation sequences, including 5 different random sampling sequences with a sampling number of N under 30 different working conditions. This section compares the reconstruction results of models under sampling numbers of N = 60 , N = 40 , and N = 20 . The noise covariance of the actual observation sequence is set to
R = diag R 1 R 2 R 3 R 4 = diag 5 × 10 4 1 × 10 2 2 × 10 2 5 × 10 3
Figure 3 and Figure 4 show the reconstruction results of each model for the normalized values of each observation variable under different sampling numbers. Since RNN-NODE does not consider the impact of control inputs on outputs, its predictions rely solely on the initial latent state, resulting in significant estimation errors. After improvement with NCDE, the model can account for control sequences, enhancing reconstruction accuracy. As shown in Table 4, RNN-NCDE reduces the estimation error by 66.3%, 65.6%, and 53.5% compared to RNN-NODE when the sampling numbers are 60, 40, and 20, respectively. Although RNN-NCDE improves the reconstruction results of RNN-NODE, its performance degrades noticeably under sparse measurements, as evidenced by a near doubling of prediction errors when the sampling number decreases from 60 to 20.
The proposed PRN-NCDE model, by focusing on the correlation between sampling points and their position information, not only further improves fitting accuracy under each sampling number but also mitigates the impact of sparse measurements on reconstruction results. The RMSE results indicate that PRN-NCDE enhances estimation accuracy by 54%, 58.5%, and 63.7% compared to RNN-NCDE when the sampling numbers are 60, 40, and 20, respectively. Moreover, when the sampling number decreases from 60 to 20, PRN-NCDE’s prediction error increases by only 50%, remaining superior to the other two methods. Therefore, the PRN-NCDE model is more effective in handling irregular sequence reconstruction for controlled systems and can significantly reduce the impact of sparse measurements on reconstruction results.

4.2. Reconstruction Performance Analysis Under High Noise Levels

Similarly, a dataset of 150 irregular sequences with three noise levels is used to compare reconstruction performance. The sampling number for each noise level is set to N = 60 , with noise covariances as follows:
  • Noise Level 1:
R 1 = diag ( 1 × 10 3 0.05 0.1 0.02 )
  • Noise Level 2:
R 2 = diag ( 2 × 10 3 0.1 0.2 0.04 )
  • Noise Level 3:
R 3 = diag ( 3 × 10 3 0.15 0.3 0.06 )
Figure 5 and Figure 6 show the reconstruction results of each model under high noise levels. Both RNN-NCDE and RNN-NODE fail to effectively recover the true values from noisy sequences, with significant drops in accuracy.
The results in Table 5 show that RNN-NCDE has larger fitting errors than RNN-NODE under all noise levels, indicating that NCDE does not improve fitting accuracy in noisy conditions. As can be seen from the above, these two methods cannot handle the sequence reconstruction problem under large noise levels effectively.
The proposed PRN-NCDE model uses a parallel reference network to obtain reference latent features and calculates weights based on the noise level of the actual observation sequence. This weighted fusion of latent features improves reconstruction accuracy under high noise levels. The results show that PRN-NCDE can accurately recover the true value trends, with prediction errors reduced by over 70% and 60% compared to RNN-NCDE and RNN-NODE, respectively.

4.3. Discussion

For a well-trained sequence reconstruction model, the parameters related to the latent feature weight, including r 1 , r 2 , q 1 and q 2 , will decide the reconstruction results. For the simulation experiments in this paper, the value of r 1 is set based on the largest measurement noise variance. While guaranteeing the noise in a reasonable range, the features of the real observation data are more utilized for sequence reconstruction. If the noise level is smaller than r 1 , there is no need to increase the weight of the reference observation feature and the model will primarily utilize the actual observation features to realize sequence reconstruction. While if the noise level is larger than r 1 due to electromagnetic interference or sensor failure, the weight of the reference feature should be increased so that the negative impact of noise on reconstruction accuracy could be reduced. In addition, in order to guarantee that the fused feature always contains some feature information of the actual data, the maximum reference feature weight is set to q 2 < 1 .
In conclusion, for the reconstruction of noisy, irregular multivariate sequences in controlled systems, the PRN-NCDE model outperforms existing methods. Simulation results confirm its effectiveness in improving reconstruction accuracy under sparse measurements and high noise levels, making it suitable for digital twin systems in complex coal-fired power plant environments.

5. Conclusions

The uncertainty and randomness in network communication and the non-synchronous sampling of sensors can cause irregularities, sparsity, and misalignment in measurement data. Existing methods struggle to achieve ideal reconstruction under sparse measurements and high noise, making them unsuitable for sequence reconstruction tasks in digital twin systems. To address these issues, this article establishes a variational autoencoder model based on parallel reference networks and neural controlled differential equations, which can effectively handle the reconstruction of noisy, irregular multivariate sequences in digital twin systems.
Firstly, a multi-channel self-attention module is incorporated into the encoder. This module not only analyzes the position information of sampled data in the sequence and focuses on the correlation between data to enhance reconstruction accuracy under sparse measurements but also handles the misalignment and irregularity of observation sequences. Secondly, a parallel reference network is established. The reference sequence provided by the digital twin model is input and mapped to reference latent features, which are then fused with the latent features of the actual observation sequence in a weighted manner to address the reconstruction of observation sequences under high noise conditions. Thirdly, a decoder is constructed using an NCDE network, which takes into account the effect of control inputs on outputs to improve the sequence reconstruction performance of controlled systems. Finally, a weighted loss function for training the PRN- NCDE model is formed based on the calculated feature weights, which helps to better train the network parameters of the model.
Simulation results show that the proposed PRN-NCDE model improves the estimation accuracy by more than 50% and 70%, respectively, compared with RNN-NCDE under different sampling numbers and noise levels, and by more than 80% and 60%, respectively, compared with RNN-NODE. It can recover the changing trend of observation sequence more accurately. Therefore, the proposed method can effectively improve the reconstruction accuracy of observation sequences under sparse measurements and large noise conditions, and is suitable for irregular sequence reconstruction tasks in digital twin systems with uncertain network transmission and non-synchronous sensor sampling.
In future work, three aspects could be further investigated. First, the structure of the deep learning network can be improved to realize better abilities of feature extraction and prediction accuracy, especially under small sampling numbers and high noise levels. Second, the situation when outliers exist in the measurements should be considered to enhance the robustness of the reconstruction method. Third, we assume that the measurement noise follows Gaussian distribution. Non-Gaussian distribution should be considered for modelling measurement noise and the reconstruction performance under this situation could be another interesting research direction.

Author Contributions

Conceptualization, H.J. and Y.Z.; methodology, H.J., Y.Z. and Y.C.; software, H.J., Y.Z. and Q.Z.; validation, H.J., Y.Z. and Q.Z.; formal analysis, H.J. and Y.Z.; investigation, H.J. and Y.Z.; resources, H.J. and Y.Z.; data curation, H.J. and Y.Z.; writing—original draft preparation, H.J. and Y.Z.; writing—review and editing, Q.Z. and Y.C.; visualization, H.J. and Y.Z.; supervision, Y.C.; project administration, H.J.; funding acquisition, H.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant 62203349, and in part by the China Scholarship Council under Grant 201806280305.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Raw data were generated using the Matlab simulator. The data presented in this study are available on request from the corresponding author.

Acknowledgments

The authors would like to thank the anonymous reviewers for their valuable comments and suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Mikołajewska, E.; Mikołajewski, D.; Mikołajczyk, T.; Paczkowski, T. Generative AI in AI-Based Digital Twins for Fault Diagnosis for Predictive Maintenance in Industry 4.0/5.0. Appl. Sci. 2025, 15, 3166. [Google Scholar] [CrossRef]
  2. Juarez, M.; Botti, V.; Giret, A. Digital Twins: Review and Challenges. J. Comput. Inf. Sci. Eng. 2021, 21, 030802. [Google Scholar] [CrossRef]
  3. Vio, R.; Strohmer, T.; Wamsteker, W. On the reconstruction of irregularly sampled time series. Publ. Astron. Soc. Pac. 2000, 112, 74–90. [Google Scholar] [CrossRef]
  4. Kreindler, D.; Lumsden, C. The effects of the irregular sample and missing data in time series analysis. In Nonlinear Dynamical Systems Analysis for the Behavioral Sciences Using Real Data; CRC Press: Boca Raton, FL, USA, 2010; pp. 149–172. [Google Scholar]
  5. Gao, J.; Ji, W.; Zhang, L.; Shao, S.; Wang, Y.; Shi, F. Fast piecewise polynomial fitting of time-series data for streaming computing. IEEE Access 2020, 8, 43764–43775. [Google Scholar] [CrossRef]
  6. Jones, R.H. Maximum likelihood fitting of ARMA models to time series with missing observations. Technometrics 1980, 22, 389–395. [Google Scholar] [CrossRef]
  7. Fan, J.; Gijbels, I. Adaptive order polynomial fitting: Bandwidth robustification and bias reduction. J. Comput. Graph. Stat. 1995, 4, 213–227. [Google Scholar] [CrossRef]
  8. Tong, Y.; Yu, L.; Li, S.; Liu, J.; Qin, H.; Li, W. Polynomial fitting algorithm based on neural network. ASP Trans. Pattern Recognit. Intell. Syst. 2021, 1, 32–39. [Google Scholar] [CrossRef]
  9. Sekiya, F.; Sugimoto, A. Fitting discrete polynomial curve and surface to noisy data. Ann. Math. Artif. Intell. 2015, 75, 135–162. [Google Scholar] [CrossRef]
  10. Kleinbaum, D.G.; Dietz, K.; Gail, M.; Klein, M.; Klein, M. Logistic Regression; Springer: New York, NY, USA, 2010; pp. 103–127. [Google Scholar]
  11. Marlin, B.M.; Kale, D.C.; Khemani, R.G.; Wetzel, R.C. Unsupervised Pattern Discovery in Electronic Health Care Data Using Probabilistic Clustering Models. In Proceedings of the 2nd ACM SIGHIT International Health Informatics Symposium, Miami, FL, USA, 28–30 January 2012; pp. 389–398. [Google Scholar]
  12. Shukla, S.; Marlin, B. A survey on principles, models and methods for learning from irregularly sampled time series. arXiv 2020, arXiv:2012.00168. [Google Scholar]
  13. Yoon, J.; Zame, W.R.; van der Schaar, M. Estimating Missing Data in Temporal Data Streams Using Multi-Directional Recurrent Neural Networks. IEEE Trans. Biomed. Eng. 2019, 66, 1477–1490. [Google Scholar] [CrossRef] [PubMed]
  14. Shukla, S.; Marlin, B. Interpolation-prediction networks for irregularly sampled time series. arXiv 2019, arXiv:1909.07782. [Google Scholar]
  15. Tan, Q.; Ye, M.; Yang, B.; Liu, S.Q.; Ma, A.J.; Yip, T.C.F.; Wong, G.L.H.; Yuen, P.C. DATA-GRU: Dual-Attention Time-Aware Gated Recurrent Unit for Irregular Multivariate Time Series. In Proceedings of the 36th AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 930–937. [Google Scholar]
  16. Zhang, Y.; Yang, X.; Ivy, J.; Chi, M. ATTAIN: Attention-based time-aware LSTM networks for disease progression modeling. In Proceedings of the 28th International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019; pp. 4369–4375. [Google Scholar]
  17. Shukla, S.; Marlin, B. Multi-time attention networks for irregularly sampled time series. arXiv 2021, arXiv:2101.10318. [Google Scholar]
  18. Donahue, J.; Krähenbühl, P.; Darrell, T. Adversarial feature learning. arXiv 2016, arXiv:1605.09782. [Google Scholar]
  19. Cao, W.; Wang, D.; Li, J.; Zhou, H.; Li, L.; Li, Y. BRITS: Bidirectional Recurrent Imputation for Time Series. In Proceedings of the International Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018; pp. 6775–6785. [Google Scholar]
  20. Rajkomar, A.; Oren, E.; Chen, K.; Dai, A.M.; Hajaj, N.; Hardt, M.; Liu, P.J.; Liu, X.; Marcus, J.; Sun, M.; et al. Scalable and accurate deep learning with electronic health records. NPJ Digit. Med. 2018, 1, 18. [Google Scholar] [CrossRef] [PubMed]
  21. Mozer, M.; Kazakov, D.; Lindsey, R. Discrete event, continuous time RNNs. arXiv 2017, arXiv:1710.04110. [Google Scholar]
  22. Chen, R.T.Q.; Rubanova, Y.; Bettencourt, J.; Duvenaud, D.K. Neural ordinary differential equations. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, Montreal, QC, Canada, 3–8 December 2018; pp. 6572–6583. [Google Scholar]
  23. Rubanova, Y.; Chen, R.T.Q.; Duvenaud, D. Latent ordinary differential equations for irregularly-sampled time series. In Proceedings of the 33rd Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  24. De Brouwer, E.; Simm, J.; Arany, A.; Moreau, Y. GRU-ODE-Bayes: Continuous modeling of sporadically-observed time series. In Proceedings of the 33rd International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  25. Huang, Z.; Sun, Y.; Wang, W. Learning continuous system dynamics from irregularly-sampled partial observations. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020; Volume 33, pp. 16177–16187. [Google Scholar]
  26. Kidger, P.; Morrill, J.; Foster, J.; Lyons, T. Neural controlled differential equations for irregular time series. In Proceedings of the 34th Conference on Neural Information Processing Systems (NeurIPS 2020), Vancouver, BC, Canada, 6–12 December 2020; Volume 33, pp. 6696–6707. [Google Scholar]
  27. Doersch, C. Tutorial on variational autoencoders. arXiv 2016, arXiv:1606.05908. [Google Scholar]
  28. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  29. Morrill, J.; Kidger, P.; Yang, L.; Lyons, T. Neural controlled differential equations for online prediction tasks. arXiv 2021, arXiv:2106.11028. [Google Scholar]
  30. Kingma, D.; Welling, M. Auto-encoding variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  31. Lv, C. System Simulation and Modelling of Large Thermal Power Unit; Tsinghua University Press: Beijing, China, 2002. [Google Scholar]
  32. Zhao, Y.; Cai, Y.; Jiang, H. Recurrent neural network-based hybrid modeling method for digital twin of boiler system in coal-fired power plant. Appl. Sci. 2023, 13, 4905. [Google Scholar] [CrossRef]
  33. Zhao, Y.; Cai, Y.; Jiang, H.; Deng, Y. A generalized data assimilation architecture of digital twin for complex process industrial systems. Meas. Sci. Technol. 2024, 35, 066003. [Google Scholar] [CrossRef]
Figure 1. Schematic diagram of non-aligned irregularly sampled sequence reconstruction.
Figure 1. Schematic diagram of non-aligned irregularly sampled sequence reconstruction.
Applsci 15 04706 g001
Figure 2. PRN-NCDE-based sequence reconstruction network framework.
Figure 2. PRN-NCDE-based sequence reconstruction network framework.
Applsci 15 04706 g002
Figure 3. Reconstruction results of gas density and gas temperature under different sampling numbers.
Figure 3. Reconstruction results of gas density and gas temperature under different sampling numbers.
Applsci 15 04706 g003
Figure 4. Reconstruction results of furnace pressure and gas oxygen content under different sampling numbers.
Figure 4. Reconstruction results of furnace pressure and gas oxygen content under different sampling numbers.
Applsci 15 04706 g004aApplsci 15 04706 g004b
Figure 5. Reconstruction results of gas density and gas temperature under high noise levels.
Figure 5. Reconstruction results of gas density and gas temperature under high noise levels.
Applsci 15 04706 g005
Figure 6. Reconstruction results of furnace pressure and gas oxygen content under high noise levels.
Figure 6. Reconstruction results of furnace pressure and gas oxygen content under high noise levels.
Applsci 15 04706 g006aApplsci 15 04706 g006b
Table 1. Physical meaning of furnace parameters.
Table 1. Physical meaning of furnace parameters.
ParameterUnitPhysical MeaningValue
Input and Output Parameters W c f k g / s Coal flow rate[54, 80]
V s a m 3 / s Secondary air flow rate[400, 600]
V g s m 3 / s Gas flow rate[400, 650]
ρ b kg / m 3 Gas density[1.3, 1.4]
T g s ° C Gas temperature in furnace[1000, 1100]
P b kpa Furnace pressure[101, 102]
O c p % Gas oxygen content[2.6, 3.8]
Constant Parameters T a /Inertial time constant0.9
C g s kJ / kg ° C Specific heat capacity of gas2.03
H a kJ / kg Air enthalpy934.1
V b m 3 Furnace volume16,855
V 0 m 3 / kg Theoretical air consumption5.834
T s v ° C Temperature of water wall326.66
ρ a kg / m 3 Air density1.273
R g s /Universal gas constant 6.52 × 10 2
Table 2. Network architecture parameters of each reconstruction model.
Table 2. Network architecture parameters of each reconstruction model.
Sequence Reconstruction ModelNetwork Architecture Parameters
PRN-NCDE(1) MCSA module: number of channels d = 4 ; number of attention heads M = 4 ; fill value τ = 1 × 10 7 ;
(2) Number of LSTM network neurons: 80;
(3) Number of encoder fully connected layer neurons: 100;
(4) Latent feature dimension n = 10;
(5) NCDE module (two-channel feedforward network): [100]-[100]-[100]-[10].
RNN-NCDE(1) Number of LSTM network neurons: 80;
(2) Number of encoder fully connected layer neurons: 100;
(3) Latent feature dimension n = 10;
(4) NCDE module (two-channel feedforward network): [100]-[100]-[100]-[10].
RNN-NODE(1) Number of LSTM network neurons: 80;
(2) Number of encoder fully connected layer neurons: 100;
(3) Latent feature dimension n = 10;
(4) NCDE module:100-100-100-10.
Table 3. Experimental environment configuration.
Table 3. Experimental environment configuration.
Hardware and SoftwareDetails
CPUIntel(R) Core(TM) i9-10900 CPU @ 2.8 GHz
GPUNVIDIA GeForce GTX 1660 SUPER
Random Access Memory32 GB
Matlab Version2024a
Deep Learning Toolbox24.1
Table 4. RMSE results of different models under different sampling numbers.
Table 4. RMSE results of different models under different sampling numbers.
Sequence Reconstruction Model RMSE N = 60 RMSE N = 40 RMSE N = 20
PRN-NCDE0.02130.02950.0322
RNN-NCDE0.04630.07110.0888
RNN-NODE0.13750.20690.1910
Table 5. RMSE results of different models under high noise levels.
Table 5. RMSE results of different models under high noise levels.
Sequence Reconstruction Model RMSE R 1 RMSE R 2 RMSE R 3
PRN-NCDE0.01580.01550.0182
RNN-NCDE0.05990.05810.0771
RNN-NODE0.04710.04790.0529
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Jiang, H.; Zhao, Y.; Zhu, Q.; Cai, Y. A Novel Reconstruction Method for Irregularly Sampled Observation Sequences for Digital Twin. Appl. Sci. 2025, 15, 4706. https://doi.org/10.3390/app15094706

AMA Style

Jiang H, Zhao Y, Zhu Q, Cai Y. A Novel Reconstruction Method for Irregularly Sampled Observation Sequences for Digital Twin. Applied Sciences. 2025; 15(9):4706. https://doi.org/10.3390/app15094706

Chicago/Turabian Style

Jiang, Haonan, Yanbo Zhao, Qiao Zhu, and Yuanli Cai. 2025. "A Novel Reconstruction Method for Irregularly Sampled Observation Sequences for Digital Twin" Applied Sciences 15, no. 9: 4706. https://doi.org/10.3390/app15094706

APA Style

Jiang, H., Zhao, Y., Zhu, Q., & Cai, Y. (2025). A Novel Reconstruction Method for Irregularly Sampled Observation Sequences for Digital Twin. Applied Sciences, 15(9), 4706. https://doi.org/10.3390/app15094706

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop