Next Article in Journal
A Deep Reinforcement Learning Approach to Optimal Morphologies Generation in Reconfigurable Tiling Robots
Previous Article in Journal
Mixture Representations for Generalized Burr, Snedecor–Fisher and Generalized Student Distributions with Related Results
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

ASIDS: A Robust Data Synthesis Method for Generating Optimal Synthetic Samples

School of Statistics and Data Science, Nanjing Audit University, Nanjing 211815, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Mathematics 2023, 11(18), 3891; https://doi.org/10.3390/math11183891
Submission received: 16 August 2023 / Revised: 9 September 2023 / Accepted: 11 September 2023 / Published: 13 September 2023
(This article belongs to the Special Issue Advances in Computational Statistics and Data Analysis)

Abstract

:
Most existing data synthesis methods are designed to tackle problems with dataset imbalance, data anonymization, and an insufficient sample size. There is a lack of effective synthesis methods in cases where the actual datasets have a limited number of data points but a large number of features and unknown noise. Thus, in this paper we propose a data synthesis method named Adaptive Subspace Interpolation for Data Synthesis (ASIDS). The idea is to divide the original data feature space into several subspaces with an equal number of data points, and then perform interpolation on the data points in the adjacent subspaces. This method can adaptively adjust the sample size of the synthetic dataset that contains unknown noise, and the generated sample data typically contain minimal errors. Moreover, it adjusts the feature composition of the data points, which can significantly reduce the proportion of the data points with large fitting errors. Furthermore, the hyperparameters of this method have an intuitive interpretation and usually require little calibration. Analysis results obtained using simulated original data and benchmark original datasets demonstrate that ASIDS is a robust and stable method for data synthesis.

1. Introduction

Synthetic data present an effective solution to the challenges of inadequate or low-quality samples, particularly in the era of big data. The use of data synthesis methods to generate synthetic data provides a cost-effective and efficient alternative to collecting and labeling large amounts of real-world data. Furthermore, these methods can address privacy concerns associated with real-world data, making it safer to share and analyze [1]. Recently, there has been a surge in the use of synthetic data in machine learning, with various synthetic methods being developed [2,3].
Representative data synthesis methods can be classified into three categories. The first category entails techniques such as interpolation, extrapolation, and other methods [4] that generate additional data points representative of the underlying distribution. These data synthesis methods can extract more information from the dataset, thereby enhancing model generalization. For image data, deep-learning-based methods such as VAE and GAN can also be used to generate new data [5,6]. These methods are useful for improving the quality of the dataset and model generalization. The second category is meant to address the issue of dataset imbalance, such as with SMOTE [7] and some of its enhanced versions [8,9,10]. These methods can synthesize minority class samples to balance the dataset and improve the model’s performance. Finally, data sharing and research require sensitive data privacy protection. Synthetic data can help protect sensitive information while still enabling data sharing and research. Normally, we can add random noise to protect original data, like in differential privacy methods [11], generating synthetic data for sharing and research purposes. Overall, these methods help optimize the representation and use of data in various applications.
For some actual tasks, original data often contain a multitude of features and unknown noise [12]. Since the synthetic method involves generating new data based on existing data, the quality of the new points depends on the quality and quantity of the original data. If the quality of the original data is poor or the quantity is insufficient, then the synthetic data may have limitations in terms of quality [13]. However, there is a lack of effective synthetic methods for datasets with a restricted size and complex noise that can expand the size of the dataset.
Based on the purpose of improving the quality and quantity of the dataset, and motivated by piecewise linear interpolation and spline interpolation, we propose a robust and stable data synthesis method named Adaptive Subspace Interpolation for Data Synthesis (ASIDS), which aims to adaptively adjust the sample size and structure of the original dataset containing unknown noise. The idea is to divide the original feature space into several subspaces with an equal number of samples, and then perform linear interpolation for the samples in the adjacent subspaces. This method achieves sample optimization via two aspects. First, it can adaptively adjust the size of the dataset, and the expanded data typically contain minimal errors. Second, it adjusts the structure of the samples, which can significantly reduce the proportion of samples with large errors, thereby minimizing the impact of the noise in model generalization. Compared with other methods, ASIDS is particularly suitable for data containing unknown noise. Its main purpose is to expand the sample size and optimize the sample structure to uncover more hidden information in the data. Also, the samples synthesized by ASIDS tend to have smaller errors.
The rest of this paper is organized as follows: The existing interpolation research is reviewed in Section 2. Section 3 details the concept of the proposed ASIDS method and provides proof of the effectiveness of this method. The experimental results are presented and analyzed in Section 4. Finally, we conclude this paper in Section 5.

2. Related Work

Traditional interpolation methods are based on the function values of known data points for extrapolation and prediction. For a given dataset, there are generally two cases. In the case of two known data points and interpolation in between, the interpolation method can be selected based on distance, such as in nearest neighbor interpolation [14]. If it is assumed that the unknown point between these two data points is consistent with the straight line between them, a linear function, such as linear interpolation or piecewise linear interpolation [15,16], can be used to fit and interpolate. For interpolation among multiple given points, a commonly used method is to predict the value of the unknown point by constructing a high-degree polynomial based on the given points, such as is performed with Lagrange interpolation or Newton interpolation [17,18]. However, with an increase in the number of data points and the degree of the polynomial, there is a risk of overfitting and numerical instability (Runge’s phenomenon) [19]. Another solution is to construct a global smooth function by fitting a low-degree polynomial in a local region. In comparison to high-degree polynomial interpolation methods, this method has better smoothness and numerical stability, such as is seen in spline interpolation [20]. Nevertheless, as it is necessary to fit multiple local low-degree polynomials, this method can lead to relatively high computational complexity.
When expanding the sample size using interpolation methods, the selection of node positions and quantities has a significant impact on the accuracy and stability of the interpolation results. Typically, equidistant nodes [21] are used for interpolation position selection, which means the nodes are equally spaced within the interpolation interval. Chebyshev nodes [22] are selected within the interpolation interval to satisfy certain conditions for better fitting of the function. In addition, the choice of the number of interpolations can lead to instability of the data and insufficient validation accuracy. One way to determine the appropriate number of interpolations is by comparing the fitting degree of the model trained under different interpolation numbers with the original data [23], but this may lead to overfitting and the model having a poor generalization ability.

3. Proposed Method

In this section, we will explain the proposed ASIDS method in detail. ASIDS consists of two algorithms: K-Space and K-Match. To clearly illustrate the proposed method, we will first provide an overview of ASIDS and then introduce the specific algorithms involved.

3.1. Overview

ASIDS is mainly based on linear interpolation to increase the size and improve the quality of the original dataset. The idea is to divide the original feature space into several subspaces with an equal number of samples, and then perform linear interpolation for the samples in the adjacent subspaces. This method requires two hyperparameters (k and η ) in advance. Parameter k is the number of samples existing in each feature subspace, while η is the number of equidistant nodes interpolated per unit distance in the linear interpolation of the samples. This proposed method is illustrated in Figure 1.
The dataset D = { x i , y i } i = 1 n is given and is assumed to be contaminated with unknown noise, where x i X = R p and y i Y = R . Assuming y ˙ i = f ( x ˙ i ) , where x ˙ i is the actual value of x i , y ˙ i is the actual value of y i , and  f ( · ) is a continuous function, then it can be taken to represent the real relationship between x i and y. Consider the model:
y ˙ i + ϵ i , y = f ( x ˙ i + ϵ i , x ) + ϵ i ,
where ϵ i , y is the noise in y ˙ i , ϵ i , x is the noise in x ˙ i , and  ϵ i represents the error term. Expression (1) can be rewritten as
y i = f ( x i ) + ϵ i .
Let x 0 = { x 0 1 , , x 0 p } , where x 0 j = i n f { x i j } i = 1 n for j = 1 , , p , and call x 0 the sample minimum point.
Given the hyperparameter k, we provide an unsupervised clustering method called K-Space. As is shown in Figure 1b, the space can be partitioned into n / k subspaces, each containing k samples; i.e.,  X = s = 1 n / k X s , X i X j = , i , j = 1 , 2 , , n k , and  i j . The datasets corresponding to different subspaces are defined with D = s = 1 n / k D s , D s = { ( x i s , y i s ) } i = 1 k , and  x i s X s . For two adjacent subspaces, since f ( · ) is a continuous function, we assume that it can be approximated as a linear function g ( · ) , and then Equation (2) can be transformed into
y i = g ( x i ) + ϵ i + ϵ i ,
where ϵ i is the linear fitting error term. When the distance between two adjacent subspaces approaches zero and the measurements of the subspaces tend to be zero, then ϵ i 0 is obtained. Next, we will perform sample interpolation between adjacent subspaces.
We need to calculate the centers of each cluster, as follows:
x ¯ s = 1 k x i s D s x i s .
To achieve ϵ i 0 , we need to ensure that the interpolation is performed between clusters that are as close in distance as possible. For  { D s } s = 1 n / k , we define D ( 1 ) , whose cluster center has the minimum distance to the sample minimum point x 0 , and define D ( d ) , whose cluster center has the minimum distance to the center of D ( d 1 ) , D ( d 1 ) D ( 1 ) , , D ( d 1 ) , and  d > 1 .
D ( 1 ) = a r g m i n D s D d i s t ( x 0 , x ¯ s ) ,
D ( d ) = a r g m i n { D s D } , D s D ( 1 ) , , D ( d 1 ) d i s t ( x ¯ 0 ( d 1 ) , x ¯ s ) ,
where x ¯ ( d 1 ) is the center of D ( d 1 ) . Interpolation is performed on { D ( d ) } d = 1 n / k sequentially according to the order of the d values, and interpolation is carried out only between adjacent subspaces (i.e., interpolate between D ( 1 ) and D ( 2 ) , between  D ( 2 ) and D ( 3 ) , and so on).
When performing linear interpolation between adjacent subspaces, we should pair the k samples from the first subspace with an equal number of samples from the second subspace. The interpolation rules between adjacent subspaces are as follows:
  • Linear interpolation can only be performed between two samples belonging to different adjacent subspace sets.
  • Interpolation must be performed for each sample.
  • Participation of each sample point is restricted to a single interpolation instance.
The number of matching schemes is k ! . As is shown in Figure 1c, we provide a matching method called K-Match. Supposing ϵ i 0 , then this method can select a good-performing matching scheme of { ( x i ( d ) , y i ( d ) ) , ( x i ( d + 1 ) , y i ( d + 1 ) ) } i = 1 k from k ! .
Assuming x and y are continuous variables, and given another hyperparameter η , then the number of samples inserted using the linear interpolation method between D ( d ) and D ( d + 1 ) is i = 1 k η · d i s t ( x i ( d ) , x i ( d + 1 ) ) . Taking ( x i ( d ) , y i ( d ) ) D ( d ) and ( x i ( d + 1 ) , y i ( d + 1 ) ) D ( d + 1 ) as an example, then { ( x ( d , d + 1 ) ( m , i ) , y ( d , d + 1 ) ( m , i ) ) } m = 1 η · d i s t ( x i ( d ) , x i ( d + 1 ) ) is the set of inserted samples, and the linear interpolation formula is defined as
x ( d , d + 1 ) ( m , i ) = x i ( d ) + m · x i ( d + 1 ) x i ( d ) η · d i s t ( x i ( d ) , x i ( d + 1 ) ) + 1 ,
y ( d , d + 1 ) ( m , i ) = y i ( d ) + m · y i ( d + 1 ) y i ( d ) η · d i s t ( y i ( d ) , y i ( d + 1 ) ) + 1 .
After ASIDS processing, the original dataset will be optimized. The main steps of the ASIDS algorithm are summarized in Algorithm 1.
Algorithm 1: ASIDS.
Mathematics 11 03891 i001
The assumptions of ASIDS are as follows:
  • f ( · ) is a continuous function.
  • The linear fitting error is ϵ i 0 .
  • x and y are continuous variables.

3.2. K-Space

The implementation of ASIDS requires an unsupervised clustering method to partition the feature space into multiple subspaces, each containing k samples. Based on this, we propose the K-Space clustering method. The clustering method has the following performance parameters:
  • Each subspace contains an equal number of samples, i.e.,  D = s = 1 n / k D s , D s = { ( x i s , y i s ) } i = 1 k ;
  • Each sample belongs to only one subset, i.e.,  D i D j = , i j .
Maintaining continuity and similarity between adjacent subspaces is essential for synthesizing data via multiple linear interpolations in ASIDS. Our objective is to minimize the linear fitting error ϵ i , which helps to satisfy ASIDS assumption 2 as much as possible.
To determine the sample set D s for subspace X s , it is necessary to determine the first sample x 1 s in D s .
x 1 s = a r g m i n x : x D , x D 1 , , D s 1 d i s t ( x , x ¯ s 1 ) ,
where s = 1 , , n k , x ¯ s 1 is the cluster center of D s 1 , and  x ¯ 0 = x 0 . We define D s = { x 1 s } and determine x d s as follows:
x d s = a r g m i n x : x D , x D 1 , , D s d i s t ( x , x ¯ s ) ,
where d = 2 , , k . x d s is obtained and D s D s { x d s } is updated.
The main steps of the K-Space algorithm are summarized in Algorithm 2.
Algorithm 2: K-Space.
Mathematics 11 03891 i002

3.3. K-Match

We can calculate the total error of the matching scheme to measure the quality of the scheme; for the sake of simplicity, let X = R , and we calculate it as follows:
i = 1 k S ( x i ( d ) , x i ( d + 1 ) ) = i = 1 k x i ( d ) x i ( d + 1 ) | f ( x ) L i ( x ) | d x ,
where L i ( x ) is the linear expression passing through the points ( x i ( d ) , y i ( d ) ) and ( x i ( d + 1 ) , y i ( d + 1 ) ) .
Theorem 1.
Let X ( d ) and X ( d + 1 ) be two adjacent subspaces; the datasets corresponding to different subspaces are D ( d ) , D ( d + 1 ) , and  ( x i ( d ) , y i ( d ) ) D ( d ) , ( x i ( d + 1 ) , y i ( d + 1 ) ) D ( d + 1 ) . Consider the model y i = f ( x i ) + ϵ i , and let ϵ i ( d ) = y i ( d ) f ( x i ( d ) ) . For  i = 1 , 2 , , k , suppose that ϵ i 0 ; then, E S ( x i ( d ) , x i ( d + 1 ) ) | x i ( d ) x i ( d + 1 ) | < E | ϵ i ( d ) | + | ϵ i ( d + 1 ) | 2 .
Proof of Theorem 1.
Since ϵ i 0 , according to Equation (3) the model can be transformed into
y i = g ( x i ) + ϵ i ,
where g ( · ) is a linear function. According to Equation (11), it follows that
S ( x i ( d ) , x i ( d + 1 ) ) = x i ( d ) x i ( d + 1 ) | g ( x ) L i ( x ) | d x .
When ϵ i ( d ) · ϵ i ( d + 1 ) < 0 , let ( x , y ) be the intersection point between y = L i ( x ) and y = g ( x ) . We can simplify S ( x i ( d ) , x i ( d + 1 ) ) using basic geometric area calculations, and according to the law of iterated expectations (LIE),   
E S ( x i ( d ) , x i ( d + 1 ) ) | x i ( d ) x i ( d + 1 ) | = E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) · ϵ i ( d + 1 ) 0 P ( ϵ i ( d ) · ϵ i ( d + 1 ) 0 ) | x i ( d ) x i ( d + 1 ) | + E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) · ϵ i ( d + 1 ) < 0 P ( ϵ i ( d ) · ϵ i ( d + 1 ) < 0 ) | x i ( d ) x i ( d + 1 ) | = E | ϵ i ( d ) | + | ϵ i ( d + 1 ) | P ( ϵ i ( d ) · ϵ i ( d + 1 ) 0 ) 2 + h 1 · E | ϵ i ( d ) | + h 2 · E | ϵ i ( d + 1 ) | P ( ϵ i ( d ) · ϵ i ( d + 1 ) < 0 ) 2 ,
where h 1 = | x x i ( d ) | | x i ( d ) x i ( d + 1 ) | and h 2 = | x x i ( d + 1 ) | | x i ( d ) x i ( d + 1 ) | . Since P ( ϵ i ( d ) · ϵ i ( d + 1 ) 0 ) + P ( ϵ i ( d ) · ϵ i ( d + 1 ) < 0 ) = 1 , and  h 1 + h 2 = 1 , it follows that E S ( x i ( d ) , x i ( d + 1 ) ) | x i ( d ) x i ( d + 1 ) | < E | ϵ i ( d ) | + | ϵ i ( d + 1 ) | 2 . □
If our approach is to randomly select a matching scheme, the validity of this method can be proved by Theorem 1. However, randomly selecting a matching scheme does not guarantee the uniqueness of the results, and it also does not guarantee that we will necessarily select a good-performing matching scheme. We found that for x i ( d ) and x i ( d + 1 ) , if  ϵ i ( d ) · ϵ i ( d + 1 ) < 0 , there will be a better interpolation effect.
Theorem 2.
Let y i ( d ) = f ( x i ( d ) ) + ϵ i ( d ) , y i ( d + 1 ) = f ( x i ( d + 1 ) ) + ϵ i ( d + 1 ) . Suppose that ϵ i 0 ; then, we can obtain E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) · ϵ i ( d + 1 ) < 0 < E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) · ϵ i ( d + 1 ) 0 .
Proof of Theorem 2.
Since ϵ i 0 , as based on the proof of Theorem 1, we can obtain
E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) ϵ i ( d + 1 ) 0 = | x i ( d ) x i ( d + 1 ) | · E | ϵ i ( d ) | + | x i ( d ) x i ( d + 1 ) | · E | ϵ i ( d + 1 ) | 2 ,
E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) ϵ i ( d + 1 ) < 0 = | x x i ( d ) | · E | ϵ i ( d ) | + | x x i ( d + 1 ) | · E | ϵ i ( d + 1 ) | 2 .
It follows that
E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) ϵ i ( d + 1 ) < 0 < E S ( x i ( d ) , x i ( d + 1 ) ) | ϵ i ( d ) ϵ i ( d + 1 ) 0 .
According to Theorem 2, we can match samples with the opposite signs of ϵ i to achieve a good data synthesis effect. Therefore, the core idea of K-Match is to judge whether the sign of ϵ i is positive or negative for each sample, and then interpolate the samples with opposite signs as much as possible.
In K-Match, we need to choose an appropriate linear regression method to fit the dataset D ( d ) D ( d + 1 ) based on the performance of the noise. For example, lasso regression, locally weighted linear regression (LWLR) [24], and other methods can be used [25,26]. In our experiments, we used the OLS or SVR method to fit and obtain g ^ ( · ) . Notably, the kernel function is linear in SVR. According to Equation (3), and supposing that the linear fitting error is ϵ i 0 for dataset D ( d ) D ( d + 1 ) , we can obtain
ϵ i = y i g ^ ( x i ) .
Then, we sort the samples in dataset D ( d ) in ascending order according to the value of ϵ i , and obtain { ( x i ( d ) , y i ( d + 1 ) ) } i = 1 k , whereas we sort the samples in dataset D ( d + 1 ) in descending order and obtain { ( x i ( d + 1 ) , y i ( d + 1 ) ) } i = 1 k . As is shown in Figure 1d, the sorted datasets D ( d ) and D ( d + 1 ) are combined into the matching scheme { ( x i ( d ) , y i ( d ) ) , ( x i ( d + 1 ) , y i ( d + 1 ) ) } i = 1 k , the main steps of the K-Space algorithm are summarized in Algorithm 3.
Algorithm 3: K-Match.
   Input: Subset D ( d ) = { ( x i ( d ) , y i ( d ) ) } i = 1 k , D ( d + 1 ) = { ( x i ( d + 1 ) , y i ( d + 1 ) } i = 1 k
   Output: Matching scheme { ( x i ( d ) , y i ( d ) ) , ( x i ( d + 1 ) , y i ( d + 1 ) ) } i = 1 k
1  Fit the dataset D ( d ) D ( d + 1 ) and obtain g ^ ( · )
2  Obtain { ϵ i } using Equation (10)
3  Sort the samples in D ( d ) and D ( d + 1 ) according to the value of ϵ i , and obtain D ( d ) and D ( d + 1 )
4  Combine D ( d ) and D ( d + 1 ) into { ( x i ( d ) , y i ( d ) ) , ( x i ( d + 1 ) , y i ( d + 1 ) ) } i = 1 k

3.4. Supplementary Notes

The proposed method can effectively expand the size of the dataset and adjust the dataset structure, reducing the proportion of samples that deviate significantly from the actual distribution, and thereby improve model generalization performance (see Figure 2).
The supplements to ASIDS are given below:
  • The choice of the hyperparameter k is crucial, as different datasets require different values of k. Conversely, the hyperparameter η tends to exhibit a better performance as its value increases, which will be illustrated in the experimental results in the following section.
  • It is necessary to normalize the data if there is a significant difference in the dimensional scale between the features of the data. This avoids the issue of generating an excessive number of samples.
  • In most cases, n / k is not an integer, and for the excess samples, we usually have two solutions for handling this. The first one is to use the LOF algorithm [27] to filter out the excess samples that are not needed for ASIDS, as is shown in Figure 2c. Another solution is to treat the excess samples as a dataset D of a subspace D = { ( x i , y i ) } i = 1 n m o d k . When interpolating between D and other subspaces D , we choose an appropriate linear regression method to fit the dataset D D and obtain g ^ ( · ) . Then, we use the same method to sort D and D . Only n m o d k interpolations are performed, with each sample in D being interpolated, while for D , only n m o d k samples are interpolated. Moreover, we interpolate the samples with the opposite signs of ϵ i as much as possible, as is shown in Figure 3.

4. Verification of the Performances of ASIDS

For the artificial data, which are also known as f ( · ) , we investigated the hyperparameter selection and optimization performance after processing with ASIDS. For the benchmark data, we studied the prediction performance using this method.

4.1. Simulated Datasets

Some verification indicators, including the proportion of samples with ϵ i greater than α and the mean square error (MSE), can be selected to test the optimization effect of ASIDS on the original sample after processing as follows:
p ( α ) = 1 n i = 1 n I ( | f ( x i ) y i |   > α ) ,
M S E = 1 n i = 1 n ( f ( x i ) y i ) 2 .
For the simulated datasets, { x i } i = 1 n is generated using N p ( 0 , E ) . Let W 1 R ( p × p 1 ) and W 2 R ( p 1 × 1 ) , where all elements of both W 1 and W 2 are independently and identically distributed as N ( 0 , 1 ) . Consider that y i = f ( x i ) + ϵ i = t a n h ( x i W 1 ) W 2 + ϵ i .
t a n h ( x ) = e x e x e x + e x .
For a given dataset, we typically cannot ascertain the distribution of noise. Hence, we constructed datasets that contain unknown noises [12]. The unknown noises are simulated by a mixture of noises that contain uniform noises and Gaussian noises. The generation of ϵ i is explained in [28]. We fixed the random effects of the generated datasets; the experiments were repeated 100 times for each simulated dataset. Table 1 shows six simulated datasets. Specifically, P is the number of features in each sample of the simulated datasets. P 1 is the dimension of the intermediate layer used to generate the y values.

4.1.1. Hyperparameter Selection

First, we investigated the determination of parameter k. To ensure that the number of feature subspaces was sufficient, we set the hyperparameter η = 1 , letting k = 1 , 2 , , n 2 . To calculate the changes in the MSE of the datasets at different values of k, we obtained k = a r g m i n k : k = 1 , 2 , , n 2 , η = 1 M S E . The change trend in the MSE index before and after the ASISO processing is shown in Figure 4. Then, by letting η = 1 , 2 , , 30 , k = k , we calculated the changes in the MSE of the datasets (Figure 5), and obtained η = a r g m i n η : η = 1 , 2 , , 30 , k = k M S E . At last, we let k = k , η = 1 , and η , and we calculated the changes in p ( α ) of the simulated datasets (Figure 6).
As can be seen from Figure 4, ASIDS demonstrates good optimization performance for datasets with varying sample sizes or feature dimensions. The experimental results also show that the performance of ASIDS does not decline dramatically as the content of noise with large variances increases. Hence, it can deal with unknown noise well and has good robustness. In addition, it was shown by the experimental results that ASIDS exhibits good optimization performance for all datasets; thus, it is also a stable data synthesis method. As can be seen from Figure 5, the hyperparameter η has a generally monotonically decreasing relationship with the MSE, and the larger the value of η , the more significant the effect. From Figure 6, it can be observed that ASIDS can adaptively adjust the sample structure, which reduces the proportion of samples with large errors.

4.1.2. Comparison of Optimization Performance

To verify the optimization performance of ASIDS, we compared it with three other methods: piecewise linear interpolation, linear extrapolation, and nearest neighbor interpolation. Specifically, we let k = k and η = η for ASIDS. Based on the given dataset, we used the samples generated by ASIDS as interpolation points to calculate the output values of linear extrapolation and nearest neighbor interpolation. Moreover, by letting k = 1 and η = η for ASIDS, we can regard it as piecewise linear interpolation. The experimental results show that ASIDS has the smallest MSE among all methods for each dataset (Figure 7).

4.2. Benchmark Datasets

To verify the prediction performance of ASIDS, four benchmark datasets were used [29,30]. We partitioned the dataset into training and testing sets using a 7:3 ratio and normalized the data using the min–max normalization method. We preprocessed the training data using ASIDS, trained multiple machine learning models on the training set, and evaluated the prediction performance of the models on the testing set. The evaluation metric we used was the mean absolute error (MAE).
M A E = 1 n i = 1 n | y i y ^ i | .
In addition, we removed features that cannot be directly used, such as “Datetime” for bike sharing demand, “Month” for forest fires, and so on. We chose the K-nearest neighbor (KNN), random forest (RF), gradient boosting decision tree (GBDT), multilayer perceptron (MLP), and support vector regression (SVR) as the machine learning prediction models. Specifically, the kernel function is the radial basis function (RBF) in SVR. Moreover, we set the number of hidden layers to 3 for the MLP and used different numbers of neurons based on the input dimensionality and sample size. Prior to and subsequent to the application of ASIDS, the MLP retains an identical network architecture. Additionally, there are negligible variations in the count of base learners within ensemble methodologies, such as RF and GBDT. On a holistic level, these variations do not impart significant changes to the model’s complexity.
The experimental results of the five models are shown in Table 2 (the parameters for both the models and algorithms across diverse datasets were optimized using grid search and cross-validation). It is evident that ASIDS performs well on each benchmark dataset. This method is highly applicable to all five models, and in most cases, it can improve the predictive performance. It is worth mentioning that the dataset contains many categorical features which are not continuous variables. Moreover, for sparse samples (Facebook Metrics and Forest Fires), it is difficult to guarantee that the linear fitting error is ϵ i 0 when interpolating between subspaces. This indicates that even if there are violations of the ASIDS assumptions in practical applications, this method may still achieve good optimization results. This further demonstrates that this method is robust and stable.

5. Conclusions

In this paper, we proposed a data synthesis method, ASIDS, which can adaptively adjust the size of the dataset, and the generated synthetic data typically contain minimal errors. Moreover, it can adjust the structure of the samples, which can significantly reduce the proportion of samples with large errors. The experimental results from the simulated datasets demonstrate that ASIDS can optimize the samples, and compared to other methods, the data generated using this method had smaller errors. ASIDS can deal with unknown noise better and has good robustness. The results from the benchmark datasets show that the proposed method is applicable for many machine models, and in most cases, it can improve the model generalization. ASIDS has some deficiencies and limitations. In practical applications, it should be considered whether the real scenario data can satisfy the assumptions of ASIDS. From the experimental results, it can be seen that the choice of hyperparameters has a great influence on the results. Future work may focus on practical applications and its integration with advanced machine learning techniques, and study of how to automatically or effectively select hyperparameters.

Author Contributions

Writing—original draft, Y.D.; Software, Y.C.; Writing—Data curation & editing, X.J.; Methodology, H.W.; review, Y.L.; review, M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by National Social Science Fund of China grant number 22BTJ021.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. ALRikabi, H.T.S.; Hazim, H.T. Enhanced data security of communication system using combined encryption and steganography. iJIM 2021, 15, 145. [Google Scholar] [CrossRef]
  2. Kollias, D. ABAW: Learning from synthetic data & multi-task learning challenges. In Computer Vision—ECCV 2022 Workshops; Springer: Cham, Switzerland, 2022; pp. 157–172. [Google Scholar]
  3. Mahesh, B. Machine learning algorithms—A review. Int. J. Sci. Res. (IJSR) 2020, 9, 381–386. [Google Scholar]
  4. Lepot, M.; Aubin, J.B.; Clemens, F.H.L.R. Interpolation in time series: An introductive overview of existing methods, their performance criteria and uncertainty assessment. Water 2017, 9, 796. [Google Scholar] [CrossRef]
  5. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef] [PubMed]
  6. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  7. Chawla, N.V.; Bowyer, K.W.; Hall, L.O.; Kegelmeyer, W.P. SMOTE: Synthetic minority over-sampling technique. J. Artif. Intell. Res. 2002, 16, 321–357. [Google Scholar] [CrossRef]
  8. Dablain, D.; Krawczyk, B.; Chawla, N.V. DeepSMOTE: Fusing deep learning and SMOTE for imbalanced data. IEEE Trans. Neural Netw. Learn. Syst. 2022. [Google Scholar] [CrossRef] [PubMed]
  9. Han, H.; Wang, W.Y.; Mao, B.H. Borderline-SMOTE: A new over-sampling method in imbalanced data sets learning. In Proceedings of the International Conference on Intelligent Computing, Hefei, China, 23–26 August 2005; Springer: Berlin/Heidelberg, Germany, 2005; pp. 878–887. [Google Scholar]
  10. Bunkhumpornpat, C.; Sinapiromsaran, K.; Lursinsap, C. Safe-level-smote: Safe-level-synthetic minority over-sampling technique for handling the class imbalanced problem. In Advances in Knowledge Discovery and Data Mining, Proceedings of the 13th Pacific-Asia Conference, PAKDD 2009 Bangkok, Thailand, 27–30 April 2009; Springer: Berlin/Heidelberg, Germany, 2009; pp. 475–482. [Google Scholar]
  11. Ha, T.; Dang, T.K.; Dang, T.T.; Truong, T.A.; Nguyen, M.T. Differential privacy in deep learning: An overview. In Proceedings of the 2019 International Conference on Advanced Computing and Applications (ACOMP), Nha Trang, Vietnam, 26–28 November 2019; pp. 97–102. [Google Scholar]
  12. Meng, D.; De La Torre, F. Robust matrix factorization with unknown noise. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, Australia, 1–8 December 2013; pp. 1337–1344. [Google Scholar]
  13. Raghunathan, T.E. Synthetic data. Annu. Rev. Stat. Its Appl. 2021, 8, 129–140. [Google Scholar] [CrossRef]
  14. Sibson, R. A brief description of natural neighbour interpolation. In Interpreting Multivariate Data; Wiley: New York, NY, USA, 1981; pp. 21–36. [Google Scholar]
  15. Tachev, G.T. Piecewise linear interpolation with nonequidistant nodes. Numer. Funct. Anal. Optim. 2000, 21, 945–953. [Google Scholar] [CrossRef]
  16. Blu, T.; Thévenaz, P.; Unser, M. Linear interpolation revitalized. IEEE Trans. Image Process. 2004, 13, 710–719. [Google Scholar] [CrossRef] [PubMed]
  17. Berrut, J.P.; Trefethen, L.N. Barycentric lagrange interpolation. SIAM Rev. 2004, 46, 501–517. [Google Scholar] [CrossRef]
  18. Musial, J.P.; Verstraete, M.M.; Gobron, N. Comparing the effectiveness of recent algorithms to fill and smooth incomplete and noisy time series. Atmos. Chem. Phys. 2011, 11, 7905–7923. [Google Scholar] [CrossRef]
  19. Fornberg, B.; Zuev, J. The Runge phenomenon and spatially variable shape parameters in RBF interpolation. Comput. Math. Appl. 2007, 54, 379–398. [Google Scholar] [CrossRef]
  20. Rabbath, C.A.; Corriveau, D. A comparison of piecewise cubic Hermite interpolating polynomials, cubic splines and piecewise linear functions for the approximation of projectile aerodynamics. Def. Technol. 2019, 15, 741–757. [Google Scholar] [CrossRef]
  21. Habermann, C.; Kindermann, F. Multidimensional spline interpolation: Theory and applications. Comput. Econ. 2007, 30, 153–169. [Google Scholar] [CrossRef]
  22. Ganzburg, M.I. The Bernstein constant and polynomial interpolation at the Chebyshev nodes. J. Approx. Theory 2002, 119, 193–213. [Google Scholar] [CrossRef]
  23. Bové, D.S.; Held, L.; Kauermann, G. Objective Bayesian Model Selection in Generalized Additive Models With Penalized Splines. J. Comput. Graph. Stat. 2015, 24, 394–415. [Google Scholar] [CrossRef]
  24. Cleveland, W.S. Robust locally weighted regression and smoothing scatterplots. J. Am. Stat. Assoc. 1979, 74, 829–836. [Google Scholar] [CrossRef]
  25. Lichti, D.D.; Chan, T.O.; Belton, D. Linear regression with an observation distribution model. J. Geod. 2021, 95, 1–14. [Google Scholar] [CrossRef]
  26. Liu, C.; Li, B.; Vorobeychik, Y.; Oprea, A. Robust linear regression against training data poisoning. In Proceedings of the 10th ACM Workshop on Artificial Intelligence and Security, Dallas, TX, USA, 3 November 2017; pp. 91–102. [Google Scholar]
  27. Breunig, M.M.; Kriegel, H.P.; Ng, R.T.; Sander, J. LOF: Identifying density-based local outliers. In Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data, Dallas, TX, USA, 16–18 May 2000; pp. 93–104. [Google Scholar]
  28. Guo, Y.; Wang, W.; Wang, X. A robust linear regression feature selection method for data sets with unknown noise. IEEE Trans. Knowl. Data Eng. 2021, 35, 31–44. [Google Scholar] [CrossRef]
  29. Cukierski, W. Bike Sharing Demand. Kaggle. 2014. Available online: https://kaggle.com/competitions/bike-sharing-demand (accessed on 25 October 2014).
  30. Dua, D.; Craff, C. UCI Machine Learning Repository. 2017. Available online: http://archive.ics.uci.edu/ml (accessed on 25 January 2017).
Figure 1. Overview of the proposed approach for ASIDS. (a) The dataset contains multiple noisy data points, and the true functional relationship between x and y is f(·). (b) In the first step, we use the K-Space algorithm to perform unsupervised clustering on the dataset and divide the original feature space into several subspaces. (c) Interpolation matching of data points between adjacent subspaces is performed using the K-Match algorithm, and data points with the same color belong to the same class to be interpolated. (d) Piecewise linear interpolation is performed on data points under different classes, where equidistant data points are inserted onto lines of different colors in adjacent subspaces.
Figure 1. Overview of the proposed approach for ASIDS. (a) The dataset contains multiple noisy data points, and the true functional relationship between x and y is f(·). (b) In the first step, we use the K-Space algorithm to perform unsupervised clustering on the dataset and divide the original feature space into several subspaces. (c) Interpolation matching of data points between adjacent subspaces is performed using the K-Match algorithm, and data points with the same color belong to the same class to be interpolated. (d) Piecewise linear interpolation is performed on data points under different classes, where equidistant data points are inserted onto lines of different colors in adjacent subspaces.
Mathematics 11 03891 g001
Figure 2. The synthetic data from ASIDS. (a) The true relationship between x and y is y = x 3 . The sample size of the original data is 200 (b) Adding Gaussian noise. (c) Let k = 6 and η = 100 ; by processing with ASIDS, the sample size of the generated synthetic data was 3808.
Figure 2. The synthetic data from ASIDS. (a) The true relationship between x and y is y = x 3 . The sample size of the original data is 200 (b) Adding Gaussian noise. (c) Let k = 6 and η = 100 ; by processing with ASIDS, the sample size of the generated synthetic data was 3808.
Mathematics 11 03891 g002
Figure 3. Interpolations in subspaces with different number of samples.
Figure 3. Interpolations in subspaces with different number of samples.
Mathematics 11 03891 g003
Figure 4. Changes in the MSE under different k values: (a) simulation study results of dataset D 1 ; (b) simulation study results of dataset D 2 ; (c) simulation study results of dataset D 3 ; (d) simulation study results of dataset D 4 ; (e) simulation study results of dataset D 5 ; and (f) simulation study results of dataset D 6 , which will not be specifically mentioned below.
Figure 4. Changes in the MSE under different k values: (a) simulation study results of dataset D 1 ; (b) simulation study results of dataset D 2 ; (c) simulation study results of dataset D 3 ; (d) simulation study results of dataset D 4 ; (e) simulation study results of dataset D 5 ; and (f) simulation study results of dataset D 6 , which will not be specifically mentioned below.
Mathematics 11 03891 g004
Figure 5. Changes in the MSE under different η values.
Figure 5. Changes in the MSE under different η values.
Mathematics 11 03891 g005
Figure 6. Changes in the p ( α ) under different η values.
Figure 6. Changes in the p ( α ) under different η values.
Mathematics 11 03891 g006
Figure 7. Comparison of MSEs of simulated datasets.
Figure 7. Comparison of MSEs of simulated datasets.
Mathematics 11 03891 g007
Table 1. Simulated datasets.
Table 1. Simulated datasets.
Simulated Datasets ϵ i DistributionSample Size( P , P 1 )
D 1 20%-N(0, 64)
30%-U(−8,8)
50%-N(0, 0.04)
500(5, 3)
D 2 20%-N(0, 64)
30%-U(−8,8)
50%-N(0, 0.04)
200(5, 3)
D 3 20%-N(0, 64)
30%-U(−8,8)
50%-N(0, 0.04)
1500(5, 3)
D 4 20%-N(0, 64)
30%-U(−8,8)
50%-N(0, 0.04)
500(1, 3)
D 5 20%-N(0, 64)
30%-U(−8,8)
50%-N(0, 0.04)
500(20, 10)
D 6 40%-N(0, 64)
45%-U(−8,8)
15%-N(0, 0.04)
500(5, 3)
Table 2. Experimental results for benchmark datasets.
Table 2. Experimental results for benchmark datasets.
DatasetsProcessingHyperparameterTesting MAE ( 10 2 )
KNNRFMLPSVRGBDT
Bike Sharing-- 2.20 0.12 0.54 4.26 0.23
ASIDS k = 150 , η = 10 2.00 0.06 0.25 4.12 0.19
Facebook-- 6.60 2.25 8.19 7.34 1.33
ASIDS k = 2 , η = 10 6.41 1.59 2.02 7.36 1.12
Air Quality-- 2.99 2.57 4.80 3.87 2.58
ASIDS k = 20 , η = 100 2.89 2.55 2.08 3.84 2.77
Forest Fires-- 4.06 4.93 10.58 7.36 4.49
ASIDS k = 10 , η = 100 3.89 4.32 4.80 10.05 4.14
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Du, Y.; Cai, Y.; Jin, X.; Wang, H.; Li, Y.; Lu, M. ASIDS: A Robust Data Synthesis Method for Generating Optimal Synthetic Samples. Mathematics 2023, 11, 3891. https://doi.org/10.3390/math11183891

AMA Style

Du Y, Cai Y, Jin X, Wang H, Li Y, Lu M. ASIDS: A Robust Data Synthesis Method for Generating Optimal Synthetic Samples. Mathematics. 2023; 11(18):3891. https://doi.org/10.3390/math11183891

Chicago/Turabian Style

Du, Yukun, Yitao Cai, Xiao Jin, Hongxia Wang, Yao Li, and Min Lu. 2023. "ASIDS: A Robust Data Synthesis Method for Generating Optimal Synthetic Samples" Mathematics 11, no. 18: 3891. https://doi.org/10.3390/math11183891

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop