Next Article in Journal
Study of a Crosstalk Suppression Scheme Based on Double-Stage Semiconductor Optical Amplifiers
Previous Article in Journal
Multi-Stage Corn-to-Syrup Process Monitoring and Yield Prediction Using Machine Learning and Statistical Methods
Previous Article in Special Issue
Robot-Assisted Augmented Reality (AR)-Guided Surgical Navigation for Periacetabular Osteotomy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Extraction of Features for Time Series Classification Using Noise Injection

by
Gyu Il Kim
1 and
Kyungyong Chung
2,*
1
Department of Computer Science, Kyonggi University, Suwon 16227, Republic of Korea
2
Division of AI Computer Science and Engineering, Kyonggi University, Suwon 16227, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(19), 6402; https://doi.org/10.3390/s24196402
Submission received: 8 August 2024 / Revised: 24 September 2024 / Accepted: 30 September 2024 / Published: 2 October 2024
(This article belongs to the Special Issue Augmented Reality-Based Navigation System for Healthcare)

Abstract

:
Time series data often display complex, time-varying patterns, which pose significant challenges for effective classification due to data variability, noise, and imbalance. Traditional time series classification techniques frequently fall short in addressing these issues, leading to reduced generalization performance. Therefore, there is a need for innovative methodologies to enhance data diversity and quality. In this paper, we introduce a method for the extraction of features for time series classification using noise injection to address these challenges. By employing noise injection techniques for data augmentation, we enhance the diversity of the training data. Utilizing digital signal processing (DSP), we extract key frequency features from time series data through sampling, quantization, and Fourier transformation. This process enhances the quality of the training data, thereby maximizing the model’s generalization performance. We demonstrate the superiority of our proposed method by comparing it with existing time series classification models. Additionally, we validate the effectiveness of our approach through various experimental results, confirming that data augmentation and DSP techniques are potent tools in time series data classification. Ultimately, this research presents a robust methodology for time series data analysis and classification, with potential applications across a broad spectrum of data analysis problems.

1. Introduction

Time series data play an important role in various fields such as finance [1], healthcare [2], and industrial processes [3], and accurate analysis of such data is essential for decision-making and performance improvement in these areas. However, due to the complexity and variability of time series data, effectively classifying them remains a challenging task. Time series data are characterized by nonlinearity and non-stationarity over time, and failure to properly capture these characteristics can lead to degraded classification performance. Existing time series classification methods include traditional statistical techniques [4], machine learning algorithms [5], and deep learning-based models [6], but they have several limitations due to data imbalance, noise, and limited data quantities [7]. For example, traditional methods struggle to capture complex patterns in data, and deep learning models require large amounts of labeled data and high computational resources. Therefore, there is a need to develop new classification methodologies that can effectively handle the complexity of time series data and apply them to real-world problems across various domains. In this paper, we propose a novel approach that combines noise injection and digital signal processing techniques to overcome these limitations. Through this method, we enhance the features of the data and demonstrate that high classification performance can be achieved even with small amounts of data.
To address these issues, data augmentation techniques and DSP methods have emerged as useful tools [8]. Data augmentation techniques enhance data diversity by applying various transformations to existing data, thereby improving the generalization performance of models. Notably, noise injection is a technique that generates new data by adding random noise to the original data, significantly increasing data diversity. This allows the model to learn from a wider range of scenarios, thereby improving classification performance [9]. DSP is a method that transforms the features of time series data, enabling more effective extraction and representation of information. Through DSP techniques such as frequency analysis, filtering, and transformation, important features of time series data can be highlighted, or noise can be removed. This enhancement enables classification models to perform better [10].
Therefore, this paper aims to enhance the classification performance of time series data by combining data augmentation through noise injection and DSP methods. This approach enables the model to learn from the diverse variability of time series data and improves classification performance by emphasizing important features. The proposed methodology consists of three main steps. First, data augmentation is performed through noise injection. Second, DSP techniques are applied to the augmented data to extract important features from the time series data. Finally, the transformed features are classified using existing time series data models. This process aims to effectively model the complex patterns and dynamic changes of time series data, achieving high classification accuracy. By providing a detailed methodology, experimental results, and insights gained from this novel approach, this paper comprehensively evaluates the impact of integrating deep learning and DSP techniques on the field of time series data analysis.
The contributions of the proposed method in this paper are as follows:
  • Novel Integration of Noise Injection and DSP for Time Series Classification: We propose a unique methodology that synergistically combines noise injection and digital signal processing techniques specifically for time series classification. This integrated approach is novel, as previous works have explored these techniques individually but not in conjunction.
  • Enhanced Feature Extraction Framework: By applying DSP techniques—including sampling, quantization, and Fourier transformation—we develop an effective framework for extracting salient frequency features from time series data. This improves feature quality and contributes to better model learning and classification performance.
  • Improved Classification Performance with Limited Data: Our method demonstrates that high classification accuracy can be achieved even with limited datasets by increasing data diversity through controlled noise injection. This addresses a common limitation in time series analysis where data collection is challenging.
  • Validation Across Diverse Domains: We validate the generalizability of our approach through experiments on various datasets from the UCR Time Series Classification Archive, covering multiple domains such as biology, healthcare, finance, and industry. This confirms that our method is not limited to a specific dataset or field.
The structure of this paper is as follows. Section 2 describes time series data augmentation and DSP. Section 3 details extraction of features for time series classification using noise injection. Section 4 presents the results and performance evaluation, and Section 5 concludes the paper.

2. Related Work

2.1. Time Series Data Augmentation

Time series data play a crucial role in various fields, but it is often challenging to secure enough data. This can lead to overfitting due to the lack of data necessary for model training, ultimately reducing the model’s generalization performance. Data augmentation is a critical technique to address this issue [11]. By applying various transformations to existing data, data augmentation artificially increases the size and diversity of the dataset. This enables the model to learn from a wider range of patterns, preventing overfitting and enhancing generalization performance.
Various data augmentation techniques can be applied to time series data, each generating new data by transforming specific aspects of the original data. Shifting involves moving the time series data by a certain time interval, creating new data by altering the temporal position of the series [12]. This makes the model more robust against temporal variations. Scaling entails enlarging or reducing the values of the time series data by a fixed ratio [13], allowing the model to learn from variations in data magnitude. Cropping involves cutting out a portion of the time series data to use as new data [14], increasing the model’s sensitivity to specific parts of the data. Flipping reverses the time series data from left to right, which can be useful for data with symmetrical patterns [15].
Noise injection is a data augmentation technique that generates new data by adding random noise to the original time series data. Noise injection helps models learn robustly despite noise in the data, ensuring stable operation even with noisy data. Magklaras et al. [16] demonstrated that noise injection can protect the integrity of time series data and enable the detection of malicious data by accessing data validity using statistical analysis and machine learning algorithms at the receiving end. Thus, data augmentation techniques, particularly noise injection, play a crucial role in improving the classification performance of time series data. This paper aims to enhance the classification performance of time series data by leveraging these techniques. Figure 1 shows the application of five data augmentation methods, namely, shifting, scaling, cropping, flipping, and noise injection; (a) shows the original data, (b) displays the result of shifting, (c) depicts scaling, (d) demonstrates cropping, (e) presents flipping, and (f) exhibits the result of noise injection.

2.2. Digital Signal Processing

DSP is a technology that uses digital computers or electronic devices to analyze and transform signals. DSP can process signals in either the time domain or the frequency domain, playing a crucial role in various applications such as noise removal, feature extraction, and signal compression. DSP is highly useful for analyzing the complex patterns of time series data and extracting important information to enhance data quality. This is particularly beneficial in the classification of time series data, where DSP can significantly improve model performance [17].
DSP encompasses various techniques, each tailored to process signals for specific purposes. Frequency analysis converts time-domain signals to the frequency domain to analyze their frequency components, with the Fourier transform being a representative method [18]. Filtering involves removing or emphasizing specific frequency components of a signal. Common types include low-pass filters, high-pass filters, and band-pass filters [19]. Wavelet transformation is a method that analyzes signals in the time-frequency domain, allowing simultaneous examination of temporal changes and frequency components, making it particularly useful for non-stationary signals [20]. This technique can transform time series data to create new representations, highlighting or modifying specific signal patterns. In this paper, features are extracted through a sequence of processes including sampling, quantization, and Fourier transformation [21,22].
DSP plays a critical role in the analysis and classification of time series data. For example, Ghaderpour et al. [23] reviewed various frequency and time-frequency decomposition methods, enabling the analysis of non-stationary time series and irregularly spaced data, and presented advanced techniques applicable across various scientific fields. Such reviews highlight the significance of DSP as a crucial tool in time series data analysis and classification, enhancing data quality and improving model performance. This paper aims to further enhance the classification performance of time series data by leveraging these DSP techniques.

3. Methodology

This paper is composed of three stages. The first stage involves data augmentation through noise injection. The second stage processes the augmented data using DSP. The third stage classifies the data based on the extracted features. Figure 2 shows the overall process of the proposed methodology in this paper.
The first step involves data augmentation through noise injection, where Gaussian noise with a level set at 30% of the standard deviation is used to augment the data tenfold. The next step is feature extraction via DSP, which effectively removes any additional noise introduced during data augmentation. Finally, time series classification is conducted using an existing time series classification model. The classification model employed is LSTM, and its performance is compared and analyzed against baseline models such as the Gated Recurrent Unit (GRU), Temporal Convolutional Network (TCN), and Transformer to demonstrate the superiority of the proposed methodology [24,25,26,27]. Additionally, the study compares the performance of different application orders of DSP and noise injection to optimize the proposed method, evaluates the impact of hyperparameter changes in the DSP stage by varying the intensity of sampling, and conducts ablation studies to analyze various influences. This comprehensive evaluation highlights the distinctiveness of the proposed methodology.

3.1. Data Augmentation Using Noise Injection

Data augmentation is a crucial technique for enhancing the generalization performance of models, especially when there is a lack of sufficient data. In this paper, we improve classification performance by increasing the diversity of time series data through noise injection. Noise injection involves adding random noise to the original data to generate new data, allowing the model to learn various variations and improve generalization performance. Noise injection is a data augmentation technique that generates new data by adding a specific level of random noise to the original data. In this paper, Gaussian noise was used, and the noise level was set to 30% of the standard deviation [28]. Due to the varying data distributions across datasets, using a fixed noise level is less effective. Instead, setting the noise level based on the standard deviation of each dataset allows for more effective data augmentation. This involves adding random values, generated based on the mean and standard deviation of the original data, to each data point. The following formula represents this process:
N o i s e   L e v e l = 0.3 × σ
where σ represents the standard deviation of the data.
In this paper, we generate tenfold-augmented data for each original data point. This means that for each sample in the original dataset, nine additional samples are created, effectively increasing the size of the entire dataset by a factor of ten. This augmentation ratio allows the model to learn from a variety of patterns and variations, thereby preventing overfitting and enhancing generalization performance. Figure 3 shows the crop data from the UCRArchive2018 dataset before and after augmentation [29].
The results in Figure 3 show that the augmented data retain a similar shape to the original data, with slight perturbations. This confirms the effectiveness of the noise injection technique in generating realistic yet diverse data samples. These visual comparisons highlight that while the augmented data are like the original, the noise injection introduces subtle variations, contributing to increased data diversity and enhanced model generalization performance.

3.2. Time Series Feature Extraction Using Signal Processing

DSP is a critical technology used to enhance data quality and extract important information through signal analysis, transformation, and feature extraction. In this paper, DSP is applied in three stages—sampling, quantization, and discrete Fourier transformation (DFT)—to improve the classification performance of time-series data. Each stage addresses specific aspects of signal processing and ultimately contributes to the extraction of high-quality features, maximizing the model’s performance.

3.2.1. Sampling

Sampling is the first stage of converting an analog signal into a digital signal, which involves measuring continuous-time domain signals at regular intervals to transform them into discrete signals. In this paper, experiments confirmed that a sampling rate of 0.5 is most suitable, and it was set as the final sampling rate [30]. A sampling rate of 0.5 captures the frequency components of the signal sufficiently while managing the data volume efficiently. The appropriate selection of the sampling rate maintains the information of the data without loss while preventing the excessive generation of unnecessary data. Figure 4 shows the application of the sampling process to the crop data.
Figure 4 shows (a) the augmented data and (b) the augmented data after applying the sampling process. While there are no significant changes in the data, it can be observed that the overall curvature becomes smoother.

3.2.2. Quantization

Quantization is the process of converting the amplitude values of sampled discrete signals into digital values by dividing them into fixed levels. In this paper, 16-bit quantization is used to convert the signals into digital values. The 16-bit quantization uses 65,536 (216) distinct values to represent the signal amplitude, which allows for sufficient detail representation of the signal while efficiently managing data size [31]. The quantization process is designed to minimize quantization noise. Figure 5 shows the application of the quantization process to the crop data after the sampling process.
Figure 5 shows (a) the data after applying the sampling process and (b) the data after applying both sampling and quantization processes. Since the quantization value is set at 16 bits, which is a very high quantization level, the actual data values do not change significantly. In the case of the crop data, the original values change very slightly, such as from 0.240 to 0.240003 and from 0.219 to 0.219003.

3.2.3. Discrete Fourier Transform

Fourier transformation is a method for converting a time-domain signal into the frequency domain to analyze the signal’s frequency components. In this paper, the DFT is used to extract the frequency features of the sampled and quantized time-series data [32]. The DFT is calculated using the following equation:
X k = n = 0 N 1 x ( n ) × e j 2 π k n / N
where X ( k ) represents the DFT of the signal, x ( n ) is the sampled and quantized signal, N is the number of points in the signal, and k is the index of the frequency component. This transformation allows for the identification of significant frequency components within the data. The frequency components obtained through DFT contain important frequency features of the time-series data, which are used for model training. Figure 6 shows the results after applying all DSP steps to the crop data.
Figure 6 shows (a) the original crop data and (b) the final processed data after applying all DSP steps. It can be observed that most of the feature values have changed significantly from the original data. Algorithm 1 shows the DSP process used in this paper. It consists of sampling, quantization, and DFT.
Algorithm 1 Digital Signal Processing
Require: signal (signal data)
Ensure: magnitude, phase
1: function DFT(signal)
2:   dft_result <- FFT(signal)
3:   frequencies <- FFT_freq(signal)
4:   magnitude <- abs(dft_result)
5:   phase <- angle(dft_result)
6:   return (frequencies, magnitude, phase)
7: end function
8: function QUANTIZE(data, levels)
9:   min_val <- min(data)
10:  max_val <- max(data)
11:  intervals <- linspace(min_val, max_val, levels-1)
12:  quantized_data <- digitize(data, intervals)
13:  mapped_values <- (max_val-min_val)/(levels-1)*quantized_data+min_val
14:  return mapped_values
15: end function
16: function DFT_TO_FEATURES(signal)
17:  (magnitude, phase) <- DFT(signal)
18:  return (magnitude, phase)
19: end function
20: procedure DSP(signal)
21:  sampling_rate <- 0.5
22:  sample_count <- length(signal)*sampling_rate
23:  sampled_indices <- linspace(0, length(signal)-1,sample_count)
24:  sampled_data <- signal[sampled_indices]
25:  quantization_levels <- 2 16
26:  quantized_data <- QUANTIZE(sampled_data, quantization_levels)
27:  (magnitude, phase) <- DFT_TO_FEATURES(sampled_data)
28:  return (magnitude, phase)
29: end procedure

3.3. Time Series Classification Based on LSTM

Time series classification (TSC) is a critical problem in various real-world applications [33]. In this paper, a vanilla long short-term memory (LSTM) model is used to perform time series classification [34]. LSTM, a type of recurrent neural network (RNN), is particularly strong in learning dependencies in long sequence data [35]. In this paper, the LSTM model is employed to classify time series data, and techniques such as data augmentation and DSP are applied to enhance the model’s performance.
In this paper, a vanilla LSTM model is used to perform time series classification. LSTM, a type of RNN, can learn dependencies in long sequence data. The key feature of LSTM is its cell state and gate structure, which allows it to selectively remember and forget information. This capability enables LSTM to effectively learn from long sequence data.
The model training is conducted for 300 epochs without pretraining, applying a tenfold augmentation and DSP to the training data. In contrast, only DSP is applied to the test data to evaluate the model’s performance. The Adam optimizer is used for each epoch to minimize the model’s loss function. Figure 7 shows the structure of the LSTM used in this paper.

4. Result and Performance Evaluation

4.1. Dataset and Performance Evaluation Metrices for Experiments

In this section, the dataset and data augmentation techniques are described. All experiments were conducted on a system equipped with an Intel® Xeon® Silver 2.40 GHz CPU, 128 GB RAM, and a single NVIDIA RTX 3090 GPU.
The performance evaluation in this paper was conducted using the UCRArchive2018 dataset. The UCR is a comprehensive collection of time series datasets used for analysis and classification, encompassing various domains. This dataset is widely employed for comparing and validating the performance of classification algorithms. One of the key features of the UCR dataset is that it includes dozens of datasets, each reflecting the characteristics of different applications and data types. This diversity is highly beneficial for evaluating the generalization ability of algorithms. However, due to the vast number of datasets, making efficient selections for experimentation is essential. In this paper, the following seven major categories are used to effectively categorize and select datasets from the UCR repository: image outline, sensor readings, motion capture, spectrographs, electric devices, ECG, and simulated. By selecting datasets that fit each of these categories, the experimental data can be utilized more efficiently for the intended purpose of the study. Categorizing the diverse UCR datasets in this manner allows for the evaluation of how consistently each algorithm performs across time series data with different characteristics. The datasets chosen from each category are split into training and testing sets, with 20% of the training data used for validation during the training process. All experiments were conducted using untrained models with the number of epochs fixed at 300.
The evaluation metric used in the experiments was accuracy. Accuracy indicates how accurately the model predicts the class of the input image and is calculated as the proportion of correctly classified data out of the total data [36]. The accuracy is computed using the following formula:
A c c u r a c y = T P + T N T P + F P + T N + F N
Accuracy indicates how accurately the model predicts the class of the input image and is calculated as the proportion of correctly classified data out of the total data. Equation (3) shows the method for calculating accuracy [37,38]. TP (true positive) refers to cases where the model correctly predicts a positive instance as positive. TN (true negative) refers to cases where the model correctly predicts a negative instance as negative. FP (false positive) refers to cases where the model incorrectly predicts a negative instance as positive. FN (false negative) refers to cases where the model incorrectly predicts a positive instance as negative.

4.2. Proposed Methods Performance Evaluation

In Section 4.2, we present a comparative performance analysis between the proposed method and existing models. The evaluation metric used for the comparison is accuracy, which represents the proportion of correctly predicted instances out of the total instances. The accuracy results of the experiments are as follows. Table 1 shows the performance evaluation results of the proposed method.
Table 1’s “Normal” represents the performance evaluation of LSTM on each UCR dataset without any preprocessing. The results in Table 1 indicate that our methodology achieves higher performance across various datasets. This demonstrates that noise injection and DSP are effective for time series data classification. Specifically, analysis of the datasets with the most significant performance improvements reveals that our method shows substantial gains in few-shot data and binary classification problems. This finding substantiates the effectiveness of the proposed methodology for few-shot data and binary classification tasks. Conversely, the analysis of datasets where performance did not improve or even declined shows that these typically involve a high number of features and multi-class classification problems. For instance, the CinCECGTorso dataset has 1640 features and four classes. Future work will aim to develop methodologies that are effective for datasets with many features and multi-class classification problems to address these limitations.

4.3. Performance Comparison with Variations in DSP Hyperparameters

In this section, we evaluate the impact of changes in the hyperparameters of the DSP stage on performance by comparing performance across different sampling intensities. This comparison focuses on exploring the flexibility of DSP and its applicability across various scenarios. The experiments were conducted by varying the sampling intensity from 0.1 to 0.9. The datasets with the most significant performance improvements—crop, FordA, and trace—were used for the experiments. Table 2 shows the performance comparison based on changes in the DSP hyperparameters.
The experimental results indicate that the highest performance is achieved when the sampling intensity is set between 0.4 and 0.6. When the sampling intensity is too low or too high, it can result in reduced representativeness of the data or overfitting due to excessive sampling. A sampling intensity between 0.4 and 0.6 provides adequate data diversity, enhancing the model’s generalization capability. Within this range, data diversity is sufficiently maintained, while important features are preserved, allowing the model to learn effectively. This finding can serve as a crucial criterion for optimizing sampling intensity in future work.

4.4. Comparison of Application Order Performance of Noise Injection and DSP

In this section, we evaluate the impact of the application order of DSP and noise injection on performance by comparing the performance based on different sequences. The experiments were conducted under two conditions. The first condition involved applying DSP before performing noise injection augmentation, and the second condition involved performing noise injection augmentation before applying DSP. Table 3 shows the performance comparison based on the application order.
Applying DSP after augmentation optimizes model performance by first increasing data diversity through data augmentation and then enhancing the signal and removing noise using DSP techniques. This sequence allows DSP to effectively eliminate additional noise introduced during data augmentation, resulting in more refined data for training. Conversely, applying data augmentation after DSP can reintroduce noise that was removed in the DSP stage, potentially degrading the quality of the final training data and negatively impacting model performance. Consequently, it is confirmed that applying DSP after data augmentation achieves higher performance. This finding suggests the optimal application sequence of noise injection and DSP techniques for time series data, providing a basis for further performance improvements in future work.

4.5. Comparison with Time Series Models

For Section 4.5, we conducted comparative experiments with various state-of-the-art time series models to evaluate the performance of the proposed methodology. The models selected for comparison included LSTM, GRU, TCN, and Transformer. By assessing the performance of each model, we demonstrated the superiority of the proposed methodology. Table 4 shows the comparative performance of different time series models.
The results in Table 4 show that the proposed methodology outperforms traditional models such as LSTM, GRU, TCN, and Transformer. The superior performance of the proposed methodology highlights the effectiveness of noise injection augmentation and DSP application for time series data classification. This finding suggests a promising approach for future work in time series analysis.

4.6. Ablation Study

For this section, we applied noise injection augmentation and DSP techniques to enhance the performance of time series data classification. To analyze the impact of each technique on model performance, we conducted an ablation study. The experiments were carried out in the following order: vanilla model, DSP application, augmentation application, and the application of both DSP and augmentation. This approach allowed us to evaluate the independent and combined effects of each technique. Table 5 shows the results of the ablation study.
The experimental results indicate that both the application of DSP and augmentation led to performance improvements compared to the vanilla model. Notably, the highest performance improvement was observed when both techniques were applied together. The application of DSP enhances model performance by removing noise from time series data and emphasizing key features of the signal, resulting in an average performance improvement of approximately 9.33% compared to the vanilla model. Additionally, data augmentation increases the diversity of the training data, thereby improving the model’s generalization ability, with an average performance improvement of approximately 8.37% compared to the vanilla model.
When both DSP and data augmentation techniques are applied, the combined advantages of both methods yield the highest performance improvement. This outcome demonstrates that enhancing the signal during the preprocessing stage and simultaneously increasing the diversity of the training data can maximize the learning effectiveness of the model. Specifically, the combined application of both techniques results in an average performance improvement of approximately 16.81% compared to the vanilla model. These findings support the validity of the proposed methodology and demonstrate that DSP and data augmentation techniques play a crucial role in enhancing the performance of time series data classification.

5. Conclusions

In this paper, we proposed a new methodology that combines data augmentation and digital signal processing techniques to enhance the performance of time series data classification. For data augmentation, we employed a noise injection technique to increase the diversity of the training data, augmenting each data sample tenfold by setting the Gaussian noise level to 30% of the standard deviation. Additionally, DSP was conducted through three stages—sampling, quantization, and Fourier transformation—to extract important frequency features of the time series data. This approach improved the quality of the training data and maximized the model’s generalization performance.
The experimental results showed that the proposed methodology significantly improved time series data classification performance compared to existing methods, demonstrating superior performance across various evaluation metrics. These findings confirm that the combination of data augmentation and DSP techniques is an effective tool for addressing time series data classification problems.
The results of this paper have the following implications. First, the integration of data augmentation and DSP techniques supports the effective learning of complex patterns in time series data, indicating that it can overcome limitations caused by data imbalances and noise. Second, achieving high classification performance with a small amount of data suggests potential applicability in fields where data collection is challenging. Furthermore, this paper contributes to existing research as follows. While existing time series classification methods have mainly focused on individual techniques, our paper confirmed the synergistic effect of combining data augmentation and DSP. This presents a new approach in the field of time series data classification and contributes to indicating the direction for future research.
In future work, we plan to further explore the scalability and applicability of the proposed methodology. We aim to develop effective methods for problems with many features and multi-class classification, thereby constructing a generalized model applicable to various time series datasets. Additionally, we will proceed with research to further enhance model performance by combining more complex signal processing techniques and advanced data augmentation methods. Through this, we expect to make practical contributions to solving time series data classification problems.

Author Contributions

Conceptualization, G.I.K. and K.C.; methodology, G.I.K.; software, G.I.K.; validation, G.I.K.; formal analysis, G.I.K.; writing—original draft preparation, G.I.K.; writing—review and editing, K.C.; visualization, G.I.K.; supervision, K.C.; project administration, K.C.; funding acquisition, K.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Kyonggi University Research Grant 2024.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available in the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fawaz, H.I.; Lucas, B.; Forestier, G.; Pelletier, C.; Schmidt, D.F.; Weber, J.; Webb, G.I.; Idoumghar, L.; Muller, P.A.; Petitjean, F. Inceptiontime: Finding alexnet for time series classification. Data Min. Knowl. Discov. 2020, 34, 1936–1962. [Google Scholar] [CrossRef]
  2. Dempster, A.; Petitijean, F.; Webb, G.I. ROCKET: Exceptionally fast and accurate time series classification using random convolutional kernels. Data Min. Knowl. Discov. 2020, 34, 1454–1495. [Google Scholar] [CrossRef]
  3. Dempster, A.; Schmidt, D.F.; Webb, G.I. Minirocket: A Very Fast (Almost) Deterministic Transform for Time Series Classification. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, New York, NY, USA, 14–18 August 2021; pp. 248–257. [Google Scholar]
  4. Yuan, A.E.; Shoi, W. A rigorous and versatile statistical test for correlations between stationary time series. PLoS Biol. 2024, 22, e3002758. [Google Scholar] [CrossRef] [PubMed]
  5. Masini, R.P.; Medeiros, M.C.; Mendes, E.F. Machine learning advances for time series forecasting. J. Econ. Surv. 2023, 37, 76–111. [Google Scholar] [CrossRef]
  6. Fawa, H.I.; Forestier, G.; Weber, J.; Idoumghar, L.; Muller, P.A. Deep learning for time series classification: A review. Data Min. Knowl. Discov. 2019, 33, 917–963. [Google Scholar]
  7. Dhar, M.; Dickinson, J.A.; Berg, M.A. Efficient, nonparametric removal of noise and recovery of probability distributions from time series using nonlinear-correlation functions: Additive noise. J. Chem. Phys. 2023, 159, 054110. [Google Scholar] [CrossRef]
  8. Kim, G.I.; Chung, K. ViT-Based Multi-Scale Classification Using Digital Signal Processing and Image Transformation. IEEE Access 2024, 12, 58625–58638. [Google Scholar] [CrossRef]
  9. Liu, B.; Zhang, Z.; Cui, R. Efficient Time Series Augmentation Methods. In Proceedings of the 2020 13th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Chengdu, China, 17–19 October 2020; pp. 1004–1009. [Google Scholar]
  10. França, R.P.; Monteiro, A.C.B.; Arthur, R.; Iano, Y. An overview of deep learning in big data, image, and signal processing in the modern digital age. In Trends in Deep Learning Methodologies; Piuri, V., Raj, S., Genovese, A., Srivastava, R., Eds.; Academic Press: Amsterdam, The Netherlands, 2021. [Google Scholar]
  11. Iwana, K.; Uchida, S. Time Series Data Augmentation for Neural Networks by Time Warping with a Discriminative Teacher. In Proceedings of the 2020 25th International Conference on Pattern Recognition (ICPR), Milano, Italy, 10–15 January 2021; pp. 3558–3565. [Google Scholar]
  12. Jiang, G.; Wang, W.; Zhang, W. A novel distance measure for time series: Maximum shifting correlation distance. Pattern Recognit. Lett. 2019, 117, 58–65. [Google Scholar] [CrossRef]
  13. Rashid, K.M.; Louis, J. Window-warping: A time series data augmentation of IMU data for construction equipment activity identification. In Proceedings of the International Symposium on Automation and Robotics in Construction (ISARC), Banff, AB, Canada, 21–24 May 2019; pp. 651–657. [Google Scholar]
  14. Iwana, B.K.; Uchida, S. An empirical survey of data augmentation for time series classification with neural networks. PLoS ONE 2021, 16, e0254841. [Google Scholar] [CrossRef]
  15. Ranjan, K.G.; Tripathy, D.S.; Prusty, B.R.; Jena, D. An improved sliding window prediction-based outlier detection and correction for volatile time-series. Int. J. Numer. Model. 2021, 34, e2816. [Google Scholar] [CrossRef]
  16. Magklaras, A.; Andriopoulos, N.; Birbas, A. Noise Injection/Machine Learning Fraud Detection Framework in Time Series Data. In Proceedings of the International Conference on Noise and Fluctuations (ICNF), Neuchâtel, Switzerland, 18–21 June 2019. [Google Scholar]
  17. Karanov, B.; Chagnon, M.; Aref, V.; Ferreira, F.; Lavery, D.; Bayvel, P.; Schmalen, L. Experimental Investigation of Deep Learning for Digital Signal Processing in Short Reach Optical Fiber Communications. In Proceedings of the IEEE Workshop on Signal Processing Systems (SiPS), Coimbra, Portugal, 20–22 October 2020; pp. 1–6. [Google Scholar]
  18. Cen, S.; Kim, D.O.; Lim, C.G. A fused CNN-LSTM model using FFT with application to real-time power quality disturbances recognition. Energy Sci. Eng. 2023, 11, 2267–2280. [Google Scholar] [CrossRef]
  19. Oryad, F.E.; Amindavar, H. Wireless positioning based on hierarchical symbolic dynamic filtering of RSSI time series. Signal Process. 2023, 206, 108903. [Google Scholar] [CrossRef]
  20. Polanco-Martínez, J.M.; Fernández-Macho, J.; Medina-Elizalde, M. Dynamic wavelet correlation analysis for multivariate climate time series. Sci. Rep. 2020, 10, 21277. [Google Scholar] [CrossRef] [PubMed]
  21. Xuan, F.; Dong, Y.; Li, J.; Li, X.; Su, W.; Huang, X.; Huang, J.; Xie, Z.; Li, Z.; Liu, H.; et al. Mapping crop type in Northeast China during 2013–2021 using automatic sampling and tile-based image classification. Int. J. Appl. Earth Obs. Geoinf. 2023, 117, 103178. [Google Scholar] [CrossRef]
  22. Shen, T.; Ren, W.; Han, M. Quantized generalized maximum correntropy criterion based kernel recursive least squares for online time series prediction. Eng. Appl. Artif. Intell. 2020, 95, 103797. [Google Scholar] [CrossRef]
  23. Ghaderpour, E.; Pagiatakis, S.D.; Hassan, Q.K. A survey on change detection and time series analysis with applications. Appl. Sci. 2021, 11, 6141. [Google Scholar] [CrossRef]
  24. Karim, F.; Majumdar, S.; Darabi, H.; Chen, S. LSTM fully convolutional networks for time series classification. IEEE Access 2017, 6, 1662–1669. [Google Scholar] [CrossRef]
  25. Elsayed, N.; Maida, A.S.; Bayoumi, M. Gated Recurrent Neural Networks Empirical Utilization for Time Series Classification. In Proceedings of the International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Atlanta, GA, USA, 14–17 July 2019; pp. 1207–1210. [Google Scholar]
  26. Koh, B.H.D.; Lim, C.L.P.; Rahimi, H.; Woo, W.L.; Gao, B. Deep temporal convolution network for time series classification. Sensors 2021, 21, 603. [Google Scholar] [CrossRef]
  27. Zerveas, G.; Jayaraman, S.; Patel, D.; Bhamidipaty, A.; Eickhoff, C. A Transformer-Based Framework for Multivariate Time Series Representation Learning. In Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery & Data Mining, Singapore, 14–18 August 2021; pp. 2114–2124. [Google Scholar]
  28. Çoban, G.; Büyüklü, A.H.; Das, A. A linearization based non-iterative approach to measure the gaussian noise level for chaotic time series. Chaos Solitons Fractals 2021, 45, 266–278. [Google Scholar] [CrossRef]
  29. Dau, H.A.; Bagnall, A.; Kamgar, K.; Yeh, C.C.M.; Zhu, Y.; Gharghabi, S.; Ratanamahatana, C.A.; Keogh, E. The UCR time series archive. IEEE/CAA J. Autom. Sin. 2019, 6, 1293–1305. [Google Scholar] [CrossRef]
  30. Pickens, A.H.; Hansen, M.C.; Hancher, M.; Stehman, S.V.; Tyukavina, A.; Potapov, P.; Marroquin, B.; Sherani, Z. Mapping and sampling to characterize global inland water dynamics from 1999 to 2018 with full Landsat time-series. Remote Sens. Environ. 2020, 243, 111792. [Google Scholar] [CrossRef]
  31. Chiarot, G.; Silvestri, C. Time series compression survey. ACM Comput. Surv. 2023, 55, 1–32. [Google Scholar] [CrossRef]
  32. Jalayer, M.; Orsenigo, C.; Vercellis, C. Fault detection and diagnosis for rotating machinery: A model based on convolutional LSTM, Fast Fourier and continuous wavelet transforms. Comput. Ind. 2021, 125, 103378. [Google Scholar] [CrossRef]
  33. Faouzi, J. Time Series Classification: A review of Algorithms and Implementations. In Machine Learning (Emerging Trends and Applications); Proud Pen: London, UK, 2022. [Google Scholar]
  34. Abbasimehr, H.; Paki, R. Improving time series forecasting using LSTM and attention models. J. Ambient. Intell. Humaniz. Comput. 2022, 13, 673–691. [Google Scholar] [CrossRef]
  35. Hu, J.; Wang, X.; Zhang, Y.; Zhang, D.; Zhang, M.; Xue, J. Time series prediction method based on variant LSTM recurrent neural network. Neural Process. Lett. 2020, 52, 1485–1500. [Google Scholar] [CrossRef]
  36. Lee, S.E.; Yoo, H.; Chung, K. Pose Pattern Mining using Transformer for Motion Classification. Appl. Intell. 2024, 54, 3841–3858. [Google Scholar] [CrossRef]
  37. Baek, J.W.; Chung, K. Meta Learning based Object Tracking Technology: A Survey. KSII Trans. Internet Inf. Syst. (TIIS) 2024, 18, 2067–2081. [Google Scholar]
  38. Jeon, B.U.; Chung, K. CutPaste-based Anomaly Detection Model using Multi-scale Feature Extraction in Time Series Streaming Data. KSII Trans. Internet Inf. Syst. (TIIS) 2022, 16, 2787–2800. [Google Scholar]
Figure 1. Five data augmentation results.
Figure 1. Five data augmentation results.
Sensors 24 06402 g001
Figure 2. Process of extraction of features for time series classification using noise injection.
Figure 2. Process of extraction of features for time series classification using noise injection.
Sensors 24 06402 g002
Figure 3. Comparison of crop data before and after noise injection.
Figure 3. Comparison of crop data before and after noise injection.
Sensors 24 06402 g003
Figure 4. Comparison of crop data before and after sampling.
Figure 4. Comparison of crop data before and after sampling.
Sensors 24 06402 g004
Figure 5. Comparison of crop data before and after quantization.
Figure 5. Comparison of crop data before and after quantization.
Sensors 24 06402 g005
Figure 6. Comparison of crop data before and after discrete Fourier transformation.
Figure 6. Comparison of crop data before and after discrete Fourier transformation.
Sensors 24 06402 g006
Figure 7. Structure of the LSTM.
Figure 7. Structure of the LSTM.
Sensors 24 06402 g007
Table 1. Performance evaluation of proposed methods.
Table 1. Performance evaluation of proposed methods.
CategoryDataLSTM
NormalOurs
Image OutlineCrop63.26%72.80%
ShapesAll69.50%72.17%
InlineSkate30.55%33.09%
Sensor readingsDodgerLoopGame52.17%52.17%
DodgerLoopWeekend73.91%73.91%
FordA78.33%94.24%
Motion captureWafer99.03%99.48%
CricketX44.87%54.87%
SpectrographsTrace63.0%88.0%
ChlorineConcentration58.80%73.59%
Electric devicesSonyAIBORobotSurface180.03%83.86%
SonyAIBORobotSurface282.69%85.67%
Table 2. Performance comparison with variations in DSP hyperparameters.
Table 2. Performance comparison with variations in DSP hyperparameters.
CropFordATrace
0.150.51%82.80%81.00%
0.266.71%92.27%87.00%
0.367.68%93.79%85.00%
0.469.12%94.17%88.00%
0.572.80%94.24%85.00%
0.672.42%93.48%83.00%
0.772.30%93.79%82.00%
0.871.74%94.02%83.00%
0.972.79%93.94%82.00%
Table 3. Comparison of application order performance of noise injection and DSP.
Table 3. Comparison of application order performance of noise injection and DSP.
CropFordATrace
DSP → Aug68.76%93.79%76.00%
Aug → DSP72.80%94.24%88.00%
Table 4. Comparison between time series models.
Table 4. Comparison between time series models.
CropFordATrace
LSTM [24]63.26%78.33%63.00%
GRU [25]61.29%79.92%43.00%
TCN [26]43.54%79.39%68.00%
Transformer [27]68.69%74.62%68.00%
Ours72.80%94.24%88.00%
Table 5. Ablation study.
Table 5. Ablation study.
LSTMDSPAugDSP + Aug
Crop63.26%66.70%71.95%72.80%
FordA78.33%92.99%79.77%94.24%
Trace63.00%73.00%78.00%88.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, G.I.; Chung, K. Extraction of Features for Time Series Classification Using Noise Injection. Sensors 2024, 24, 6402. https://doi.org/10.3390/s24196402

AMA Style

Kim GI, Chung K. Extraction of Features for Time Series Classification Using Noise Injection. Sensors. 2024; 24(19):6402. https://doi.org/10.3390/s24196402

Chicago/Turabian Style

Kim, Gyu Il, and Kyungyong Chung. 2024. "Extraction of Features for Time Series Classification Using Noise Injection" Sensors 24, no. 19: 6402. https://doi.org/10.3390/s24196402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop