Next Article in Journal
Incorporating Concreteness in Multi-Modal Language Models with Curriculum Learning
Next Article in Special Issue
Homomorphic Encryption Based Privacy-Preservation for IoMT
Previous Article in Journal
Application of Emerging Cell Disintegration Techniques for the Accelerated Recovery of Curcuminoids from Curcuma longa
Previous Article in Special Issue
Trust No One? A Framework for Assisting Healthcare Organisations in Transitioning to a Zero-Trust Network Architecture
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sample Reduction for Physiological Data Analysis Using Principal Component Analysis in Artificial Neural Network

by
Cid Mathew Santiago Adolfo
1,*,†,‡,
Hassan Chizari
1,‡,
Thu Yein Win
2,‡ and
Salah Al-Majeed
3
1
School of Computing & Engineering, University of Gloucestershire, Cheltenham GL50 2RH, UK
2
Department of Computer Science and Creative Technologies, University of West England, Bristol BS16 1QY, UK
3
School of Computer Science, University of Lincoln, Lincoln LN6 7TS, UK
*
Author to whom correspondence should be addressed.
Current Address: 4 Fairhaven Street, Cheltenham GL53 7PL, UK.
These authors contributed equally to this work.
Appl. Sci. 2021, 11(17), 8240; https://doi.org/10.3390/app11178240
Submission received: 20 July 2021 / Revised: 24 August 2021 / Accepted: 29 August 2021 / Published: 6 September 2021
(This article belongs to the Special Issue Cyber Security in Human Health and Medical Devices)

Abstract

:
With its potential, extensive data analysis is a vital part of biomedical applications and of medical practitioner interpretations, as data analysis ensures the integrity of multidimensional datasets and improves classification accuracy; however, with machine learning, the integrity of the sources is compromised when the acquired data pose a significant threat in diagnosing and analysing such information, such as by including noisy and biased samples in the multidimensional datasets. Removing noisy samples in dirty datasets is integral to and crucial in biomedical applications, such as the classification and prediction problems using artificial neural networks (ANNs) in the body’s physiological signal analysis. In this study, we developed a methodology to identify and remove noisy data from a dataset before addressing the classification problem of an artificial neural network (ANN) by proposing the use of the principal component analysis–sample reduction process (PCA–SRP) to improve its performance as a data-cleaning agent. We first discuss the theoretical background to this data-cleansing methodology in the classification problem of an artificial neural network (ANN). Then, we discuss how the PCA is used in data-cleansing techniques through a sample reduction process (SRP) using various publicly available biomedical datasets with different samples and feature sizes. Lastly, the cleaned datasets were tested through the following: PCA–SRP in ANN accuracy comparison testing, sensitivity vs. specificity testing, receiver operating characteristic (ROC) curve testing, and accuracy vs. additional random sample testing. The results show a significant improvement in the classification of ANNs using the developed methodology and suggested a recommended range of selectivity ( S c ) factors for typical cleaning and ANN applications. Our approach successfully cleaned the noisy biomedical multidimensional datasets and yielded up to an 8% increase in accuracy with the aid of the Python language.

1. Introduction

It is clear that biomedical sensors will keep improving exponentially, as they become more accessible; more readily available on the market; more intelligent; smaller and more compact; and integrated into personal belongings, such as cellular phones, watches, and eyeglasses. It is an outstanding contribution to human health, entertainment, the military, security, sports, and leisure and in analysing a patient’s physiological data and their interpretation. These sensors are an integral part of biomedical devices; however, this innovation has an ever-evolving trend coupled along with challenges to perform intelligently [1]. The integration of these sensors, subjected to different environments in both wired and wireless implementations, adds random noise to the system. The sudden upswing of the noise floor yields a degradation in the signal-to-noise ratio (C/N) and low energy bit per noise reference (Eb/No) during data processing. The drastic increase in the probability of error (Pe) eventually decreases the accuracy in such a system [2]. This interference results in flawed calculation and decision making, especially for artificial neural networks (ANNs) in biomedical applications [3].
An artificial neural network (ANN) is a part of a computing system that simulates the ability of the human neuron to learn the complex characteristics of the environment, to recognise patterns, and to generalise the inter-relationships between the features, including multidimensional datasets. It is part of the umbrella of artificial intelligence (AI) and deep learning, and it solves problems that would be impossible or difficult to solve by human effort or statistical criteria. ANNs have self-learning skills, enhancing their efficiency as more effective and reliable data become available to them. Primarily, ANNs enable the complex inter-relationships between the features within a given dataset to be identified and have seen widespread adoption in different applications, including biomedical and signal processing, which is readily and publicly available by wearable sensors [4].
Since data are everywhere—such as on the Internet—neglecting the integrity of a source of information poses a significant threat for ANNs to misinterpret the data. For critical aspects of the medical field, the efficacy of a diagnosis is threatened if the analysis that aids medical experts in research has a weak source, especially if dealing with the multitude of datasets that result in the presence of noisy and biased data, which often occurs from seemingly reliable sources [5].
Low-powered sensors are a critical source of unreliable information. Sudden changes in conditions can introduce environmental noise, and there are many possible avenues for noise and interference to blend into any part of the system. A sensor’s scalability is another problem, given that the sensors nowadays are low power and with limited computation. Placement or location is another significant constraint, and if they are placed too close to other devices, they may be affected by crosstalk interference [6].
To understand the data, first, one needs to understand their composition. Data comprise valid and dirty data. Accurate data contain information that is accurate, holds predictive power, and is generalised to the entire set. Dirty data contain information that is misleading, noisy, or erroneous data, such as pragmatic contexts and the semantic- and syntactic-biased errors shown in Figure 1 [7].
Noisy data are any random fluctuations considered unwanted, unpleasant, loud, or disruptive that hinder the generalisation of the entire dataset. For an ANN with predictive power to analyse the data, the dataset needs to be free from noise to reduce errors in classification and prediction, which would significantly impact medical experts in their professional biomedical interpretation. Given recent technological solutions, one relevant control algorithm is highlighted in solving these challenges [8].
Principal component analysis (PCA) is usually part of data visualisation. PCA converts the features of datasets from high dimensional data and can help convert data to low-dimensional data with the aid of covariance, eigenvalues, and eigenvectors in identifying and sorting the strength of the predictive power of all features [9]. PCA allows for the extraction of critical feature vectors from multidimensional datasets in terms of data visualisation and loses some crucial information along the process [10], which is a considerable disadvantage, primarily if it is used extensively in the field of machine learning. This study used PCA not as a dimension reduction but as a sample reduction in order to remove the unwanted noisy samples in multidimensional datasets.
This paper proposes an application of a PCA–sample reduction process (SRP) to improve the prediction accuracy of ANNs. Using publicly available biomedical datasets from a different field provides qualitative analysis and demonstrates the effectiveness of said method in improving the accuracy of ANNs. By cleaning the publicly available biomedical datasets before training and testing the ANN, the classification problem is expected to increase the accuracy and to lower the computational cost, which is a great help in dealing with the the analysis and prediction of big multidimensional datasets.
A study to reinforce the methodology of an ANN using PCA–sample reduction is proposed in this paper to determine the proposed system’s accuracy and performance with different sizes of multidimensional biomedical datasets. Furthermore, we investigate heart disease, voice and speech analysis for gender recognition, breast cancer classification, and cancer patient datasets to prove the versatility and flexibility of the proposed data-cleansing technique.
The rest of the paper is structured as follows. Section 2 provides a literature review of data cleaning in biomedical applications. Section 3 discusses the basic concepts of principal component analysis and related topics. Section 4 discusses how the PCA–SRP is integrated with an ANN. Section 5 provides the recommended S c ranges. Section 6 interprets the result of publicly available biomedical datasets. Lastly, Section 7 provides a discussion, the conclusion, and future research directions.

2. Data-Cleaning Applications

The volume of data collected nowadays is vastly increasing, and since most data acquired are polluted, the dependability of the data is declining. Various data-cleaning methodologies are available to rectify this issue, but data cleansing remains difficult when working with large data requirements. Data cleaning, also known as data cleansing, is no longer a recent area of research. It aims to increase data quality by detecting and eliminating errors and inconsistencies [11]. As of now, there are two classifications of data cleansing: traditional data cleansing and data cleansing for big data. Traditional data cleansing techniques are so called because it is not used to manage massive volumes of data, such as Potter’s Wheel and Intelliclean [12].
Meanwhile, the techniques in Table 1 such as Cleanix [13], SCARE [14], KATARA [15], and BigDansing [16] are developed specifically for big data. Regarding the emerging trends in data-cleaning techniques, one of the new challenges that researchers are about to face is scalability [17]. One of the perennial problems in data analytics is identifying and restoring dirty data, and failure to do so will result in faulty analytics and unreliable decisions. New abstractions and scalability are among the various facets of this issue and are considered when developing data-cleaning methods to cope with the amount and diversity of data [13,14,16]—see Table 2. Given the significant amount of data, it needs time to be processed to be suitable for big data analysis and decision making. The data’s volume, veracity, and velocity must also be considered when analysing the proposed approaches; however, the researchers mentioned that “Data analytics is not about having the information known, but about discovering the predictive power behind the data collected” [12].
Cleanix, SCARE, and BigDansing focus on the scalability issue in the data cleansing process. Moreover, SCARE and BigDansing do not require any human-domain expert in the cleansing process. SCARE needs an extensive set of rules to update the dataset; however, no expert is present in the process. Nevertheless, the process is expensive, and if the authorities fail to identify correct fixes for the dirty dataset, it will result in redundancy of the training data and a threshold machine learning parameter that is hard to set precisely [15]. Furthermore, BigDansing also requires a set of data-quality rules for optimisation of the cleansing process that requires too many regulations to calibrate and put into place before the start of the cleaning process, as shown in Table 2 below; however, it needs no human-domain expertise to monitor the whole process, although adjusting such parameters is crucial in maintaining the essence of the information in the datasets [12].
These data-cleansing techniques support various data-cleaning tasks such as abnormal value detection and correction, incomplete data filling, de-duplication and conflict resolution (Cleanix), value modification (SCARED), the identification of correct and incorrect data, and the generation of top-k possible repairs for inaccurate data (KATARA) and rules into a series of transformations that enable distributed computations and several optimisations (BigDansing).
There are two kinds of dirty data: erroneous and noisy—see Figure 1. The current big-data-cleansing techniques—Cleanix [13], SCARE [14], KATARA [15], and BigDansing [16]—are focused mostly on erroneous dirty data such as a solution in duplicate entries, missing values, wrong values, and wrong formats. In medical applications, these big-data-cleansing techniques play a critical role in medical record management in medical facilities; however, they do not address noisy data acquired via biomedical sensors. Cleanix, KATARA, and BigDansing are not able to predict the correct values through a machine learning approach and cannot determine how to eliminate the mixed random number [12]; however, the SCARE technique, though constructed via machine learning, can only replace the missing values with the most precise value. When dealing with noisy data, the common approach in applications with machine learning and artificial neural network (ANN) is to address noise reduction and suppression [11,18,19].
Given all of the above data-cleaning techniques, noise and random sample reduction is a significant part of their scope of discussion, which is vital in biomedical data signal analyses acquired in wearable biomedical sensors due to its subjectiveness in a different kind of environment—an environment often filled with varying sorts of noises, such as thermal and acoustic noises and interference. Moreover, since most wearable biomedical sensors are of low power, by suppressing these kinds of noises through signal processing by applying the filtering threshold method, the unsupervised classification is not effective under a low SNR. When the spectral characteristics of the noise are so similar or near that of the sensor-received signal, the detection performance may be degraded [20]. When it applies to the classification problem of an artificial neural network (ANN), obtaining the correct values through comprehensive and extensive quantisation in data-signal processing is essential. Still, it is not sufficient to identify which of the gathered data have a predictive power, as some authors mention that “Data analytics is not about having the information known, but about discovering the predictive power behind the data collected” [12]. Correct values of the data do not guarantee that it holds a valid value in predicting and summarising the entirety of the multidimensional datasets. Mining those correct values is an integral part of data mining in every machine learning and big data analysis.
Complicated and straightforward mistakes are unavoidably present in data input and acquisition. Although much effort could be expended into this front-end procedure to reduce entry mistakes, the truth remains that mistakes in massive datasets are prevalent. Field error rates usually are about 5% or higher unless an organisation takes extraordinary precautions to prevent data inaccuracies [21]. Moreover, this rate is still quite high, at such a rate that it might lead to erroneous interpretation and decision making.
In the case of cleaning the noisy data via biomedical sensors, most researchers use principal component analysis, such as a reduction in dimensionality or feature space [22,23,24,25,26], feature extraction in further data visualisation [27,28,29], and feature selection tools in machine and deep learning applications [30,31,32,33,34] for machine learning applications.
Principal component analysis (PCA) is usually utilized in dimension or feature reduction and provides a significant increase in accuracy and efficiency along with other machine learning techniques in many applications aside from biomedical application [35,36,37,38,39]. Nevertheless, some of the information is lost during the process of dimension reduction [10].
Figure 2 shows a sample Excel dataset, and PCA reduces the dimensions vertically (by column), shown in red; for the proposed methodology, it reduces the noisy samples horizontally (by rows). We emphasise that the extraction of features, such as covariance, eigenvalues, eigenvectors, and dimension reduction, is not a novel technique that we propose here, but instead, we propose the implementation of sample reduction using PCA. Identifying the noisy samples that cause irregularity in the multidimensional datasets and omitting or reducing them is the main focus of this study.
The observation and the application of the PCA–sample reduction process are utilised for data cleansing of the noisy multidimensional datasets to increase the accuracy of classification problems in artificial neural networks and to identify recommended threshold ranges. To the best of our knowledge, we are the first to propose this unique approach. We discuss a specific technique for qualitative noise detection and omission of such. These techniques are explained with a motivating example highlighting deep learning classification problems of artificial neural networks in biomedical applications using publicly available biomedical datasets. The simplicity of this technique makes it portable and can apply to a variety of tasks for fast and accurate classification of typical or commonly used artificial neural networks.
In terms of software implementation, developers and data engineers prefer Python programming. It is the best option for projects or programming involving AI and big data analysis as Python is a simple language with a mature and supportive Python community; an abundance of support from renowned corporate sponsors; an extensive and popular selection of libraries; and the ability to work with heavy-hitting frameworks such as TensorFlow, Sci-kit-learn, OpenCV, and Keras [40].

3. Principal Component Analysis–Sample Reduction Process

Principle component analysis (PCA) aims to minimise the dimensionality of a dataset with many connected variables while keeping as much variance as feasible. The conversion of the new collection of uncorrelated variables known as principal components (PCs) preserves most of the variance included in the original variables. A new set of dimensions or orthogonal measurements are linearly independent and ranked according to the data variance. This means the more crucial principal axis occurs first (more important = more variance/more spread-out data). In general, understanding the PCA, variance, covariance, eigenvalues, and eigenvectors has an essential role in this concept [41].
Figure 3 shows large positive covariance means that X and Y are completely related, i.e., as X increases, Y also increases. Negative covariance portrays the exact opposite relation. Zero covariance means that X and Y have no relation.
The visualisation of data is an excellent approach to understand the patterns that lie in the multidimensional datasets. When information is placed in the horizontal and vertical axes (two-dimensional plane), it is straightforward to understand and discern the pattern that lies in it; however, the difficulty of conceiving it visually in multidimensional data with many features to consider and computing the data analysis becomes complex. Principal component analysis prerequisites require the discovery of patterns between the datasets so that the data are distributed across each dimension by first analysing the contributions of each feature in providing information to the overall dataset through eigenvector analysis. It then performs dimension reduction by keeping the feature columns with the highest eigenvalues.
To demonstrate the efficacy of PCA, Figure 4, which projects a set of multidimensional data onto a two-dimensional space, is used as an example. Due to the high dimensional nature of the data points, it can be challenging to identify a linear correlation between the data points. The points are represented as column vectors before being aggregated into a matrix M. PCA is then applied to M as shown in Equation (1):
P C A ( M ) = W × V
where W and V are the corresponding eigenvalues and eigenvectors, respectively.
V P C 1 is generated using covariant eigenvalues V with the corresponding top k eigenvalue in W, which best represents the data points, and are then selected, before plotting them as shown in Figure 5 [30]. The principal axes ( P C 1 and P C 2 ) are denoted by the red lines passing through the points that provide a graphical representation of the covariance amongst the points.
Lastly, the vital part of the proposed PCA–sample reduction process ( P C A S R P ) is as follows:
P C A S R P ( M ) = | M × V P C 1 |
where P C A S R P ( M ) is the loading score of each sample in dataset M and V P C 1 is the P C 1 eigenvectors of the samples in matrix M.
S = 0 D 1 if P C A S R P ( M ) > S c × P C A S R P ( M ) h i g h e s t 0 otherwise
S is the number of samples above the loading score of set-biased S c , and S D , where D is the number of samples in the dataset, including the random and noise samples.
p a s s r a t e = S / D
The p a s s r a t e is the rate of samples cleaned using PCA–SRP in dataset M based on Equation (4). Any samples above said threshold are accepted, and those below it will be rejected for the new process dataset D P S as per Equation (3).

4. PCA–SRP Implementation in ANN

The proposed framework addresses the limitations of the ANN in processing multidimensional datasets through the use of PCA with the sample reduction process (PCA–SRP). PCA is used to analyse the multidimensional data, and if there exists a significant relationship between the features of a dataset, it is arranged from the most significant samples to the least to be visualised and the multidimensional data are put into perspective. It focuses on implementing dimension reduction by removing data from the multidimensional dataset feature columns that do not have high inter-feature covariance; however, the proposed framework using the loading scores can dissect all of the samples or rows in the multidimensional dataset. Figure 6 provides a graphical description of the proposed PCA–SRP-based ANN solution.
The PCA–SRP Python program generates a new set of multidimensional data with fewer samples based on the screen plot to identify the most significant samples. It is used in the Python ANN program as an input in order to ensure correct classification.
Figure 7 and Figure 8 show the algorithm used in Python in the implementation and analysis of the concept. The implementation starts by extracting the data in .csv format and by injecting it into the PCA–SRP process. Major principal components are extracted through the threshold mentioned above based on the input dataset, resulting in a new set of multidimensional datasets used in ANN.
The proposed PCA–SRP utilised in the ANN consists of two hidden layers consisting of 32 neurons and 16 neurons. The learning rate was set at 0.1 and trained for 100 epochs. The input parameters in the PCA–SRP-based ANN are shown in Table 3 [42].
The accuracy of the PCA–SRP-based ANN was compared against a model trained using ANN alone on different datasets.

5. Sc Range Identification

Accuracy testing was conducted for the specified dataset in order to identify the S c range shown in Figure 9, to maximise the number of samples of the cleaned set S, and to maximise the number of samples of the removed set R in the processed D P C A S R P .
Assume that there are two sets of data D: cleaned data S and R.
D = S + R
% N o i s e = R / S
R = % N o i s e × S
Therefore,
D = S + % N o i s e × S
D = S × ( 1 + % N o i s e )
By indexing the dataset D for identifier, where n is the number of cleaned data S, we have the following:
D = S if 0 index n R if n < index n × ( 1 + % N o i s e )
D R randomly distributes R in S to thoroughly mix the noise in the cleaned data by maintaining its index for identification. Then, D R is processed in PCA–sample reduction and sorted based on its loading score through the concept of covariance, eigenvalues and eigenvectors, and D P C A S R P .
S c is the selectivity of the data based on the equation in Table 3, D P C A S R P , where the cut-off is biased to separate the clean-processed data D P S and to remove processed data D P R . The dataset was transformed from random to sorted PCA-SR-processed data.
Transformation:
D D R D P C A S R P
Thus,
D = S R
D P C A S R P = S P C A S R P R P C A S R P
By definition, S a c c u r a c y (true positive rate or sensitivity) is the number of correct samples in a clean-processed S in D P S with a relative to number of n.
S a c c u r a c y = S D P S / n
However, R a c c u r a c y (True Negative Rate or Specificity) is the number of correct samples of R in D P R .
R a c c u r a c y = R D P R / ( n × % N o i s e )
Therefore, the most efficient S c is where S a c c u r a c y and R a c c u r a c y meet.

5.1. Test Selectivity ( S c t e s t )

To find S c t e s t , the actual cleaned S set and random R set are needed to identify the accuracy by identifying the S a c c u r a c y and R a c c u r a c y lines and by applying Equation (16).
S c t e s t = i n d e x [ m a x ( S a c c u r a c y × R a c c u r a c y ) ]
For S c m i n , S a c c u r a c y and S a c c u r a c y are maximised:
( S a c c u r a c y ) ( S c ) = 0
( R a c c u r a c y ) ( S c ) = 0
As shown in Equation (16),
F v a l u e = S a c c u r a c y × R a c c u r a c y
F v a l u e ( S c ) = S a c c u r a c y × ( R a c c u r a c y ) ( S c ) + R a c c u r a c y × ( S a c c u r a c y ) ( S c ) = 0
2 F v a l u e ( S c ) 2 < 0
Maximum F v a l u e with respect to S c is the S c t e s t by satisfying the Equation (21).
Note: Be cautious when using S c t e s t as S c ; it might remove all of the unwanted sample sets but may lose some information.

5.2. Minimum Selectivity ( S c m i n )

The minimum S c is the rate of samples in P C 1 , which shows that all included samples pass the minimum requirements of having the same attributes as the entire set.
S c m i n = P C 1 r a t e
Note: By using S c m i n as S c , while it might not lose important information, the dataset D includes an abundance of random and noise samples.
The acceptable S c is shown in Equation (23) in Figure 10.
0 S c m i n S c S c t e s t 100
Then,
S c m i n S c S c t e s t 100
Therefore, the recommended S c is between the range given in Equation (25). It varies depending on how much cleaning and removing are needed in the dataset D.
S c m i n S c 100

5.3. ANN Selectivity ( S c A N N )

The ANN requires samples that allow for strong predictive power (see Section 7 for more details concerning the ANN requirements); we assumed the following:
S c t e s t S c A N N 100
However, S c t e s t is unknown due to unidentifiable R samples, and it is suggested that S c A N N is close to 100%.

6. Multidimensional Datasets

The multidimensional datasets for classification problems are tested under two layers of artificial neural network with 32 and 16 neurons, respectively, using ReLU and SoftMax activators, a testing size of 20%, a training size of 80%, 100 epochs, a learning rate of 10% [42], and S c = 98% of the highest loading score in PCA analysis. Noisy data based on a standard distribution are added to the original dataset before training to simulate noisy data samples. Figure 11 shows how a given multidimensional dataset is preprocessed and used in the proposed PCA–SRP approach. Each multidimensional dataset or corpora has selected the variation in its dimensions, primarily sample sizes; the number of classifications shown in Table 4; the noise sample contents; and the purpose it serves in any field of biomedical applications in classification problems.
The different multidimensional datasets are shown in Table 4 and were acquired from different scientific and medical laboratories [43,44,45,46,47]. Specifically, to test our procedure, we vary the number of samples and dimensions. Moreover, researchers also are trying to show that the Python implementation works with a variety of dataset sizes.

6.1. Heart Disease Dataset

The experiment with the Cleveland patients dataset is concentrated on simply attempting to distinguish presence (values 1) from absence (values 0) to find other trends in heart data to predict certain cardiovascular events or to find clear indications of heart health [43].
Figure 12 and Figure 13 show the result of the scree plot based on the values, indicating the PCA value distributions of the features in the dataset and the S c -threshold cutoff limit, respectively. Based on the values, 78.48% of the top samples were selected in the training of the ANN model for a 98% selectivity.

6.2. Gender Voice Recognition Dataset

The gender voice recognition dataset is based upon acoustic properties of the voice and speech to identify a voice as male or female. The dataset consists of 3167 recorded voice samples collected from male and female speakers. The voice samples are preprocessed by acoustic analysis in R using the Seewave and TuneR packages, with an analysed frequency range of 0–280 hz (human vocal range) [44].
Figure 14 and Figure 15 show the threshold cutoff limit and the result of the scree plot show the PCA distributions of the features in the dataset, respectively. In the training of the ANN model, the proposed method selected 23.82% of the top samples from the dataset.

6.3. Breast Cancer Classification Dataset

The computed dataset consists of a series of digitised images of a mass breast fine needle aspirate (FNA). They describe characteristics of the cell nuclei present in the three-dimensional image as described in [45].
Figure 16 and Figure 17 show the threshold cutoff limit and the results of the scree plot showing the PCA distributions of the features in the dataset, respectively. The selection of the top eigenvalues in the samples obtains 93.67% of the original dataset.

6.4. Cancer Patients Dataset

The data comprise information about hundreds of cancer patients and their lifestyles. It consists of three classes (low, medium, and high severity), based on the cancer patients dataset [47].
Figure 18 and Figure 19 provide the threshold cutoff limit and a graphical description of the PCA distributions amongst the features in the dataset, respectively. The selection of high eigenvalues features resulted in 45.41% of the samples in the original dataset used for the model’s training.

7. Discussion and Results

Upon acquiring the given datasets, they are contiminated with noise and random samples. By definition, noise is small, unwanted forms of energy or samples; on the other hand, random samples are the unconscious and unspecified values within the range of normal values.
The datasets with noise and random samples are processed using PCA–SRP with 98% selectivity ( S c ), shown in Table 5, and then subjected to following tests:
  • PCA–SRP + ANN comparison accuracy testing compares the validation model accuracy with and without PCA–SRP in an ANN.
  • Sensitivity vs. specificity testing is a diagnostic test to find the approximate S c range values ( S c t e s t ).
  • Receiver operating characteristic (ROC) curve testing compares the methodology PCA–SRP in different datasets in terms of organisation and classification of samples. Moreover, ROC curves also provide a practical evaluation of machine learning techniques.
  • Accuracy vs. additional random samples testing is a diagnostic test responding to the sudden increase in noise and random sample spikes.

7.1. PCA–SRP + ANN Comparison Accuracy Testing

We compared the validity model accuracy using PCA–SRP in an ANN classification problem as suggested by the results presented in Figure 20 and determined its effect upon being subjected to noise and random samples. Table 6 and Table 7 display the validation accuracy using ANN + PCA–SRP and ANN only, showing a significant increase as shown in Figure 21 for both noise and random samples.
The difference between dataset accuracy subjected to noise and random samples are the definition of the accuracy curve line between ANN + PCA–SRP to ANN only, even though both of them have a significant accuracy increase. The accuracy of datasets under the influence of random samples swings and deviates more and is less defined in comparison with accuracy with noise samples.

7.2. Sensitivity vs. Specificity Testing

Sensitivity measures how many true positives remain in S set, and it is described as a sudden dip as S c increases. Specificity measures how many true negatives remain in the removed R set and increases as S c increases. Figure 22 presents the diagnostic testing yielding a recommended S c range in normal cleaning and ANN applications, as demonstrated in Equations (25) and (26), respectively.
Table 8 shows the relation between S c m i n and S c t e s t , but in practicality, S c t e s t is hard to determine due to unknown R sets, and whether it consists of noise or random samples. In general, for the normal cleaning process, S c is close to or above the S c m i n or P C 1 rate. For the given datasets, it is suggested that S c n o r m a l is as follows:
80 % S c n o r m a l 96 %
However, ANN applications need highly predictive samples [12], so it is suggested that selectivity ( S c ) should be almost near 100%. Table 8 shows S c A N N at the following range:
90 % S c A N N 99 %
A high S c value loses information and true positives in the S set but increases the specificity of the dataset; however, a low value of S c adds noise and random samples that yield fewer true negatives and increases sensitivity. Careful adjustment of S c vouches for a good result as a cleaning agent in the system.

7.3. Receiver Operating Characteristic (ROC) Curve Testing

As observed in Figure 23, receiver operating characteristic (ROC) datasets have shown an organisation of the samples, even the random and noise samples were strongly mixed into the cleaned set. The larger the area under the curve (AUC), the better the classifier methodology for the true positive rate (sensitivity) vs. true negative rate (1 − specificity) diagram, as seen in Figure 24. Ideally, the objective is the perfect classifier; nevertheless, a result above the random classifier line would allow us to conclude that the methodology is acceptable.
Given all of the datasets, the cancer patients dataset is acceptable but by not as much as the other datasets; it has a smaller AUC.

7.4. Accuracy vs. Additional Random Samples Testing

Since the cancer patients dataset shows the least-effective classifier, as shown in Figure 23, it was tested for its response to the injection of the a sudden spike in additional random samples up to 100%. Figure 25 shows that, upon increasing the random samples, the result of PCA–SRP decreases gradually.
Figure 26 and Figure 27 present the validation accuracy of both ANN + PCA(SRP) and ANN only to additional noises up to 100% with selectivities (Sc) of 88 and 98%, respectively. It has also shown a reasonable increase in accuracy using 98% selectivity ( S c ) instead of 88%.
The high-valued data have been preserved and maintain their accuracy until a specific additional noise point. Nevertheless, the predictive power remains intact until that point, even though some data were lost in the process.
The A N N + P C A S R P maintains the highest performance of cancer patient classifications with high S c values. The methodology allows for a significant advantage by gradually lessening the decrease in classification problem accuracy over the sudden increase in noise in the system.
The ANN classification problem requires strong training sets, which requires a lot of high S c while disregarding the low ones; based on the observation in both Figure 26 and Figure 27, S c A N N is better, as described in Equation (26).

8. Conclusions and Future Research

The material presented in this paper shows a significant improvement in the accuracy of an ANN in classification problems with the aid of principal component analysis–sample reduction process (PCA–SRP). The ANN cast off 10% of the learning rate, two layers with 32 and 16 hidden neurons, ReLU activators in hidden layers, and SoftMax in output activators with 100 epochs or iterations, as shown in Table 3 based on the PCA–SRP and ANN The Python implementation program was implemented on the multidimensional datasets gathered, namely the heart disease, gender voice recognition, breast cancer classification, and cancer patients datasets provided in Table 4. These datasets were then used in the PCA–SRP + ANN accuracy comparison testing, sensitivity vs. specificity testing, receiver operating characteristic (ROC) curve testing, and accuracy vs. additional random samples testing; the results show significant improvements.
PCA–SRP removed dirty and imprecise datasets, based on the results shown in Table 5, which allowed us to reduce the number of samples in the process and allowed for a significant increase in accuracy, as shown in Table 6 and Table 7. Furthermore, we also determined the recommended S c range values for normal cleaning and the ANN classification problem.
Future research will further investigate the performance in massive biomedical datasets and determine how to load them into PCA–SRP cleansing agents; one of the suggestions is loading through batch processing. Another suggestion is to utilise knowledge-based techniques of PCA–SRP in different neural network architectures such as CNNs, RNNs, LSTMs, and GNNs. Furthermore, an investigation into various field applications to explore the incorporation of the investigated cleaning techniques, such as real-time biomedical automation, image-based medical diagnosis classification, and human thoughts processes [48,49,50,51] could be a desirable research avenue.
The proposed methodology can be applied to a wearable EEG or similar device in order to extract chaotic data from the brain’s unique biometric featured samples for use in cryptography or especially for steganography. Lastly, this formal basis is necessary to design and construct high-quality and helpful software tools to support the data cleansing process of PCA–SRP and its application for artificial neural networks (ANNs).

Author Contributions

Conceptualisation, C.M.S.A., T.Y.W. and H.C.; methodology, C.M.S.A., T.Y.W. and H.C.; software, C.M.S.A.; validation, H.C., T.Y.W. and S.A.-M.; formal analysis, C.M.S.A. and H.C.; investigation, C.M.S.A., T.Y.W. and H.C.; resources, H.C., T.Y.W. and S.A.-M.; data curation, C.M.S.A.; writing—original draft preparation, C.M.S.A.; writing—review and editing, C.M.S.A., T.Y.W. and H.C.; visualisation, C.M.S.A., T.Y.W. and H.C.; supervision, T.Y.W. and H.C.; project administration, C.M.S.A. and H.C.; funding acquisition, C.M.S.A. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

The publication of this article was funded by the Department of Science and Technology—Science Education Institute (DOST-SEI), Republic of the Philippines, through the Foreign Graduate Scholarship.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

The sample dataset is publicly available, but specific invisible health data will not be provided as consent was not given.

Data Availability Statement

The data presented in this study are available from the corresponding author upon request.

Acknowledgments

The authors thank Christofer Yalung for their encouragement of the first author to perform research and for their proof reading of the English spelling and grammar in our research paper. Furthermore, we thank the faculty members of the University of Gloucestershire who supported the authors in their research.

Conflicts of Interest

The authors declare no conflict of interest.

Sample Availability

The sample datasets are available from the given URL in the references.

Abbreviations

The following abbreviations are used in this manuscript:
AIArtificial Intelligence
ANNArtificial Neural Network
AUCArea Under Curve
CNNConvolutional Neural Network
GNNGraph Neural Network
LSTMLong Short-Term Memory
PCAPrincipal Component Analysis
ReLURectified Linear Unit
RNNRecurrent Neural Network
ROCReceiver Operating Characteristic
SCAREScalable Automatic Repairing
SRPSample Reduction Process

References

  1. Engin, M.; Demirel, A.; Engin, E.Z.; Fedakar, M. Recent developments and trends in biomedical sensors. Measurement 2005, 37, 173–188. [Google Scholar] [CrossRef]
  2. Hannan, M.A.; Abbas, S.M.; Samad, S.A.; Hussain, A. Modulation Techniques for Biomedical Implanted Devices and Their Challenges. Sensors 2012, 12, 297. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  3. Gholami, M.; Cai, N.; Brennan, R. An artificial neural network approach to the problem of wireless sensors network localization. Robot. -Comput.-Integr. Manuf. 2013, 29, 96–109. [Google Scholar] [CrossRef]
  4. Cheng, L.; Guan, Y.; Zhu, K.; Li, Y. Recognition of human activities using machine learning methods with wearable sensors. In Proceedings of the 2017 IEEE 7th Annual Computing and Communication Workshop and Conference (CCWC), Las Vegas, NV, USA, 9–11 January 2017; pp. 1–7. [Google Scholar]
  5. Barnett, M.P. The Information Explosion. Nature 1964, 203, 585–585. [Google Scholar] [CrossRef]
  6. Schwiebert, L.; Gupta, S.K.S.; Weinmann, J. Research challenges in wireless networks of biomedical sensors. In Proceedings of the 7th Annual International Conference on Mobile Computing and Networking, Rome, Italy, 16 July 2001; pp. 151–165. [Google Scholar]
  7. Fung, B.C.; Wang, K.; Yu, P.S. Top-down specialization for information and privacy preservation. In Proceedings of the 21st International Conference on Data Engineering (ICDE’05), Tokyo, Japan, 5–8 April 2005; pp. 205–216. [Google Scholar]
  8. Zhang, L.; Wu, X. Recent Progress in Challenges of Wireless Biomedical Sensor Network. In Proceedings of the 2009 3rd International Conference on Bioinformatics and Biomedical Engineering, Beijing, China, 11–13 June 2009; pp. 1–4. [Google Scholar]
  9. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  10. Geiger, B.C.; Kubin, G. Relative information loss in the PCA. In Proceedings of the 2012 IEEE Information Theory Workshop, Lausanne, Switzerland, 3–7 September 2012; pp. 562–566. [Google Scholar]
  11. Rahm, E.; Do, H.H. Data cleaning: Problems and current approaches. IEEE Data Eng. Bull. 2000, 23, 3–13. [Google Scholar]
  12. Ridzuan, F.; Wan Zainon, W.M.N. A Review on Data Cleansing Methods for Big Data. Procedia Comput. Sci. 2019, 161, 731–738. [Google Scholar] [CrossRef]
  13. Wang, H.; Li, M.; Bu, Y.; Li, J.; Gao, H.; Zhang, J. Cleanix: A big data cleaning parfait. In Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, Shanghai, China, 3–7 November 2014; pp. 2024–2026. [Google Scholar]
  14. Yakout, M.; Berti-Équille, L.; Elmagarmid, A.K. Don’t be scared: Use scalable automatic repairing with maximal likelihood and bounded changes. In Proceedings of the 2013 ACM SIGMOD International Conference on Management of Data, New York, NY, USA, 22–27 June 2013; pp. 553–564. [Google Scholar]
  15. Chu, X.; Morcos, J.; Ilyas, I.F.; Ouzzani, M.; Papotti, P.; Tang, N.; Ye, Y. KATARA: A Data Cleaning System Powered by Knowledge Bases and Crowdsourcing. In Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, Melbourne, VIC, Australia, 31 May–4 June 2015; pp. 1247–1261. [Google Scholar]
  16. Khayyat, Z.; Ilyas, I.F.; Jindal, A.; Madden, S.; Ouzzani, M.; Papotti, P.; Quiané-Ruiz, J.-A.; Tang, N.; Yin, S. Bigdansing: A system for big data cleansing. In Proceedings of the 2015 ACM SIGMOD International Conference on Management of Data, Melbourne, VIC, Australia, 31 May–4 June 2015; pp. 1215–1230. [Google Scholar]
  17. Chu, X.; Ilyas, I.F.; Krishnan, S.; Wang, J. Data Cleaning: Overview and Emerging Challenges. In Proceedings of the Proceedings of the 2016 International Conference on Management of Data, San Francisco, CA, USA, 26 June–1 July 2016; pp. 2201–2206. [Google Scholar]
  18. Badri, L. Development of neural networks for noise reduction. Int. Arab J. Inf. Technol. 2010, 7, 289–294. [Google Scholar]
  19. Kalantari, N.K.; Bako, S.; Sen, P. A machine learning approach for filtering Monte Carlo noise. ACM Trans. Graph. 2015, 34, 122–1. [Google Scholar] [CrossRef]
  20. Kim, K.H.; Kim, S.J. Neural spike sorting under nearly 0-dB signal-to-noise ratio using nonlinear energy operator and artificial neural-network classifier. IEEE Trans. Biomed. Eng. 2000, 47, 1406–1411. [Google Scholar]
  21. Maletic, J.I.; Marcus, A. Data Cleansing: Beyond Integrity Analysis; Pennsylvania State University: University Park, PA, USA, 2000; pp. 200–209. [Google Scholar]
  22. Jilani, T.A.; Yasin, H.; Yasin, M.M. PCA-ANN for classification of Hepatitis-C patients. Int. J. Comput. Appl. 2011, 14, 1–6. [Google Scholar] [CrossRef]
  23. Deshpande, N.T.; Ravishankar, S. Face Detection and Recognition using Viola-Jones algorithm and Fusion of PCA and ANN. Adv. Comput. Sci. Technol. 2017, 10, 1173–1189. [Google Scholar]
  24. Uğuz, H. A biomedical system based on artificial neural network and principal component analysis for diagnosis of the heart valve diseases. J. Med Syst. 2012, 36, 61–72. [Google Scholar] [CrossRef]
  25. Narendra, N.; Alku, P. Dysarthric speech classification from coded telephone speech using glottal features. Speech Commun. 2019, 110, 47–55. [Google Scholar] [CrossRef]
  26. Moreira, L.P.; Silveira, L.; Pacheco, M.T.T.; da Silva, A.G.; Rocco, D.D.F.M. Detecting urine metabolites related to training performance in swimming athletes by means of Raman spectroscopy and principal component analysis. J. Photochem. Photobiol. B Biol. 2018, 185, 223–234. [Google Scholar] [CrossRef] [PubMed]
  27. Benhar, H.; Idri, A.; Fernández-Alemán, J. Data preprocessing for heart disease classification: A systematic literature review. Comput. Methods Programs Biomed. 2020, 195, 105635. [Google Scholar] [CrossRef]
  28. Kara, S.; Güven, A.; İçer, S. Classification of macular and optic nerve disease by principal component analysis. Comput. Biol. Med. 2007, 37, 836–841. [Google Scholar] [CrossRef]
  29. Krisciukaitis, A.; Tamosiunas, M.; Jakuska, P.; Veteikis, R.; Lekas, R.; Saferis, V.; Benetis, R. Evaluation of ischemic injury of the cardiac tissue by using the principal component analysis of an epicardial electrogram. Comput. Methods Programs Biomed 2006, 82, 121–129. [Google Scholar] [CrossRef]
  30. Kumar, V.; Sachdeva, J.; Gupta, I.; Khandelwal, N.; Ahuja, C.K. Classification of brain tumors using PCA-ANN. In Proceedings of the 2011 World Congress on Information and Communication Technologies, Mumbai, India, 11–14 December 2011; pp. 1079–1083. [Google Scholar]
  31. Walse, K.H.; Dharaskar, R.V.; Thakare, V.M. PCA Based Optimal ANN Classifiers for Human Activity Recognition Using Mobile Sensors Data. In Proceedings of the First International Conference on Information and Communication Technology for Intelligent Systems, Ahmedabad, India, 28–29 November 2015; Volume 1, pp. 429–436. [Google Scholar]
  32. Landi, A.; Piaggi, P.; Pioggia, G. Backpropagation-based non linear pca for biomedical applications. In Proceedings of the 2009 Ninth International Conference on Intelligent Systems Design and Applications, Pisa, Italy, 30 November–2 December 2009; pp. 635–640. [Google Scholar]
  33. Mostaar, A.; Sattari, M.; Hosseini, S.; Deevband, M. Use of Artificial Neural Networks and PCA to Predict Results of Infertility Treatment in the ICSI Method. J. Biomed. Phys. Eng. 2019, 9, 679. [Google Scholar] [CrossRef]
  34. Mehta, D.; Verma, C. Prediction of cancer diagnosis patients from fine-needle aspirates using machine learning. In Proceedings of the International Conference on Intelligent Computing and Smart Communication, Tehri, India, 20–21 April 2019; pp. 337–348. [Google Scholar]
  35. Chen, Y.; Hao, Y. Integrating principle component analysis and weighted support vector machine for stock trading signals prediction. Neurocomputing 2018, 321, 381–402. [Google Scholar] [CrossRef]
  36. Zhong, X.; Enke, D. Predicting the daily return direction of the stock market using hybrid machine learning algorithms. Financ. Innov. 2019, 5, 1–20. [Google Scholar] [CrossRef]
  37. Adisa, J.A.; Ojo, S.O.; Owolawi, P.A.; Pretorius, A.B. Financial Distress Prediction: Principle Component Analysis and Artificial Neural Networks. In Proceedings of the 2019 International Multidisciplinary Information Technology and Engineering Conference (IMITEC), Vanderbijlpark, South Africa, 21–22 November 2019; pp. 1–6. [Google Scholar]
  38. Tran, N.M.; Burdejová, P.; Ospienko, M.; Härdle, W.K. Principal component analysis in an asymmetric norm. J. Multivar. Anal. 2019, 171, 1–21. [Google Scholar] [CrossRef] [Green Version]
  39. Richards, J.E. Recovering dipole sources from scalp-recorded event-related-potentials using component analysis: Principal component analysis and independent component analysis. Int. J. Psychophysiol. 2004, 54, 201–220. [Google Scholar] [CrossRef]
  40. Chen, Z.; Oni, S.; Hoban, S.; Jademi, O. A Comparative Study of Data Cleaning Tools. Int. J. Data Warehous. Min. 2019, 15, 48–65. [Google Scholar] [CrossRef]
  41. Dubey, A. The Mathematics Behind Principal Component Analysis. Available online: https://tinyurl.com/PCA-Maths (accessed on 12 April 2020).
  42. Agarap, A.F. Deep learning using rectified linear units (ReLU). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  43. Janosi, A.; Steinbrunn, W.; Pfisterer, M.; Detrano, R. Heart Disease Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Heart+Disease (accessed on 1 April 2019).
  44. Becker, K. Gender Recognition by Voice. Available online: https://www.kaggle.com/primaryobjects/voicegender (accessed on 1 February 2020).
  45. Wolberg, W.H.a.S.; William, N.; Mangasarian, O.L. Breast Cancer Wisconsin (Diagnostic) Data Set. Available online: https://archive.ics.uci.edu/ml/datasets/Breast Cancer Wisconsin (Diagnostic) (accessed on 12 November 2020).
  46. Barreto, G.d.A.; Neto, A. R.; Filho, H. A. Vertebral Column Data Set. Available online: http://archive.ics.uci.edu/ml/datasets/Vertebral Column (accessed on 1 August 2019).
  47. Da Marla, R. Cancer Patients Data Set. Available online: https://data.world/cancerdatahp/lung-cancer-data/workspace (accessed on 1 March 2020).
  48. Vyas, S.; Das, S.; Ting, Y.-P. Predictive modeling and response analysis of spent catalyst bioleaching using artificial neural network. Bioresour. Technol. Rep. 2020, 9, 100389. [Google Scholar] [CrossRef]
  49. Yalung, C.N.; Adolfo, C.M.S. Analysis of obstacle detection using ultrasonic sensor. Int. Res. J. Eng. Technol. 2017, 4, 1015–1019. [Google Scholar]
  50. Tirupattur, P.; Rawat, Y.S.; Spampinato, C.; Shah, M. Thoughtviz: Visualizing human thoughts using generative adversarial network. In Proceedings of the 26th ACM international conference on Multimedia, Seoul, South Korea, 22–26 October 2018; pp. 950–958. [Google Scholar]
  51. Yalung, C.; Majeed, S.A.; Adolfo, C.M.; Karam, J.; Mirtskhulava, L. Gyroscope explorer terrain angles classification. In Proceedings of the 2016 IEEE East-West Design & Test Symposium (EWDTS), 14–17 October 2016, Yerevan, Armenia; pp. 1–5.
Figure 1. Dirty data.
Figure 1. Dirty data.
Applsci 11 08240 g001
Figure 2. Sample Excel multidimensional dataset.
Figure 2. Sample Excel multidimensional dataset.
Applsci 11 08240 g002
Figure 3. Covariance and their values.
Figure 3. Covariance and their values.
Applsci 11 08240 g003
Figure 4. Projection of data points in two dimensions.
Figure 4. Projection of data points in two dimensions.
Applsci 11 08240 g004
Figure 5. Projection of data points in the two newly constructed dimensions.
Figure 5. Projection of data points in the two newly constructed dimensions.
Applsci 11 08240 g005
Figure 6. Conceptual flowchart of the proposed approach.
Figure 6. Conceptual flowchart of the proposed approach.
Applsci 11 08240 g006
Figure 7. Detailed breakdown of the steps in PCA.
Figure 7. Detailed breakdown of the steps in PCA.
Applsci 11 08240 g007
Figure 8. Detailed breakdown of the steps in ANN.
Figure 8. Detailed breakdown of the steps in ANN.
Applsci 11 08240 g008
Figure 9. Detailed sensitivity vs. specificity testing procedure.
Figure 9. Detailed sensitivity vs. specificity testing procedure.
Applsci 11 08240 g009
Figure 10. Sc—accuracy testing representation.
Figure 10. Sc—accuracy testing representation.
Applsci 11 08240 g010
Figure 11. Pre-processing an application of multidimensional datasets in the proposed approach.
Figure 11. Pre-processing an application of multidimensional datasets in the proposed approach.
Applsci 11 08240 g011
Figure 12. Scree plot of the heart disease dataset.
Figure 12. Scree plot of the heart disease dataset.
Applsci 11 08240 g012
Figure 13. Threshold cutoff limit of the heart disease dataset.
Figure 13. Threshold cutoff limit of the heart disease dataset.
Applsci 11 08240 g013
Figure 14. Scree plot of the gender voice recognition dataset.
Figure 14. Scree plot of the gender voice recognition dataset.
Applsci 11 08240 g014
Figure 15. Threshold cutoff limit of the gender voice recognition dataset.
Figure 15. Threshold cutoff limit of the gender voice recognition dataset.
Applsci 11 08240 g015
Figure 16. Scree plot of the breast cancer classification.
Figure 16. Scree plot of the breast cancer classification.
Applsci 11 08240 g016
Figure 17. Threshold cutoff limit of the breast cancer classification dataset.
Figure 17. Threshold cutoff limit of the breast cancer classification dataset.
Applsci 11 08240 g017
Figure 18. Scree plot of the cancer patients dataset.
Figure 18. Scree plot of the cancer patients dataset.
Applsci 11 08240 g018
Figure 19. Threshold cutoff limit of the cancer patients dataset.
Figure 19. Threshold cutoff limit of the cancer patients dataset.
Applsci 11 08240 g019
Figure 20. PCA–SRP to ANN process.
Figure 20. PCA–SRP to ANN process.
Applsci 11 08240 g020
Figure 21. Validation model accuracy of the given datasets.
Figure 21. Validation model accuracy of the given datasets.
Applsci 11 08240 g021
Figure 22. Sensitivity vs. specificity diagram.
Figure 22. Sensitivity vs. specificity diagram.
Applsci 11 08240 g022
Figure 23. Receiver operating characteristic (ROC) curve.
Figure 23. Receiver operating characteristic (ROC) curve.
Applsci 11 08240 g023
Figure 24. ROC curve interpretation.
Figure 24. ROC curve interpretation.
Applsci 11 08240 g024
Figure 25. PCA cleaned data percentage (%).
Figure 25. PCA cleaned data percentage (%).
Applsci 11 08240 g025
Figure 26. +%Noise response for cancer patients dataset (Sc = 88%).
Figure 26. +%Noise response for cancer patients dataset (Sc = 88%).
Applsci 11 08240 g026
Figure 27. +%Noise response for the cancer patients dataset (Sc = 98%).
Figure 27. +%Noise response for the cancer patients dataset (Sc = 98%).
Applsci 11 08240 g027
Table 1. Data-cleansing methods for big data.
Table 1. Data-cleansing methods for big data.
MethodsKey FeaturesApproach
CleanixScalability, unification,Rule Selection
and Usability
SCAREScalabilityMachine Learning
Technique
KATARAEasy Specification,Knowledge-base and
Pattern ValidationCrowdsourcing
Data Annotation
BigDansingEfficiency, ScalabilityRule Specification
and ease of use
Table 2. Currently used data-cleaning technique.
Table 2. Currently used data-cleaning technique.
VolumeVeracityVelocity
Cleaning Technique DataScalabilityNo need for Extensive Data Quality Rule OptimisationNo need for Human Domain Expert
Cleanixx
SCAREx x
KATARA xx
BigDansingx x
Table 3. PCA–SRP and ANN parameters.
Table 3. PCA–SRP and ANN parameters.
PCA–SRP and ANN ParametersValues
Most significant> S c × ( P C A S R P ( M ) h i g h e s t )
sample size
Learning rate0.1%
Epochs100
Number of ANN2 hidden layers (32 and
neurons16, respectively)
Activation functionsReLU in hidden layers
usedSoftMax in final layer
Table 4. Datasets used and their metadata.
Table 4. Datasets used and their metadata.
DatasetsNumber of DimensionsSample SizeTarget Classification
Heart Disease143012
Gender Voice Recognition2131672
Breast Cancer315682
Cancer Patients2410983
Table 5. Datasets sample status after PCA–sample reduction process.
Table 5. Datasets sample status after PCA–sample reduction process.
DatasetSamples Used%Passrate
Heart Disease259/33078.48
Gender Voice Recognition786/330023.82
Breast Cancer562/60093.67
Cancer Patients499/109945.41
Table 6. A ccuracy of datasets with noisy samples.
Table 6. A ccuracy of datasets with noisy samples.
DatasetANNANN%
Onlywith SRPIncrease
Heart Disease76.1979.893.7
Gender Voice Recognition90.693.292.69
Breast Cancer90.4992.361.86
Cancer Patients79.8283.273.45
Table 7. Accuracy of datasets with random samples.
Table 7. Accuracy of datasets with random samples.
DatasetANNANN%
Onlywith SRPIncrease
Heart Disease73.6981.768.07
Gender Voice Recognition85.7291.25.42
Breast Cancer83.386.463.12
Cancer Patients80.9088.817.91
Table 8. S a c c u r a c y ( S c t e s t m i n ) result.
Table 8. S a c c u r a c y ( S c t e s t m i n ) result.
Dataset Sc min SensitivitySpecificity Sc test SensitivitySpecificity
Heart Disease9698.0473.689794.1581.04
Gender Voice8088.752.879053.0884.103
Recognition
Breast Cancer9699.4757.649994.5181.84
Cancer Patients8893.8325.869581.839.1
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Adolfo, C.M.S.; Chizari, H.; Win, T.Y.; Al-Majeed, S. Sample Reduction for Physiological Data Analysis Using Principal Component Analysis in Artificial Neural Network. Appl. Sci. 2021, 11, 8240. https://doi.org/10.3390/app11178240

AMA Style

Adolfo CMS, Chizari H, Win TY, Al-Majeed S. Sample Reduction for Physiological Data Analysis Using Principal Component Analysis in Artificial Neural Network. Applied Sciences. 2021; 11(17):8240. https://doi.org/10.3390/app11178240

Chicago/Turabian Style

Adolfo, Cid Mathew Santiago, Hassan Chizari, Thu Yein Win, and Salah Al-Majeed. 2021. "Sample Reduction for Physiological Data Analysis Using Principal Component Analysis in Artificial Neural Network" Applied Sciences 11, no. 17: 8240. https://doi.org/10.3390/app11178240

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop