Next Article in Journal
Directional Differences in Thematic Maps of Soil Chemical Attributes with Geometric Anisotropy
Previous Article in Journal
Precise Tensor Product Smoothing via Spectral Splines
Previous Article in Special Issue
Self-Organizing Topological Multilayer Perceptron: A Hybrid Method to Improve the Forecasting of Extreme Pollution Values
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the (Apparently) Paradoxical Role of Noise in the Recognition of Signal Character of Minor Principal Components

by
Alessandro Giuliani
1,* and
Alessandro Vici
2
1
Environment and Health Department, Istituto Superiore di Sanità, Viale Regina Elena 299, 00161 Rome, Italy
2
Department of Oncology and Molecular Medicine, Istituto Superiore di Sanità, Viale Regina Elena 299, 00161 Rome, Italy
*
Author to whom correspondence should be addressed.
Stats 2024, 7(1), 54-64; https://doi.org/10.3390/stats7010004
Submission received: 14 December 2023 / Revised: 7 January 2024 / Accepted: 9 January 2024 / Published: 11 January 2024
(This article belongs to the Special Issue Statistical Learning for High-Dimensional Data)

Abstract

:
The usual method of separating signal and noise principal components on the sole basis of their eigenvalues has evident drawbacks when semantically relevant information ‘hides’ in minor components, explaining a very small part of the total variance. This situation is common in biomedical experimentation when PCA is used for hypothesis generation: the multi-scale character of biological regulation typically generates a main mode explaining the major part of variance (size component), squashing potentially interesting (shape) components into the noise floor. These minor components should be erroneously discarded as noisy by the usual selection methods. Here, we propose a computational method, tailored for the chemical concept of ‘titration’, allowing for the unsupervised recognition of the potential signal character of minor components by the analysis of the presence of a negative linear relation between added noise and component invariance.

1. Introduction

1.1. The Peculiar Role of Principal Component Analysis in Life Sciences

Principal component analysis (PCA) is the most-widespread statistical method to deal with multidimensional data sets coming from biomedical research. The huge success of PCA is somewhat related to its location in a twilight zone between general-purpose statistical techniques and quantitative models developed for specific applications [1,2]. Even if any statistical method (e.g., Analysis of Variance) implies some general hypothesis on the analyzed phenomenon (e.g., linearity), these hypothesis pertain to a meta-level largely decoupled from the nature of the analyzed system. On the contrary, other approaches (e.g., the Renormalization Group, a formal apparatus that allows for systematic investigation of the changes of a physical system as viewed at different scales [3]) are linked to a specific field of investigation and allow only for a ‘metaphorical’ extension to other science areas [4]. PCA blurs the distinction between ‘content-agnostic’ and ‘content-derived’ methods: while being a general-purpose statistical technique, PCA allows generating a statistical mechanics theoretical frame for biological systems without the need of strong a priori theoretical assumptions [2]. This is made possible by the analogy of PCA to the two main pillars of systems approaches in biomedicine: complex networks and order/organization quantification. In the first case, the analogy stems from the equivalence between a network adjacency matrix (with nodes corresponding to variables and edges to their pairwise correlations) and a correlation (or covariance) matrix [5]. In the second case, the analogy builds upon the immediate translation of the eigenvalues’ distribution in terms of the order and the organization of the system at hand [6,7,8,9]. In both cases, PCA acts as a hypothesis-generation tool; thus, it is of utmost importance to assign a biologically relevant meaning to the extracted components. This feature contrasts with hypothesis testing approaches (e.g., Analysis of Variance, Student’s t-test), where the biological meaning is set before the statistical analysis.
In the following, we will concentrate on a proposal for recognizing biologically relevant signals ‘hidden’ in very minor (in terms of the amount of explained variance) noise, a very common situation encountered in biological experimentation [10].
It is important to stress that, not only in biology, the interpretation of component meaning is of crucial importance: Vilenchik and colleagues in their paper [11] put the problem in a very straightforward way, ‘To interpret or not to interpret PCA? This is our question’. The paper [11] dealt with the features of social networks coming from different platforms (e.g., Linkedin). The authors collected between 11 and 15 features per network that describe the user’s activity in the network and the feedback that a user receives from other users. They constructed a ‘interpretability score’ based on the intuitive idea that the component most easily interpretable is the one in which only one feature (original variable) has a relevant (non-zero in the paper, which sets a threshold for loading relevance) loading. On the contrary, the presence of non-zero loadings for all the variables points to a very low interpretability score. This approach resembles the classical Shannon definition of information entropy, low interpretability scores being strictly related to the high values of the eigenvalues’ distribution entropy, which in turn was demonstrated [9] to be the one of the most-efficient complexity estimators.

1.2. Why in Biomedical Sciences the Relevant Information Often Hides in Minor Components

When applying PCA in a content-agnostic mode (i.e., without the need for interpreting the meaning of extracted components) as in de-noising or dimensionality reduction for feature extraction applications [12,13], the choice of p < n (with n = the number of original variables), the number of components to be retained for further analysis, relies on their relative proportion of explained variance. When the focus is on the biological interpretation of extracted components, relying exclusively on the relative amount of variance explained as the selection criterion can be highly misleading. The order of selection by the eigenvalue deals with the concept of the ’noise floor’, i.e., the set of minor components corresponding to extremely low and almost identical proportions of explained variance [14] that are dominated by stochasticity and, consequently, do not carry relevant information about the system at hand [5,14]. While the above statement is surely correct, the peculiar character of biological systems often generates a different situation in which the most-important thing is to investigate the information carried by specific minor components making part of the noise floor. One of the most-common situations in the life sciences is the presence of a dominant first component carrying semantically irrelevant information together with a set of minor components, where noise and weak (but semantically relevant) signals are mixed. An exemplary case is gene expression data sets coming from the microarray methodology [7,10]. These are extremely high-dimensionality data sets made of thousands of different gene expression values, in which the first component is related to the tissue of origin [7]. This is a natural consequence of the need for each specific tissue to maintain a peculiar balance among different gene expressions to carry out its physiological role, in other words, the dominant component neatly emerging from the noise floor, in terms of the explained variance, conveying semantically irrelevant information, corresponding to the kind of analyzed biological material that the scientist knows in advance.
In general, the first component is the component ecologists and morphologists traditionally define as ‘size’ as opposed to ‘shape’ components [15]. A ‘size’ component shows loadings of the same sign for all the variables. This is because of the presence of a major order parameter (typically the size of an organism), influencing almost equally all the studied variables. In this case, the relevant information resides in the minor components, in which the presence of both negative and positive loadings corresponds to changes in the ‘shape’ (in the wide sense of the individual profile across the different descriptors) of the statistical units. The most-direct way to discriminate between a weak signal (a semantically informative shape component) and a ‘pure noise’ component is through a supervised approach: a weak signal component, at odds with a pure noise one, has a significant correlation with an external variable of interest (Roden).
In the presence of a relevant correlation between a minor component and an external variable of interest, it is straightforward to sketch a mechanistic hypothesis by the analysis of both component loadings and scores [10]. What can we do if we have no external variable acting as the probe? This is not pure imagination: in many situations (e.g., neuroimages [16]), we do not have any external reference to rely upon, but, despite that, it is of crucial importance to determine the signal character of a comparatively minor detail.
The proposed method faces the problem of determining the signal character of minor components when in the absence of external covariates, relying upon the different behavior of the signal and noise when the original data matrix is corrupted by increasing amounts of added noise.

2. Methods

Noise–Signal Discrimination by Noise Titration

In [17], the authors described a toy model of a situation characterized by a semantically irrelevant major component, two minor (weak, but informative) components, and two pure noisy (measurement error) components. This configuration stems from the aforethought choice of a poor measurement frame: 33 Europeans cities were described by their distances (manually estimated on a geographic map at a 1 : 3 × 10 6 scale by a ruler, with a 1 mm precision corresponding to 3 km) from the five main towns of Latium (the central Italian region around Rome). The aim of the study was to reconstruct the mutual positions of the European towns. The fallacy of the measurement frame came from the very low size of Latium with respect to Europe (the ratio between Latium and the European areas is equal to 1.65 × 10 4 ). While the choice of reference points covering a space of the same-order-of-magnitude data set range of variation could guarantee a precise reconstruction of the statistical units’ configuration [18], the reference frame of [17] generated a strongly biased configuration. PCA was applied to the data set spanned by the distances from the five Latium towns, generating a five-component solution with the following distribution of explained variance proportion (Table 1).
PC1 was a size component with all-positive and near-unity loading, corresponding to the distances of the 33 European towns from the geographic center of Latium (approximately coincident with Rome). PC1, even if almost entirely explaining the data set’s variance, carries useless information for reconstructing the data set’s configuration: the circular symmetry of the distance operator makes cities like Barcelona and Athens have very similar PC1 scores even if they are very far apart (Barcelona and Athens being at the opposite west and east sides with respect to Latium). Both PC2 and PC3 are shape components with both (near-zero) positive and negative loadings with the five original variables. Adopting a supervised approach, the authors demonstrated that these two components, despite their very low eigenvalues, allow reconstructing the Europe cities’ spatial configuration: the bearing between European cities and Latium was predicted by a linear combination of the PC2 and PC3 scores with a Pearson r = 0.97 ( p < 0.0001 ) [17].
On the contrary, PC4 and PC5 were pure noise components, derived from the measurement error, with no relation to any geographical information.
The eigenvalue distribution of this geographical data set was coincident with many biological data sets in which a dominant (but semantically irrelevant) signal ‘squashes’ on the noise floor the semantically relevant, but weak signals.
In the above-described paper, the authors, having demonstrated the informative character of PC2 and PC3 by a supervised approach (correlation with an external covariate), gave a proof of concept of the suitability for the identification of very weak signals, using an unsupervised approach based on a ‘noise titration’ procedure [19]. This implies the generation of n ‘dirty’ copies of the original data matrix, adding to each variable a zero-mean Gaussian noise with different standard deviation values. As noise levels increase (the signal-to-noise ratio (SNR) decreases), the weak signal components, extracted from the noise-corrupted data sets, showed a linear negative relation between the amount of added noise and their Pearson correlation with the corresponding components of the original data set. On the contrary, the pure noise (PC4, PC5) components did not show any significant relation between the amount of added noise and their Pearson correlation with the corresponding components of the original data set.
Poon and Barahona [19] explicitly set the metaphor of ‘noise titration’ in the framework of deterministic chaos/stochasticity discrimination. The authors equated time series data to acid–base solutions with deterministic chaos being most ‘acidic’ and white noise most ‘alkaline’. This metaphor implies the consideration of deterministic chaos as a ‘buffer’, i.e., a ‘weak acid’ able to neutralize the effect of the added alkaline substance (noise) with no p H change of the solution until a given volume (called the volume of titration) is reached. At this point, the pH of the solution initiates a dramatic rise.
An essential component in chemical titration is a sensitive indicator that specifically reveals the changes in pH around the equivalence point of acid–base neutralization. The indicator for the numerical titration is not specified by the authors, who stated, ‘…could in theory be any noise-tolerant technique that can reliably detect nonlinear dynamics in short noisy series’ [19].
Our proposal, even if tailored for the Poon and Barahona metaphor, is more general. We adopted a more-fundamental (not necessarily linked to buffering) definition of titration as a method to determine the concentration of any substance (analyte) in a solution. A reagent (a substance known to react with the analyte) is added at different (known) volumes to the solution containing the analyte. The volume of added reagent corresponding to reaching a plateau of the chemical reaction is termed the ‘titration volume’ and corresponds to the volume of the analyte present in the solution (and consequently, to its concentration). We did not focus on the reaching of a ‘neutralization point’ nor on a quantification of the relative amount of ‘determinism’ and ‘stochasticity’, but limited ourselves to the recognition of the most-promising ‘signal candidates’ among the noise floor components. We hypothesized that a weak signal component should show a strong negative relation between added noise and the Pearson correlation between its original and noisy versions. This relation is expected to be less evident in ‘pure noise’ components. This mirrors the results reported in [17], which, in a titration framework, correspond to the progressive approaching of the end of the ‘reaction’ between the signal and added noise when the analyte (signal) is exhausted by the reaction with the increasing amount of the reactant (added noise). Given that noise cannot react with itself, the ‘reaction’ ends at an amount of added noise corresponding to the erasure of any residual correlation structure, and the relation between the amount of added noise and a suitable ‘signal indicator’ (i.e., the correlation between the initial signal and its noise-corrupted counterpart) vanishes. The main difficulty for an immediate generalization of the case described in [17] is that PC4 and PC5 are pure material noise deriving from actual measurement and are neatly separated by weak signal components in terms of the proportion of explained variance (see Table 1), thus greatly enhancing the discrimination power of the procedure. Moreover, the clear-cut separation in terms of explained variance between the weak signal and pure noise components keeps the ordering of the components across noise-corrupted replicas invariant, so allowing for an immediate recognition of the corresponding components among the different data sets.
Thus, to develop a selection procedure able to identify putative weak signals endowed with a wide application spectrum, we designed a simulation framework making this discrimination as difficult as possible. In the following, we will describe the subsequent steps of the suggested procedure, explaining the motivations for each step:
  • Using a relatively short series:
    The first step of our simulation is the generation of a time series ( S 0 ) by a Gaussian distribution at zero mean and unit standard deviation N ( 0 , 1 ) . This series, by the action of a 15-dimensional embedding procedure at l a g = 1 [14], gives rise to a multivariate matrix ( A 0 ) with 15 variables and 86 statistical units. The relatively low number of statistical units is consistent with the numerosity of a great part of biological experimentation.
  • Presence of correlation structures in the extracted components:
    The character of PCA as a filter for correlated information makes the component scores relative to the A 0 matrix show a certain amount of internal correlation, so we do not have any ‘pure noise’ component. In the second step of the procedure, we added to the original series an extremely weak signal (corresponding to a ’saw-wave’), composed of alternating 0.1 and −0.1 values. The extremely low power of the added signal makes the resulting series ( S 1 ) practically identical to S 0 (Pearson r = 0.994 ; see Section 3) and have a largely superimposable eigenvalue distribution of the original ( A 0 ) and signal-added ( A 1 ) embedding matrices (see Section 3). Looking at component loading matrix, the 14th component, well inside the noise-floor, keeps trace of the square-wave signal. This component is the ‘analyte’ we expect to ‘react’ with the added noise.
  • Adding noise:
    Having checked the superposition between the S 0 and S 1 series and the consequent coincidence between the eigenvalue distribution of the relative A 0 and A 1 15-dimensional embedding matrices, as the third step, we generated 20 noise-corrupted copies of S 1 . These contaminated series are named C 1 C 20 (according to the increasing amount of added zero-mean Gaussian noise), from a minimal σ of 0.05 ( C 1 ) to a maximum of 1 ( C 20 ), with each copy differing 0.05 σ units from the previous one.
  • Titration:
    As the fourth step, the 15-dimensional embedding matrices relative to the C 1 C 20 series are analyzed by PCA and the Pearson correlation between the PC6 and PC15 (noise floor) component scores relative to the A 1 matrix and each corresponding component relative to each of the 20 embedding matrices’ noise-added series are computed. It is worth noting that, due to their almost identical eigenvalues, the PC6–PC15 ordering varies both across noise-corrupted data sets and with respect to the original A 1 matrix. Thus, for any value of added noise, we picked up the component having the higher correlation with the original one as the ‘corresponding component’, independent of its relative order of explained variance.
  • Recognition of the weak signal:
    The fifth and last step of the procedure is to check if the ‘weak signal’ component (PC14 in the original A 1 matrix) shows a significantly higher R 2 between the original corrupted versions of the PC14 (weak signal) correlation and the amount of added noise with respect to the other minor components. It is worth noting that the recognition of the weak signal relies on the expected effect of added noise in decreasing the correlation between the original and noise-contaminated versions of the component (the Pearson r between added noise and original noise-added components are all negative).
    All the analyses were performed in R using the ‘rnorm’ function for randomly generating data from a normal distribution and the ‘princomp’ for PCA. Both functions belong to the ‘stats’ package.

3. Results and Discussion

The component loadings and eigenvalue distribution of the A 1 (15-dimensional embedding of the S 1 series) matrix is reported in Table 2.
The eigenvalue distribution of the embedding matrix of the series generated by N ( 0 , 1 ) , by the very weak added signal, is not completely flat (as in the ideal case of pure noise), but shows a slow decay from the 10.1% explained variance of PC1 to the 3.1% of PC15. This result is consistent with the character of the correlation filter of PCA, exalting any correlation structure present in the data. The plot reported in Figure 1 suggests a five-component solution as the bona fide signal according to the LEV criterion [20], setting the signal/noise border at the reaching of a linear relation between the logarithm of the variance explained and the number of components.
The saw-wave signal (alternating 0.1/−0.1 values) (see Section 2) does not correspond to the ‘signal components’ as selected by the LEV criterion, but is embedded in PC14, well inside the noise floor. PC14’s loadings show a clear resemblance to the added saw-like alternating weak signal with alternating opposite sign loadings (the only exceptions being T7, T8, and T9, which in turn have the lowest absolute loading values on PC14).
The addition of a saw-wave of alternating 0.1/−0.1 values (see Section 2) to S 0 did not alter the original signal (original series = S 0 ; original series + low signal = S 1 ), the original and signal-added series showing a Pearson r = 0.994 (Figure 2). Consequently, the A 0 and A 1 matrices are practically identical.
The 20 noise-contaminated replicas of the S 1 series (with the amount of Gaussian added noise going from σ = 0.1 to σ = 1) were transformed into 20 multivariate matrices ( C 1 C 20 ) by a 15-dimensional unit lag embedding and analyzed by PCA (see Section 2). The Pearson correlation coefficient between the corresponding noise floor components of the original ( A 1 ) and noise-corrupted ( C 1 C 20 ) matrices were computed. As previously underlined in the Methods Section, the fact that PC6–PC15 constitute the noise floor implies that their eigenvalues are almost coincident. This, in turn, provokes the alteration of their ordering in the noise-added C 1 C 20 matrices. Thus, for any value of added noise, we picked the component having the higher correlation with the original one as the ‘corresponding components’, independent of the relative order of explained variance. The last step is the computations of the R 2 between the amount of added noise and the PC6–PC15 original/noisy pairs. The entire procedure going from the generation of noise-corrupted series and the R 2 of the relation between the amount of added noise and the correlation between the original/noisy component pairs was iterated 100 times starting from different seeds. The R 2 relative to each iteration were eventually averaged, obtaining the following values for PC6–PC15 (Table 3).
As expected, PC14, mirroring the added weak signal, had a higher R 2 despite its extremely low eigenvalue. The average R 2 for PC1–PC16 was 0.58, with a standard deviation equal to 0.13, and the confidence interval at 99% was [0.45–0.72], thus pointing to a statistically significant departure of PC14 from the other minor components. This result, far from being a definitive proof of a general methodology to select putative signal components in a unsupervised way, seems to be a promising avenue to follow, to complement the more-reliable supervised approach based on the correlation with external variables [10].

4. Conclusions

The titration metaphor predicts the end of the ‘reaction’ when the analyte can no longer react with the added reactants: the stopping of the reaction comes from the total conversion of the analyte into the reaction product. In the case of noise titration, this implies that the relationship between the correlation of the corrupted signal with its original version arrives at a bottom end of near-zero correlation. Figure 3 depicts this condition.
The above result tells us that, analogous to chemical titration, even in the case of computational titration, it is of utmost importance to carefully select both the range of added noise and the dosage schedule. It is worth underlining that our procedure relies on the classical effects of noise in degrading self-correlated signals that (as demonstrated in [17]) do not need to be time series or to be generated by a specific function.
To go beyond a purely empirical demonstration of the proposed methodology, we can refer to [21]. Formally, we wish to choose between the two hypothesis:
H 0 : x [ n ] = w [ n ] H 1 : x [ n ] = A + w [ n ]
for n = 0 , , N 1 (Equation (1)). We have that x is our data, A is the known signal (saw signal), and w is the Gaussian noise. In this case, the hypothesis H 0 corresponds to the situation where the data are just noise, while the hypothesis H 1 is the signal plus the Gaussian noise. A reasonable approach might be to average the samples and compare the value obtained to a threshold γ , so that we could decide H 1 if
T ( x ) = 1 N n = 0 N 1 x [ n ] > γ x
where T ( x ) is the statistic test. If w [ n ] is a Gaussian mixture, the optimal likelihood ratio test (LRT) is used to decide if H 1 is true. With a fixed probability of false alarm (level of significance α ) and, hence, threshold γ , it will produce the maximum probability of detection P D (power of the test). We can use another approach to find a better value for P D . We assumed a realization of white Gaussian noise (WGN) with variance σ 2 , which is independent of w [ n ] . Calling this u [ n ] , we have y [ n ] = x [ n ] + u [ n ] . It follows that we decide for H 1 if
T ( y ) = 1 N n = 0 N 1 y [ n ] > γ y
In [21], it was proven that this approach provides a higher value for P D with respect to the statistical test T ( x ) . This result can be easily generalized to R 2 .
The crucial issue of the paper of Kay [21] was the demonstration of the additive character of the noise ( w [ n ] ) with respect to the signal (A) evident in Equation (1). This corresponds to the titration metaphor in which the analyte (A) reacts with a reactant that corresponds to a different chemical ( w [ n ] ) added to the solution.
A last remark is related to the use of R 2 between the original and noise corrupted (hypothetical) signal instead of the ‘titration volume’ (i.e., the reaching of a flat correlation/added noise condition), like in analytical chemistry. At odds with chemistry, in which the goal of titration is the quantification of the concentration of an analyte in a solution, we are interested in a qualitative yes/no response about the signal character of a minor component. For this last task, the evidence of a reaction whose rate scales with the amount of added reactant rather than the precise amount of reactant saturating the initial analyte (volume of titration) is much more cogent. That is to say that the higher the R 2 , the higher the tenability of H 1 = A + w [ n ] corresponding to a pre-existing analyte concentration (A) independent of the added reactant ( w [ n ] ).
It is worth noting that the same fundamental distinction between the signal and noise reported in Equation (1) holds for sparse PCA [22] in which the authors, as for the estimation of the ‘real covariance matrix’, depurated by the effect of noise, explicitly stated ‘… in practice, one does not have access to the population covariance, but instead must rely on a “noisy” version of the form Σ = Σ + Δ , where Δ = Δ n denotes a random matrix perturbation’. Sparsity is frequent in biological data analysis, and in [22], the authors suggested an estimation strategy based on the search for the invariance of the PCA solution with different choices of reduced (and thus, perturbed) data sets. Sparsity is the focus of the interpretability score discussed in [11] and shares the same epistemological fundamentals of [22]: in order to interpret an emergent correlation structure (i.e., a component), we must be able to demonstrate that the component has a preferential link (e.g., correlation) with only one or at most a few variables, these being internal to the study [11] or external, as in [10]. In our work, we complemented this basic information theory concept [9] with a chemistry-inspired procedure relying on the hypothesis that a ‘true’ signal decays by its ‘reaction’ (titration) with added noise until it becomes no longer distinguishable by randomness (no preferential correlation).
As we stated before, the methodology we proposed in this paper can be of help in any situation in which we need to ascertain the signal character of a minor component in the absence of an external covariate. This implies that the proposed method could be adopted in both the feature selection and explanation procedures in machine learning [23]. Similar considerations hold in dynamical systems analysis, which, in the case of biological systems, often implies the simultaneous presence of different dynamical modes [24]. Besides the resemblance to already-established theoretical results and the wide range of applicability, what we considered the main motivation for this work was the need for statistical science to get back to its empirical roots. It is not by chance that the fathers of modern statistics (e.g., Ronald Fisher, Francis Galton) were biologists and their works were deeply inspired by the need to face peculiar biological phenomena. Here, we started from a very common situation encountered in experimental biology, i.e., the decoupling between the syntactical and semantical value of information mirrored by extremely strong signals (main component) devoid of any relevant semantic interest, going together with minor components in which semantically relevant information can hardly be distinguished from noise. We proposed an empirical solution to this problem based on a routine analytical chemistry method (titration), which holds promise to be of help in a wide range of situations.

Author Contributions

A.G. and A.V. contributed equally to the conceptualization, methodology, formal analysis, and writing—original draft preparation. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The original contributions presented in the study are included in the article. Further inquiries can be directed to the corresponding authors.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Preisendorfer, R. Principal component analysis in meteorology and oceanography. Elsevier Sci. Publ. 1988, 17, 425. [Google Scholar]
  2. Giuliani, A. The application of principal component analysis to drug discovery and biomedical data. Drug Discov. Today 2017, 22, 1069–1076. [Google Scholar] [CrossRef] [PubMed]
  3. Pelissetto, A.; Vicari, E. Critical phenomena and renormalization-group theory. Phys. Rep. 2002, 368, 549–727. [Google Scholar] [CrossRef]
  4. Giuliani, A.; Mancini, A.M.; Ghirardi, O.; Ramacci, M.T.; Voronina, T.; Sirabella, P.; Colosimo, A. Micro-and macrostructure of learning in active avoidance: A quantitative approach. Neurobiol. Learn. Mem. 1996, 65, 82–90. [Google Scholar] [CrossRef] [PubMed]
  5. David, C.C.; Jacobs, D.J. Principal component analysis: A method for determining the essential dynamics of proteins. Protein Dyn. Methods Protoc. 2014, 1084, 193–226. [Google Scholar]
  6. Gorban, A.N.; Smirnova, E.V.; Tyukina, T.A. Correlations, risk and crisis: From physiology to finance. Biophys. Rev. 2021, 389, 3193–3217. [Google Scholar]
  7. Zimatore, G.; Tsuchiya, M.; Hashimoto, M.; Kasperski, A.; Giuliani, A. Self-organization of whole-gene expression through coordinated chromatin structural transition. Phys. Stat. Mech. Its Appl. 2010, 389, 3193–3217. [Google Scholar] [CrossRef]
  8. Giuliani, A.; Vici, A. Stability/Flexibility: The tightly coupled homeostasis generator is at the same time the driver of change. Ann. Ist. Sup. San. 2023, in press. [Google Scholar]
  9. Giuliani, A.; Colafranceschi, M.; Webber, C.L., Jr.; Zbilut, J.P. A complexity score derived from principal components analysis of nonlinear order measures. Phys. Stat. Mech. Its Appl. 2001, 301, 567–588. [Google Scholar] [CrossRef]
  10. Roden, J.C.; King, B.W.; Trout, D.; Mortazavi, A.; Wold, B.J.; Hart, C.E. Mining gene expression data by interpreting principal components. BMC Bioinform. 2006, 7, 194. [Google Scholar] [CrossRef] [PubMed]
  11. Vilenchik, D.; Yichye, B.; Abutbul, M. To interpret or not to interpret PCA? This is our question. In Proceedings of the International AAAI Conference on Web and Social Media, Münich, Germany, 11–14 June 2019; Volume 13. [Google Scholar]
  12. Jade, A.M.; Srikanth, B.; Jayaraman, V.K.; Kulkarni, B.D.; Jog, J.P.; Priya, L. Feature extraction and denoising using kernel PCA. Chem. Eng. Sci. 2003, 58, 4441–4448. [Google Scholar] [CrossRef]
  13. Song, F.; Guo, Z.; Mei, D. Feature selection using principal component analysis. In Proceedings of the 2010 International Conference on System Science, Engineering Design and Manufacturing Informatization, Yichang, China, 12–14 November 2010; IEEE: Washington, DC, USA, 2010; Volume 1, pp. 27–30. [Google Scholar]
  14. Broomhead, D.S.; King, G.P. Extracting qualitative dynamics from experimental data. Phys. Nonlinear Phenom. 1986, 20, 217–236. [Google Scholar] [CrossRef]
  15. Jolicoeur, P.; Mosimann, J.E. Size and shape variation in the painted turtle. A principal component analysis. Growth 1960, 24, 339–354. [Google Scholar] [PubMed]
  16. Hansen, L.K.; Larsen, J.; Nielsen, F.Å.; Strother, S.C.; Rostrup, E.; Savoy, R.; Lange, N.; Sidtis, J.; Saver, C.; Paulson, O.B. Generalizable patterns in neuroimaging: How many principal components? NeuroImage 1999, 9, 534–544. [Google Scholar] [CrossRef] [PubMed]
  17. Giuliani, A.; Colosimo, A.; Benigni, R.; Zbilut, J.P. On the constructive role of noise in spatial systems. Phys. Lett. A 1998, 247, 47–52. [Google Scholar] [CrossRef]
  18. Sneath, P.H.A. Distortions of taxonomic structure from incomplete data on a restricted set of reference strains. Microbiology 1983, 129, 1045–1073. [Google Scholar] [CrossRef]
  19. Poon, C.S.; Barahona, M. Titration of chaos with added noise. Proc. Natl. Acad. Sci. USA 2001, 98, 7107–7112. [Google Scholar] [CrossRef] [PubMed]
  20. Jolliffe, I.T. Principal Component Analysis; Springer: New York, NY, USA, 2002; pp. 111–149. [Google Scholar]
  21. Kay, S. Can detectability be improved by adding noise? IEEE Signal Process. Lett. 2009, 7, 8–10. [Google Scholar] [CrossRef]
  22. Amini, A.; Wainwright, M.J. High-dimensional analysis of semidefinite relaxations for sparse principal components. In Proceedings of the 2008 IEEE International Symposium on Information Theory, Toronto, ON, Canada, 6–11 July 2008; IEEE: Washington, DC, USA, 2008. [Google Scholar]
  23. Garg, A.; Belarbi, M.O.; Tounsi, A.; Li, L.; Singh, A.; Mukhopadhyay, T. Predicting elemental stiffness matrix of FG nanoplates using Gaussian Process Regression based surrogate model in framework of layerwise model. Eng. Anal. Bound. Elem. 2022, 143, 779–795. [Google Scholar] [CrossRef]
  24. Mukdasai, K.; Sabir, Z.; Raja, M.A.Z.; Sadat, R.; Ali, M.R.; Singkibud, P. A numerical simulation of the fractional order Leptospirosis model using the supervise neural network. Alex. Eng. J. 2022, 61, 12431–12441. [Google Scholar] [CrossRef]
Figure 1. LEV diagram of the signal. The logarithm of the eigenvalues is shown on the h-axis. The number of principal components is chosen accordingly to the value of y, for which the graph assumes, approximately, a straight-line trend. In this case, the number of bona fide signal components suggested by the method was five.
Figure 1. LEV diagram of the signal. The logarithm of the eigenvalues is shown on the h-axis. The number of principal components is chosen accordingly to the value of y, for which the graph assumes, approximately, a straight-line trend. In this case, the number of bona fide signal components suggested by the method was five.
Stats 07 00004 g001
Figure 2. The X-axis shows the sequence of values subsequently extracted in order; the Y-axis shows the simulated values for the original series S 0 (black line) and the original series + small signal S 1 (red line). The two series are practically coincident and score a Pearson correlation near unity. The weak signal is almost perfectly masked by the Gaussian noise.
Figure 2. The X-axis shows the sequence of values subsequently extracted in order; the Y-axis shows the simulated values for the original series S 0 (black line) and the original series + small signal S 1 (red line). The two series are practically coincident and score a Pearson correlation near unity. The weak signal is almost perfectly masked by the Gaussian noise.
Stats 07 00004 g002
Figure 3. For increasing noise levels (X-axis), the correlation of the corrupted signal with its original version (Y-axis) reaches a bottom end of near-zero correlation. A unit value of added noise can be considered as the approximate ‘volume of titration’ for the analyte (signal) in the solution; when reaching this point, the linear correlation/added noise relation is lost.
Figure 3. For increasing noise levels (X-axis), the correlation of the corrupted signal with its original version (Y-axis) reaches a bottom end of near-zero correlation. A unit value of added noise can be considered as the approximate ‘volume of titration’ for the analyte (signal) in the solution; when reaching this point, the linear correlation/added noise relation is lost.
Stats 07 00004 g003
Table 1. The table reports the explained variance proportion of the components extracted in [17]. PC1 accounts for the, by far, major part of the variance, but does not convey any semantically relevant information. PC2 and PC3 are the minor, but semantically relevant components, able to reconstruct the European cities’ configuration. PC4 and PC5 are pure noise components corresponding to measurement error.
Table 1. The table reports the explained variance proportion of the components extracted in [17]. PC1 accounts for the, by far, major part of the variance, but does not convey any semantically relevant information. PC2 and PC3 are the minor, but semantically relevant components, able to reconstruct the European cities’ configuration. PC4 and PC5 are pure noise components corresponding to measurement error.
PC1PC2PC3PC4PC5
0.9960.00290.000060.000040.000005
Table 2. Explained variance for normal data distribution + low-amplitude signal. As expected in the case of a Gaussian distribution, the amount of explained variance gently decays along components with no clear-cut separation between the signal and noise floor. PC14’s loading pattern partially mirrors the added low-amplitude saw-signal.
Table 2. Explained variance for normal data distribution + low-amplitude signal. As expected in the case of a Gaussian distribution, the amount of explained variance gently decays along components with no clear-cut separation between the signal and noise floor. PC14’s loading pattern partially mirrors the added low-amplitude saw-signal.
VariablesComponents
PC1PC2PC3PC4PC5PC6PC7PC8
T00.2930.0000.4430.0000.1910.0000.2590.000
T10.3070.0000.0000.4470.141−0.2720.175−0.130
T20.180−0.167−0.168−0.1050.473−0.298−0.2540.000
T30.133−0.3160.0000.0000.2580.597−0.1820.000
T4−0.114−0.434−0.4000.000−0.1160.1430.4020.180
T5−0.321−0.3260.000−0.4030.000−0.1510.218−0.395
T6−0.383−0.2020.1900.121−0.2630.138−0.506−0.154
T7−0.4280.0000.1590.1020.0000.0000.1670.638
T8−0.2870.3450.2230.1720.1050.0000.314−0.437
T9−0.1310.411−0.3770.1610.000−0.130−0.310−0.179
T100.0000.396−0.138−0.3720.1180.253−0.1070.289
T110.2410.221−0.1230.000−0.3400.4480.213−0.208
T120.1920.000−0.321−0.202−0.481−0.2850.0000.000
T130.2390.0000.388−0.490−0.168−0.1920.0000.000
T140.263−0.1580.2500.339−0.4080.000−0.2200.000
Explained Variance0.1010.0960.0840.0830.0820.0730.0680.066
VariablesComponents
PC9PC10PC11PC12PC13PC14PC15
T00.2060.4470.4390.2090.1830.2630.114
T10.0000.370−0.4950.1690.000−0.3570.000
T20.466−0.2590.0000.144−0.2310.3950.000
T30.2770.1220.000−0.362−0.150−0.3410.242
T40.1470.122−0.128−0.1810.3850.349−0.225
T50.1690.0000.0000.3770.209−0.2400.354
T60.1400.4080.0000.240−0.1250.166−0.320
T70.3150.000−0.1940.000−0.2810.0000.323
T80.414−0.1870.000−0.3160.0000.000−0.328
T90.1490.2680.000−0.1790.3530.1710.469
T100.2350.000−0.1650.3060.344−0.293−0.353
T110.0000.000−0.2990.314−0.3160.3320.235
T120.3500.1980.416−0.105−0.293−0.233−0.119
T130.0000.197−0.454−0.4510.0000.1560.000
T140.333−0.4490.0000.0000.404−0.1270.101
Explained Variance0.0580.0570.0560.0510.0480.0370.031
Table 3. The table reports the adjusted R 2 of the relation between the original/noise-corrupted correlation and the amount of added noise in the case of the ‘noise floor’ (PC6–PC15) components. It is worth noting the lack of any evident link between the component ordering and correlation values and the decidedly higher sensitivity to the added noise of PC14 with respect to the other components.
Table 3. The table reports the adjusted R 2 of the relation between the original/noise-corrupted correlation and the amount of added noise in the case of the ‘noise floor’ (PC6–PC15) components. It is worth noting the lack of any evident link between the component ordering and correlation values and the decidedly higher sensitivity to the added noise of PC14 with respect to the other components.
Value of Adjusted R-Squared
PC6PC7PC8PC9PC10PC11PC12PC13PC14PC15
0.3930.5880.3500.5300.6120.6620.6090.6000.8000.681
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Giuliani, A.; Vici, A. On the (Apparently) Paradoxical Role of Noise in the Recognition of Signal Character of Minor Principal Components. Stats 2024, 7, 54-64. https://doi.org/10.3390/stats7010004

AMA Style

Giuliani A, Vici A. On the (Apparently) Paradoxical Role of Noise in the Recognition of Signal Character of Minor Principal Components. Stats. 2024; 7(1):54-64. https://doi.org/10.3390/stats7010004

Chicago/Turabian Style

Giuliani, Alessandro, and Alessandro Vici. 2024. "On the (Apparently) Paradoxical Role of Noise in the Recognition of Signal Character of Minor Principal Components" Stats 7, no. 1: 54-64. https://doi.org/10.3390/stats7010004

Article Metrics

Back to TopTop