Skip to Content
You are currently on the new version of our website. Access the old version .
  • Proceeding Paper
  • Open Access

22 September 2021

A Generative Adversarial Network Based Autoencoder for Structural Health Monitoring †

,
,
,
,
,
and
1
Dipartimento di Ingegneria Civile ed Ambientale, Politecnico di Milano, Piazza L. da Vinci 32, 20133 Milano, Italy
2
Laboratoire MSSMat UMR 8579, Université Paris Saclay—CNRS—CentraleSupélec, 3 rue Joliot Curie, 91190 Gif-sur-Yvette, France
3
Dipartimento di Matematica, Politecnico di Milano, Piazza L. da Vinci 32, 20133 Milano, Italy
*
Author to whom correspondence should be addressed.

Abstract

Civil structures, infrastructures and lifelines are constantly threatened by natural hazards and climate change. Structural Health Monitoring (SHM) has therefore become an active field of research in view of online structural damage detection and long term maintenance planning. In this work, we propose a new SHM approach leveraging a deep Generative Adversarial Network (GAN), trained on synthetic time histories representing the structural responses of both damaged and undamaged multistory building to earthquake ground motion. In the prediction phase, the GAN generates plausible signals for different damage states, based only on undamaged recorded or simulated structural responses, thus without the need to rely upon real recordings linked to damaged conditions.

1. Introduction

Bridges, power generation systems, aircrafts, buildings and rotating machinery are only few instances of structural and mechanical systems which play an essential role in the modern society, even if the majority of them are approaching the end of their original design life [1]. Taking into account that their replacement would be unsustainable from an economic standpoint, alternative strategies for early damage detection have been actively developed so to extend the basis service life of those infrastructures. Furthermore, the advent of novel materials whose long-term behaviour is still not fully understood drives the effort for effective Structural Health Monitoring (SHM), resulting in a saving of human lives and resources [1].
SHM consists of three fundamental steps: (i) measurement, at regular intervals, of the dynamic response of the system; (ii) selection of damage-sensitive features from the acquired data; (iii) statistical analysis of those attributes to assess the current health state of the structure. To characterize the damage state of a system, the method relying on hierarchical phases, originally proposed by [2] represents the currently adopted standard. The latter prescribes several consecutive identification phases (to be tackled in order), namely: check the existence of the damage, the location of the damage, its type, extent and the system’s prognosis. Damaged states are identified by comparison with a reference condition, assumed to be undamaged. The detection of the damage location relies upon a wider awareness of the structural behaviour and the way in which it is influenced by damage. This information, along with the knowledge of how the observed features are altered by different kinds of damage, allows to determine the type of damage. The last two phases require an accurate estimation of the damage mechanisms in order to classify its severity and to estimate the Remaining Useful Life (RUL).
All the steps mentioned above rely on continuous data acquisition and processing to obtain information about the current health condition of a system. In the last few years, the concept of Digital Twin has emerged, combining data assimilation, machine learning and physics-based numerical Simulations [1], the latter being essential to completely understand the physics of the structure and damage mechanisms. A suitable tool able to extract main dominant features from a set of data is represented by neural networks [3], especially generative models such as Generative Adversarial Networks (GANs) [4] and Variational Autoencoders (VAEs) [5].
In this paper, an application of the generative neural network RepGAN, proposed by [6], is presented in the context of SHM. Section 2 provides an overview on existing works. In Section 3, the application of RepGAN to Structural Health Monitoring is presented. In Section 4, extensive numerical results are illustrated, while Section 5 gathers some concluding remarks.

3. Methods

With the purpose of learning a semantically meaningful and disentangled representation of the SHM time-histories, we adopted in this study the architecture called RepGAN, originally proposed in [6]. RepGAN is based on an encoder-decoder structure (both represented by deep CNNs made of stacked 1D convolutional blocks), with a latent space Z = C , S , N . C 0 , 1 d C a categorical variable representing the damage class(es), with C p C which is generally chosen as a categorical distribution over d C classes, i.e., p C = Cat ( d C ) . S R d S is a continuous variable of dimension d S , with S p S , generally p S = N ( 0 , I ) or the uniform distribution p S = U ( 1 , 1 ) . Finally, N R d N is a random noise of d N independent components, with N p N , generally p N N ( 0 , I ) . RepGAN adopts the conceptual frameworks of VAEs and InfoGAN, combining the learning of two representations x z ^ x ^ and z x ^ z ^ , respectively. The x z ^ x ^ scheme must learn to map multiple data instances x ( i ) into their images (via encoder F θ X ) in a latent manifold z ^ ( i ) = F θ X ( x ( i ) ) and back into a distinct instance in data space x ^ ( i ) = G θ Z F θ X ( x ( i ) ) (via decoder G θ Z ), providing satisfactory results in reconstruction. z x ^ z ^ maps multiple data latent instances into the same data representation, in order to guarantee impressive generation and clustering performance. Combining the two surjective mappings, in RepGAN the two learning tasks x z ^ x ^ and z x ^ z ^ are performed together with shared parameters in order to obtain a bijective mapping x z . In practice, the training of z x ^ z ^ is iterated five times more than the x z ^ x ^ . This ability to learn a bidirectional mapping between the input space and the latent space is achieved through a symmetric adversarial process. The Empirical Loss function can be written as:
L S = D J S ( p X ^ | ( C , S , N ) | | p X ) + D J S ( q C ^ | X | | p C ) + D J S ( q S ^ | X | | p S ) + D J S ( q N ^ | X | | p N ) E p C E p X ^ | C ln q C ^ | X E p S E p X ^ | S ln q S ^ | X E p X E q ( C , S , N ) | X ln p X | ( C , S , N )
with the terms:
  • E p C E p X ^ | C ln q C ^ | X minimizing the conditional entropy S ( C | X ) ;
  • E p S E p X ^ | S ln q S ^ | X minimizing the conditional entropy S ( S | X ) .
are introduced in order to constrain a deterministic and injective encoding mapping (see Appendix B). On the other hand, the term
  • E p X [ E q ( C , S , N ) | X ln p X | ( C , S , N ) ] .
penalizes the learning scheme, in order to minimize the conditional entropy S ( X | ( C , S , N ) ) , i.e., in order to grant a good reconstruction.
Following the original RepGAN formulation:
  • E p X E q ( C , S , N ) | X ln p X | ( C , S , N ) is enforced penalizing the L 1 -norm | X G θ Z F θ X ( X ) | ;
  • E p S E p X ^ | S ln q S ^ | X corresponds to the InfoGAN L I penalty, and it is maximized via the reparametrization trick (structuring the S branch of the encoder-decoder structure as a VAE, see [5]).
Finally, E p C E p X ^ | C ln q C ^ | X is maximized in a supervised way, considering the actual class of labeled signals x ( i ) : x d ( i ) corresponding to a damaged structure and x u ( i ) to an undamaged one, respectively. RepGAN provides an informative and disentangled latent space associated with the damage class C . The most significant aspect of the approach is the efficiency in generating reasonable signals for different damage states only on the basis of undamaged recorded or simulated structural responses. Both generators F θ X , G θ Z and discriminators D ω X , D ω C , D ω S and D ω N are parametrized via 1D CNN (and strided 1D CNN), following [8]. Our RepGAN model has been designed using the Keras API, and trained employing a Nvidia Tesla K40 GPU (on the supercomputer Ruche, the cluster of the Mésocentre Moulon of Paris Saclay University).

4. Results and Discussion

In the following, a case study is considered in order to prove the ability of the new architecture to achieve the three fundamental tasks of semantic generation, clustering and reconstruction. The reference example is a shear building subject to an earthquake ground motion whose signals are taken from the STEAD seismic database [14]. STEAD [14] is a high-quality, large-scale, and global data set of local earthquake and non-earthquake signals recorded by seismic instruments. In this work, local earthquake wave forms (recorded at local distances within 350 km of earthquakes) have been considered. Seismic data are constituted by three wave forms of 60 s duration, recorded in east–west, north–south, and up-dip directions, respectively. The structure is composed of 39 storeys. The mass and the stiffness of each floor, in undamaged conditions, are, respectively, m = 625 × 10 3 kg and k = 1 × 10 9   k N m . Damage is simulated through the degradation of stiffness. In the present case, the stiffness reduction has been set equal to 50% of the above mentioned value. The structural response of the system is evaluated considering one degree-of-freedom (dof) per floor. To take into account damping effects, a Rayleigh damping model has been considered.
The following results have been obtained considering 100 signals in both undamaged and damaged conditions for a total of 200 samples, with separated training and validation data sets. Each signal is composed of 2048 time steps with dt = 0.04 s. The training process has been performed over 2000 epochs. The reconstruction capability of the proposed network has been evaluated through the Goodness-of-Fit (GoF) criteria [15] where both the fit in Envelope (EG) and the fit in Phase (FG) are measured. An example is shown in Figure 1. The values 9.17 and 9.69, respectively, related to EG and PG testify the excellent reconstruction quality.
Figure 1. Time–Frequency Goodness-of-Fit criterion: the black line represents the original time-histories x ( i ) while the red time history depicts the result of the RepGAN reconstructions G Z F X ( x ( i ) ) . GoF is evaluated between 0 and 10: the higher the score, the better is the reconstruction. Frequency Envelope Goodness (FEG), Time–Frequency Envelope Goodness (EG), Time Envelope Goodness (TEG), Frequency Phase Goodness (FPG), Time–Frequency Phase Goodness (PG) and Time Phase Goodness (TPG).
The capability of reproducing signals for different damage scenarios can be appreciated from Figure 2 which presents the original structural response (black) and the corresponding generated one (orange) in both undamaged (left panel in Figure 2) and damaged (right panel in Figure 2) conditions. Regarding the classification capability, the classification report and the confusion matrix in Figure 3 highlight the fact that the model is able to correctly assign the damage class to the considered time histories.
Figure 2. Examples of reconstructed signals for undamaged (left) and damaged (right) time-histories. The black lines represent the original time-histories x u ( i ) and x d ( i ) , respectively. The orange time histories represent the result of the RepGAN reconstructions G Z F X ( x u ( i ) ) and G Z F X ( x d ( i ) ) , respectively. The proposed examples represent the normalized displacement of the 1st floor of the building in object.
Figure 3. Evaluation of the classification ability of the model. On the left panel, precision, recall, f1-score and accuracy values are reported. A precision score of 1.0 for a class C means that every item labelled as belonging to class C does indeed belong to class C, whereas a recall of 1.0 means that every item from class C was labelled as belonging to class C. F1-score is the harmonic mean of the precision and recall. Accuracy represents the proportion of correct predictions among the total number of cases examined. On the right panel, the confusion matrix allows to visualize the performance of the model: each row of the matrix represents the instances in the actual class, while each column depicts the instances in the predicted class.

5. Conclusions

In this paper, we introduce a SHM method based on a deep Generative Adversarial Network. Trained on synthetic time histories that represent the structural response of a multistory building in both damaged and undamaged conditions, the new model achieves high classification accuracy (Figure 3) and satisfactory reconstruction quality (Figure 1 and Figure 2), resulting in a good bidirectional mapping between the input space and the latent space. However, the major innovation of the proposed method is the ability to generate reasonable signals for different damage states, based only on undamaged recorded or simulated structural responses. As a consequence, real recordings linked to damaged conditions are not requested. In our future work, we would like to extend our approach to real-time data. We will further consider a dataset constituted by a far larger number of time histories.

Author Contributions

Conceptualization, G.C., L.R., F.G., S.M. and A.C.; methodology, G.C. and F.G.; software, G.C., L.R. and F.G.; validation, G.C. and F.G.; formal analysis, G.C. and F.G.; investigation, G.C. and F.G.; resources, F.G.; data curation, G.C., L.R. and F.G.; writing–original draft preparation, G.C., L.R., M.T., F.G., S.M., A.M. and A.C.; writing–review and editing, G.C. and F.G.; visualization, G.C. and F.G.; supervision, F.G., S.M. and A.C.; project administration, F.G. and A.C.; Funding acquisition, G.C., F.G. and A.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

All data generated during the study are available from the corresponding author upon reasonable request.

Acknowledgments

The training and testing of the neural network has been performed exploiting the supercomputer resources of the Mésocentre Moulon (http://mesocentre.centralesupelec.fr, last accessed 14 February 2022), the cluster of CentraleSupélec and ENS Paris-Saclay, hosted within the Paris-Saclay University and funded by the Contrat Plan État Région (CPER). This work has been developed thanks to the scholarship “Tesi all’estero—a.y. 2020/2021—second call” funded by Politecnico di Milano.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Shannon’s Entropy

  • Shannon’s entropy for a probability density function p X :
    S ( X ) = S p X = E X p X ln 1 p X = E X p X ln p X 0
  • Conditional Shannon’s entropy for X and Z:
    S X | Z = E Z P Z S p X | Z = E ( X , Z ) p X Z ln 1 p X | Z
    S ( X , Z ) = S ( Z | X ) + S ( X ) = S ( X | Z ) + S ( Z )
  • Cross-entropy:
    S ( p X Z | | q X Z ) = E ( X , Z ) p X Z ln 1 q X Z = E X p X E Z p Z | X ln 1 q X Z
  • Given a data set of identically independent distributed (i.i.d.) samples S = x ( i ) i = 1 N , the true yet unknown probability p X of extracting an instance x ( i ) can be approximated by the likelihood p θ X x ( i ) i = 1 N , whose entropy is
    S ( p θ X ) = ln p θ X x ( i ) i = 1 N = i N ln p θ X x ( i )

Appendix B. Kullback–Leibler Divergence

  • Kullback-Liebler divergence (non-symmetric):
    D K L ( p X Z | | q X Z ) = E ( X , Z ) p X Z ln p X Z q X Z = S ( p X Z ) + S ( p X Z | | q X Z ) S ( p X Z | | q X Z )
    D K L ( p X Z ] | | q X Z ) + S ( X ) = S ( X | Z ) E X p X S ( p Z | X ) + S ( p X Z | | q X Z ) S ( p X Z | | q X Z )
    D K L ( p X Z | | q X Z ) + S ( X ) D K L ( p X Z | | q X Z ) ]
  • Mutual Information between X and X | Z :
    I ( X , Z ) = S ( X ) S ( X | Z ) 0
    If p X | Z = p X ( ( X , Z ) are independent) then I ( X , Z ) = 0 . If p X | Z = δ ( Z f ( X ) ) with f deterministic, then I ( X , Z ) = max X , Z I ( X , Z ) = S ( X ) .
  • S ( Z | X ) = E Z p Z E X p X | Z ln p Z | X = = E X p X E Z p Z | X ln p Z | X q Z | X E Z p Z E X p X | Z ln q X | Z = = E X p X D K L p Z | X | | q Z | X E Z p Z E X p X | Z ln q X | Z E Z p Z E X p X | Z ln q X | Z

Appendix C. Generative Adversarial Networks (GAN)

  • Given X belonging to the probabilistic space ( Ω X , E X , Pr X ) with class C d , g (“d” corresponding to data and “g” to generated, and a discriminator D : Ω X 0 , 1 acting as an expert/critic:
    -
    P ( C = d ) = α ; P ( C = g ) = 1 α
    -
    P ( C = d | x ( i ) ) = D ( x ( i ) )
    -
    P ( C = d | x ( i ) ) = 1 D ( G ( z ( i ) ) )
    S ( C | X ) = E X p X E C p C | X ln p C | X = E C p C E X p X | C ln p C | X
    S ( C | X ) = α E X p X | C = d ln p C = d | X ( 1 α ) E X p X | C = g ln p C = g | X
    S ( C | X ) = α E X p X ln D X ( 1 α ) E Z p Z ln 1 D G ( Z )
    For tuneable conditional probability distributions D ω :
    max I ( X , C ) S ( C ) + max S ( C | X ) = S ( C ) + min S ( C | X )
    max I ( X , C ) S ( C ) + min G max D α E X p X ln D X + ( 1 α ) E Z p Z ln 1 D G ( Z )
    Thus, minimizing S + min G max D S ( C | X ) represents an upper bound for the Mutual Information between C and X , which is maximized by maximizing S ( C | X ) . For an optimum training, D must not be able to discriminate between x ( i ) and x ^ ( i ) , therefore α = 1 2 .

Appendix D. Standard Autoencoder

In the standard Autoencoder formulations [16,17], F and G are trained by maximizing I ( X , Z ) , namely:
F , G = arg max F , G I ( X , Z ) = arg min F , G H ( X | Z ) = arg min F , G E X p X E Z q Z | X ln 1 p X | Z
If the encoder and decoder are parametrized as neural networks, respectively, as F θ X and G θ Z , the AE loss can be approximated by the Empircal Loss:
θ X , θ Z = arg max θ X , θ Z i = 1 N ln p X | Z x ( i ) | Z = F θ X x ( i )
Given the fact that the Gaussian distribution has maximum entropy relative to all probability distributions covering the entire real line, the Empirical Loss in Equation (A2) can be maximized by the Empirical Loss with p X | Z = N G θ Z Z , σ 2 I :
θ X , θ Z = arg max θ X , θ Z i = 1 N 1 2 σ 2 x ( i ) G θ Z F θ X x ( i ) 2 + d Z 2 ln 2 π σ 2

References

  1. Farrar, C.R.; Worden, K. Structural Health Monitoring: A Machine Learning Perspective; Wiley: Oxford, UK, 2013. [Google Scholar] [CrossRef] [Green Version]
  2. Rytter, A. Vibrational Based Inspection of Civil Engineering Structures. Ph.D. Thesis, University of Aalborg, Aalborg, Denmark, 1993. [Google Scholar]
  3. Bishop, C.M. Neural Networks for Pattern Recognition; Oxford University Press: Oxford, UK, 1995. [Google Scholar]
  4. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative adversarial nets. In Proceedings of the 28th Annual Conference on Neural Information Processing Systems (NIPS 2014), Montreal, QC, Canada, 8–13 December 2014. [Google Scholar] [CrossRef]
  5. Kingma, D.P.; Welling, M. Auto-Encoding Variational Bayes. arXiv 2013, arXiv:1312.6114. [Google Scholar]
  6. Zhou, Y.; Gu, K.; Huang, T. Unsupervised Representation Adversarial Learning Network: From Reconstruction to Generation. In Proceedings of the 2019 International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, 14–19 July 2019. [Google Scholar] [CrossRef] [Green Version]
  7. Lindsay, B.G. Mixture Models: Theory, Geometry and Applications. In NSF-CBMS Regional Conference Series in Probability and Statistics; Institute of Mathematical Statistics and American Statistical Association: Suitland, MD, USA, 1995; Volume 5, p. 163. ISBN 0-940600-32-3. [Google Scholar]
  8. Radford, A.; Luke, L.M.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
  9. Makhzani, A.; Shlens, J.; Jaitly, N.; Goodfellow, I. Adversarial Autoencoders. In Proceedings of the International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  10. Dumoulin, V.; Belghazi, I.; Poole, B.; Mastropietro, O.; Lamb, A.; Arjovsky, M.; Courville, A. Adversarially Learned Inference. arXiv 2016, arXiv:1606.00704v3. [Google Scholar]
  11. Donahue, J.; Krähenbühl, P.; Darrell, T. Adversarial Feature Learning. arXiv 2017, arXiv:1605.09782. [Google Scholar]
  12. Li, C.; Liu, H.; Chen, C.; Pu, Y.; Chen, L.; Henao, R.; Carin, L. ALICE: Towards understanding adversarial learning for joint distribution matching. In Proceedings of the 31st International Conference on Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; Curran Associates Inc.: Red Hook, NY, USA, 2017; pp. 5501–5509. [Google Scholar]
  13. Chen, X.; Duan, Y.; Houthooft, R.; Schulman, J.; Sutskever, I.; Abbeel, P. InfoGAN: Interpretable Representation Learning by Information Maximizing Generative Adversarial Nets. In Proceedings of the 30th International Conference on Neural Information Processing Systems, Barcelona, Spain, 5–10 December 2016. [Google Scholar]
  14. Mousavi, S.M.; Sheng, Y.; Zhu, W.; Beroza, G.C. STanford EArthquake Dataset (STEAD): A Global Data Set of Seismic Signals for AI. IEEE Access 2019, 7, 179464–179476. [Google Scholar] [CrossRef]
  15. Kristekova, M.; Kristek, J.; Moczo, P. Time-frequency misfit and goodness-of-fit criteria for quantitative comparison of time signals. Geophys. J. Int. 2009, 178, 813–825. [Google Scholar] [CrossRef]
  16. Vincent, P.; Larochelle, H.; Bengio, Y.; Manzagol, P.-A. Extracting and Composing Robust Features with Denoising Autoencoders. In Proceedings of the 25th International Conference on Machine Learning, Helsinki, Finland, 5–9 July 2008; pp. 1096–1103. [Google Scholar]
  17. Vincent, P.; Larochelle, H.; Lajoie, I.; Bengio, Y.; Manzagol, P.-A. Stacked Denoising Autoencoders: Learning Useful Representations in a Deep Network with a Local Denoising Criterion. J. Mach. Learn. Res. 2010, 11, 3371–3408. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.