Next Article in Journal
Assessment of Spatial Variability in Ground Models Using Mini-Cone Penetration Testing
Previous Article in Journal
4D Printing: Bridging the Gap between Fundamental Research and Real-World Applications
Previous Article in Special Issue
Application of Distributed Acoustic Sensing in Geophysics Exploration: Comparative Review of Single-Mode and Multi-Mode Fiber Optic Cables
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MA_W-Net-Based Dual-Output Method for Microseismic Localization in Strong Noise Environments

College of Geoexploration Science and Technology, Jilin University, Xi Min Zhu Street No. 938, Changchun 130026, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(13), 5668; https://doi.org/10.3390/app14135668
Submission received: 26 May 2024 / Revised: 22 June 2024 / Accepted: 26 June 2024 / Published: 28 June 2024
(This article belongs to the Special Issue Seismic Data Processing and Imaging)

Abstract

:
With the continuous depletion of conventional oil and gas reservoir resources, the beginning of exploration and development of unconventional oil and gas reservoir resources has led to the rapid development of microseismic monitoring technology. Addressing the challenges of low signal-to-noise ratio and inaccurate localization in microseismic data, we propose a new neural network MA_W-Net based on the U-Net network with the following improvements: (1) The foundational U-Net model was refined by evolving the single-channel decoder into a two-channel decoder, aimed at enhancing microseismic event localization and noise suppression capabilities. (2) The integration of attention mechanisms such as the convolutional block attention module (CBAM), coordinate attention (CA), and squeeze-and-excitation (SE) into the encoder to bolster feature extraction. We use synthetic data for evaluating the proposed method. Comparing with the normal U-net network, our accuracy in seismic recordings with a signal-to-noise ratio of −15 is improved from 78 percent to 93.5 percent, and the average error is improved from 2.60 m to 0.76 m. The results show that our method can accurately localize microseismic events and denoising processes from microseismic records with a low signal-to-noise ratio.

1. Introduction

The microseismic monitoring technique is an important tool for monitoring the dynamic state of the formation and has been widely used in hydraulic fracturing [1,2,3] and geothermal development [4,5]. The key component of microseismic monitoring is the localization of microseismic events. Microseismic localization is the process of accurately reflecting the location of an earthquake source from microseismic signals recorded by geophones distributed on the surface or in wells. Accurately locating microseismic events is critical to better understanding rupture and fault behavior [6].
Microseismic events are relatively weak in energy, ranging in magnitude from −3 to 1 MW [7]. These events often exhibit a low signal-to-noise ratio (SNR) due to the influence of the complex monitoring environment (e.g., traffic noise, industrial interference, drilling interference), current noise, and spike-like noise. The interference of large amounts of noise makes subsequent microseismic localization extremely difficult. Denoising helps to localize microseismic events with low SNRs more accurately. Traditional denoising approaches have been employed to mitigate noise within microseismic data sets [8,9,10]. Furthermore, deep learning techniques, such as leveraging neural networks to construct mapping functions [11], have proven effective in addressing complex challenges. Specifically, in microseismic exploration, deep learning can solve the localization problem in the case of a poor SNR [12,13,14,15,16].
Recently, microseismic localization methods based on deep learning have become increasingly popular [17,18,19,20,21]. Trained networks can directly predict source locations from recorded microseismic data and have the potential to monitor microseismic events in real-time. Traditional deep learning approaches for localizing a singular source often struggle with accuracy, as they can only provide the location of one microseismic event, which may be compromised by the noise from concurrent events. Perol et al. [17] introduced ConvNetQuake, the inaugural convolutional neural network designed for earthquake detection and localization, aiming to tackle seismic monitoring challenges through supervised classification. Zhang et al. [18] crafted a deep neural network tailored to seismic source offset imaging, capable of identifying two distinct events within diffraction stacked images from continuous seismic records. Wang et al. [19] developed two convolutional neural networks for identifying multiple earthquake sources: one to categorize event counts and another to predict several epicenter locations. Vinard et al. [20] leveraged synthetic data to train a U-Net model, subsequently applying it to field data for microseismic event localization, demonstrating that real data can be effectively augmented with synthetic data for network training purposes. Feng et al. [21] unveiled a technique for microseismic source localization utilizing deep reinforcement learning.
In this study, we propose a novel neural network based on U-Net, named MA_W-Net, which combines low-resolution semantic information with high-resolution localized spatial information using skip connections. However, the existing U-Net architectural system has some limitations, such as the relatively small number of backbone layers and the feature extraction capability not being as good as that of other networks, which may lead to lower accuracy in localizing the earthquake source point. To solve these problems, we have made the following improvements: (1) We built a deep neural network with one encoder and two decoder structures based on U-Net, which is a multi-task learning based on the parameter hard-sharing mechanism to realize microseismic localization and denoising in a strong noise environment. Among them, the main task is to localize the microseismic source and the auxiliary task is denoising. The auxiliary task can help the network perform better in localizing microseismic events with a low SNR. And we define the label of the localization task as a Gaussian mixture model which contains the source location information of different microseismic events. (2) We have integrated multiple attention mechanisms into MA_W-Net, including CBAM in the encoder and the CA and SE modules in the decoder. The attention modules can learn to use global information to selectively emphasize useful information features and suppress useless features without increasing computational complexity, thus enabling better processing of input data.

2. Methods and Principles

Multi-task learning (MTL) is an important branch in the field of machine learning that aims to improve the generalization ability and efficiency of models by learning multiple related tasks simultaneously. Based on the assumption that all tasks, or at least some, are relevant, learning multiple tasks jointly has been found empirically and theoretically to lead to better performance than learning individual tasks independently [22]. The hard-sharing mechanism of implicit layer parameters is one of the most common approaches in multi-task learning. Multiple tasks share part of the hidden layer and learn some common low-level features while retaining the output layer for different tasks. Throughout the learning process, the tasks exchange beneficial information, thereby amplifying the efficacy of each task’s learning process, enhancing the model’s overall generalization capabilities, and diminishing the likelihood of overfitting.
U-Net is a model obtained by improving on the full convolutional neural network [23]. U-Net has low requirements on the size of training data, high segmentation accuracy, and is widely used in image segmentation tasks. U-Net and its variants have achieved remarkable results in the geophysical field due to good feature extraction capabilities [24,25,26]. The architecture of U-Net is inherently symmetrical, featuring a contraction path (encoder), dedicated to feature extraction, and an expansion path (decoder), which employs up-sampling through transposed convolution to restore the feature map to the original image’s dimensions. U-Net facilitates the transfer of features from the contraction to the expansion path via skip connections, effectively integrating surface-level high-resolution details with profound abstract features.
The input data to the network are low-SNR grayscale microseismic recording maps. The encoder first resizes the input data using a convolutional layer, followed by a Residual (RES) module and a maximum pooling operation. Decoder 1 outputs a Gaussian mixture distribution containing information about the location of the microseismic source after an up-sampling, convolution, and convolution and coordinate attention (CCA) module. Decoder 2 outputs a microseismic record after noise suppression after an up-sampling, convolution, and convolution and squeeze-and-excitation attention (CSEA) module.
The following section describes RES module, CCA module, CSEA module, and the overall architecture of the MA_W-Net network:

2.1. RES Module

Convolutional module attention module (CBAM) [27] can solve the problem of what and where the image features are utilizing a channel attention module and a spatial attention module. Therefore, we form an improved RES module by combining residual connectivity between three layers of convolution and CBAM. The RES module structure is shown in Figure 1.
The attention module, a pivotal data processing technique in machine learning, empowers neural networks to concentrate on specific local features within the input data for effective information extraction. CBAM contains two sequential sub-modules, namely the channel attention module (CAM) and the spatial attention module (SAM), as shown in Figure 2.
M c ( F ) = σ ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) ) = σ W 1 W 0 F a v g c + W 1 W 0 F m a x c
CAM (Figure 3a) is used to learn by integrating the spatial information of the input feature maps after integrating them into a multilayer perceptron (MLP) through average pooling and maximum pooling operations, and the activation function is ReLU. The output features of the two MLPs are summed element-by-element to generate the channel attention value M C through the Sigmoid activation function.
M s ( F ) = σ ( f 7 × 7 ( A v g P o o l F ; M a x P o o l F ) ) = σ f 7 × 7 F a v g s ; F m a x s
SAM (Figure 3b) takes the feature maps output from the CAM as input and performs channel-based global maximum pooling and global tie pooling operations on the feature maps to obtain two H × W × 1 feature maps, where H and W are the height and width of the feature map. The Concat operation splices the two feature maps, performs a convolution operation with a convolution kernel of 7 × 7, and reduces the number of channels to 1, i.e., we obtain a feature map of H × W × 1. Finally, a Sigmoid activation function generates the spatial attention value M S . This is then multiplied with the input feature map to obtain the final result.

2.2. CCA and CSEA Modules

In neural networks, the integration of an attention mechanism within the decoder enables the capture of feature correlations. Utilizing coordinate attention (CA) [28] and squeeze-and-excitation attention (SE) [29], we have developed two innovative attention modules, CCA and CSEA, to enhance the denoising ability and localization of the network. Attention mechanisms help networks to better handle complex tasks and data by optimizing resource allocation, improving model accuracy, enhancing flexibility and adaptability, and reducing redundant information interference. During the downward propagation of the neural network, there will be important information that may be lost due to the maximum pooling operation. To alleviate this problem, we adopt a U-Net-like framework for the joining process in the CCA and CSEA modules, which mainly performs time-consuming convolutional computations in low-resolution space. Among them, the fusion of shallow and deep features is widely used. Such feature fusion not only facilitates a seamless information flow from superficial to in-depth layers but also proves beneficial for maintaining features, supporting efficient gradient backpropagation, and minimizing information loss during subsequent propagation.

2.2.1. CCA Module

The structure of the CCA module is shown in Figure 4, which mainly consists of two 2D convolutions and CA. We can enhance the localization effect by adding CA.
CA reintroduces the channel attention mechanism along with the spatial attention mechanism, which embeds location information into the channel attention in the attention mechanism. The implementation of CA is shown in Figure 5 and can be divided into two main parallel stages. CA pools the input feature maps in both height and width directions with global average pooling to obtain the feature maps in both height and width directions, respectively. We enter the dimensions of the feature layer as [H, W, C], where H and W are the height and width of the feature map, respectively, and C is the number of channels. After the above operations, the feature had dimensions obtained in height and width are [1, W, C] and [H, 1, C], respectively. Afterward, the two parallel stages are merged to transpose the height and width to pass a dimension, and then stacking is performed to fuse the height and width features, at which point the feature layer size obtained is [1, H + W, C], followed by convolution, normalization, and ReLU activation function to obtain a feature layer size of [1, H + W, C/r], where r is the scaling factor.
The process is bifurcated into two parallel stages, where the dimensions of height and width are segregated into [1, W, C/r] and [1, H, C/r], respectively. Following this, a transpose operation is executed, yielding two distinct feature layers: [W, 1, C/r] and [1, H, C/r]. The application of a Sigmoid function then determines the attention distribution across both the vertical and horizontal dimensions. Finally, merging the original features through multiplication culminates in the implementation of the CA mechanism.

2.2.2. CSEA Module

The architecture of the CSEA module is shown in Figure 6, which mainly consists of two 2D convolutions and the squeeze-and-excitation attention (SE).
Figure 7 illustrates the structure of the SE module. This module has the capability to dynamically allocate weights across various channels, a process autonomously learned by the network with minimal human oversight. The SE module is adept at explicitly mapping the interrelations among convolutional feature channels, thereby enhancing the network’s representational power. It harnesses global information to judiciously highlight salient features while diminishing the less relevant ones. Incorporating the SE module significantly aids in the noise reduction process of seismic data.

2.3. Network Architecture

The MA_W-Net network introduced in this study based on an enhanced U-Net framework, shown in Figure 8. This architecture consists a single encoder and dual decoders. The two decoders are responsible for seismic source localization and noise suppression, respectively. They collaboratively utilize the encoder for the end-to-end parallel training of features extracted from noisy microseismic data. The encoder is tasked with isolating crucial features from these recordings. Utilizing these features, Decoder 1 constructs a Gaussian mixture model to provide the location details of microseismic events. Subsequently, Decoder 2 then attempts to reconstruct the clean microseismic record using the extracted features and outputs a noise-compressed microseismic record. In this study, it is assumed that microseismic events have been detected, and the input data are microseismic recordings with a low SNR received by geophones with a size of 512 × 64 × 1. In the encoder part, we first resize the input data using 4 convolutional layers, each with a filter of 64 and a convolutional kernel of 5 × 5. After two more RES modules, the output size is 64 × 64 × 64. The remainder of the encoder contains three down-sampling modules, each of which consists of a 2 × 2 max-pooling [30] and two RES modules. This configuration effectively doubles the channel count while reducing the feature map’s dimensions by half with each down-sampling iteration.
The decoder contains a series of up-sampling processes. In Decoder 1, each process is up-sampled using a 2 × 2 transposed convolution, doubling the size of the feature map and halving the number of channels. This process involves integrating feature maps from both the encoder and decoder, following a single convolution and two CCA modules. The final output of Decoder 1 has a size of 64 × 128 and represents a Gaussian mixture distribution containing information about the location of the seismic source. In Decoder 2, each process is up-sampled using a 2 × 2 transposed convolution, doubling the size of the feature map and halving the number of channels. Here, after merging encoder and decoder feature maps through one convolution and two CSEA modules, transposed convolution is then applied to revert the feature maps back to the original input dimensions. The final output size of Decoder 2 is 512 × 64, representing a microseismic record after noise suppression. Except for the two output convolutional layers, which use a Sigmoid activation function [31], all other convolutional layers are activated with a ReLU activation function.

3. Numerical Example

3.1. Data Organization

For microseismic source localization, we typically need a good understanding of the background velocity model of the study area. In this study, we assume that we could obtain a good velocity model of the study area using many velocity inversion methods. We selected the Marmousi [32] model as the background velocity model, which can simulate complex geological conditions. We assumed that the true velocity model is known. We assessed the model’s predictive performance across various SNRs. Then, a comparison was made with the single-task localization model.
We use the finite difference method to generate simulated microseismic records. Figure 9 shows the velocity model, which has an extent of 640 m × 1280 m and a grid spacing of 10 m: A total of 64 geophones were positioned at every other grid point on the velocity model’s surface. We use a Ricker wavelet with a sampling rate of 1 ms as a function of the source time, and the source location and excitation moment are randomly generated. Data were generated using one seismic source at a time, 4000 training data and labels were prepared, and the simulated data were divided into training and test sets in a 9:1 ratio.
Our computing device is NVIDIA Quadro RTX 4000, our neural network framework is TensorFlow-GPU 2.7.0, with 8 GB of video memory. We use ReLU as the activation function of the network, with the Sigmoid function applied in the final layer. We trained the network using Adam’s algorithm [33]. The entire training process was carried out for 600 iterations, the learning rate was set to 0.0001, and the number of training data in each batch was 32. This setting provides a basis for subsequent model evaluation and comparison.

3.2. Data Preprocessing

The input data to the neural network are noisy microseismic records, generated by adding Gaussian noise to clean microseismic records. The main task is to localize the source location of microseismic events from noisy data. Traditional deep learning approaches typically predict the exact coordinates of seismic sources, necessitating a predefined number of sources to structure the model’s output layer. However, with an increasing count of seismic sources, there arises a need for developing multiple models, each with a unique architecture. To address this, we employ a Gaussian mixture model to both represent the microseismic source locations and serve as the output label for this key task. The probability distribution of the source location of each microseismic event can be generated by a Gaussian distribution [34]:
f i x , z = e x p ( ( x x 0 ) 2 + ( z z 0 ) 2 2 σ 2 )
where ( x 0 , y 0 ) are the coordinates of the epicenter point and σ is the variance.
A Gaussian mixture model is essentially a composite of K Gaussian distributions. In scenarios where the microseismic record encompasses multiple events, this model is formulated by the linear aggregation of Gaussian distributions attributed to each event. Within the resultant Gaussian mixture model, the peak’s location signifies the microseismic event’s source location. To ensure high resolution across all sources, the standard deviation for the Gaussian distribution is uniform for each source.
For enhanced network learning, each composite Gaussian model undergoes normalization to a range of [0, 1].
Traditional deep learning approaches for denoising employ supervised learning, utilizing noisy data for input and pristine data for labels. However, this method faces significant challenges in the denoising of seismic data, primarily due to the difficulty in acquiring clean seismic data for labels. Lehtinen et al. [35] introduced a novel self-supervised learning technique for denoising that relies solely on noisy data for training. This method hinges on an L2 paradigm characteristic: the L2 norm remains constant for two vectors substituted with random numbers anticipated to be identical. Consequently, utilizing the L2 loss function for training ensures that network parameters remain stable, even in the presence of noise-contaminated label data. Thus, obtaining clean seismic data is not a prerequisite for effective denoising.

3.3. Model Training

The loss function serves as a metric to quantify the discrepancy between the model’s predicted outputs and the actual values. The value of the loss function directly affects the predictive performance of the model, the smaller the value of the loss function, the better the predictive performance of the model. The loss function consists of two parts: a weighted mean square error loss function for the main task and a conventional mean square error loss function for the auxiliary task. The overall loss is calculated as follows:
L = λ 1 L m a i n + λ 2 L a u x
where L m a i n denotes the loss of the main task, L a u x denotes the loss of the auxiliary task, and λ denotes the weight of the two tasks.
The mean square error loss function finds the sum of the squares of the errors at the points corresponding to the predicted and true values.
L a u x = 1 N i = 1 N y i t r u e y i p r e d 2
where N denotes the sample size, y t r u e denotes the true value, and y p r e d denotes the prediction result. The smaller the value of the loss function, the more closely the model’s predictions align with the actual values.
This loss function gives equal weight to each pixel point. A lower standard deviation within the Gaussian mixture model for each dimension correlates with enhanced source resolution. There are only some higher values near the epicenter location, with most values close to 0. Due to the small percentage of pixels in the whole label, if the same weights are used, the loss of the network is small even if the seismic location is incorrectly predicted. This leads to a propensity for predicting non-seismic locations for most pixels, adversely affecting network convergence. Conversely, a larger standard deviation decreases the resolution but increases the convergence. To ensure higher resolution, a weighted mean-variance loss function is used to give higher weights to pixel points near the epicenter, which will be penalized by a larger gradient in case the network predicts incorrectly. The weighted loss function is calculated as follows:
L m a i n = 1 N i = 1 N ω 1 y i t r u e + ω 2 1 y i t r u e y i t r u e y i p r e d 2
where N denotes the sample size, y t r u e denotes the true value, and y p r e d denotes the prediction result.

3.4. Results

Microseismic records from geophones contain not only microseismic waveforms but also varying degrees of noise. Random noise with an SNR of −5, −10, and −15 were added to the training data to simulate the field data. An example of the input data is shown in Figure 10, where it can be seen that as the noise increases, the in-phase axis of the seismic record becomes less and less distinct and the continuity deteriorates. Seismic signals are barely visible at SNR = −15. We trained the network for 600 epochs. Figure 11 shows the loss curve of the network, which achieved convergence in the later stages of training.
S N R = 10 l g P s P n
where P s is the power of the signal, P n is the power of the noise, and lg denotes a logarithm with a base of 10.
After the network is trained, we verify its performance using a test set. To evaluate the proposed algorithm in different SNR scenarios, we add random noise with SNR of −5, −10, and −15 to the microseismic records in the test set, respectively. Figure 10 shows the seismic recordings at different noise levels. We input the test set into the model, with the primary task outputting microseismic locations and the secondary task outputting microseismic records after noise suppression. Figure 12 shows the localization results under different noises, from which it can be seen that the resolution of the source location imaging is high and the model-predicted source location almost exactly matches the real source location. Figure 13 shows the microseismic recordings after suppression noise. Each record is essentially free of noise residue, and the in-phase axis becomes clear and continuous. Since the denoising task is only an auxiliary task and the network focuses on extracting robust features of the microseismic recordings, it is less capable of recovering the local details of the microseismic waveforms, and some erroneous waveforms can be clearly seen below the in-phase axis. In Table 1, the localization effects are summarized for different SNRs.
To show the superiority of the proposed method, we trained a localization model using an ordinary U-Net network. Both the training and test sets use seismic recording data with SNR = −15. Figure 14 shows a sample of events in the test set and their localization results. The localization results of the model did not accurately locate the source position of the single event and disturbances were observed in the surroundings. We randomly selected 400 events in the test set, and the average error distance of their localization results was 2.60 m, with an accuracy of 78% for an error of zero. Compared with the proposed model, the model still has a large gap between the average error distance and accuracy and is less noise-resistant. This suggests that auxiliary tasks are critical for improving the noise immunity of localization models.
To further analyze the predictive ability of the model, we generated microseismic records containing two and three events on the velocity model and added random noise with SNR = −10, as shown in Figure 15 and Figure 16. As shown in the figure, the model can successfully predict the location of the microseismic source, and the noise in the seismic record is well suppressed.

4. Conclusions

We propose a deep learning method for the localization of microseismic sources within noisy data environments. Based on multi-task learning, a decoder responsible for the auxiliary task is added in addition to the main task of seismic source localization, which improves the model’s localization accuracy at low SNRs. Furthermore, CBAMs were integrated into the encoder; CBAM effectively integrates both spatial and channel dimensions of the feature map, which can improve the generalization ability and performance of the model. CA and SE attention modules were also integrated into the decoder, and CA can enhance the localization of microseismic source locations. SE can learn to use global information to selectively emphasize informative features and suppress useless ones, enhancing the denoising capability of the network. We use a Gaussian mixture model to represent source location information for microseisms. To improve the stability and prediction ability of the network, the weighted mean square error is used as the loss function for the main task and different weights are assigned to the losses of the two tasks. The self-supervised assisted denoising task overcomes the difficulty of obtaining noise-free data as labels. Our method can improve the accuracy of seismic recordings with a SNR of −15 from 78% to 93.5% and the average error from 2.60 m to 0.76 m comparing with a normal U-net network. Experimental results show that our method can localize microseismic events with low SNRs and can keep the error within a small range. We plan to introduce real seismic data in future research to further validate the model’s effectiveness and generalization ability in practical applications.

Author Contributions

Conceptualization, Q.L., F.Z. and L.H.; Methodology, Q.L., F.Z. and L.H.; Software, Q.L.; Validation, Q.L., F.Z. and L.H.; Formal analysis, Q.L.; Investigation, Q.L.; Writing—original draft, Q.L., F.Z. and L.H.; Writing—review & editing, Q.L., F.Z. and L.H.; Visualization, Q.L.; Supervision, F.Z.; Project administration, F.Z. and L.H.; Funding acquisition, F.Z. and L.H. All authors have read and agreed to the published version of the manuscript.

Funding

Please add: This research is supported by the National Natural Science Foundation of China (No.42130805, No.42074154, No.41974138).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Shapiro, S.A.; Dinske, C.; Rothert, E. Hydraulic-fracturing controlled dynamics of microseismic clouds. Geophys. Res. Lett. 2006, 33, L14312. [Google Scholar] [CrossRef]
  2. Maxwell, S.C.; Rutledge, J.; Jones, R.; Fehler, M. Petroleum reservoir characterization using downhole microseismic monitoring. Geophysics 2010, 75, 75A129–75A137. [Google Scholar] [CrossRef]
  3. Li, B.Q.; Einstein, H.H. Direct and Microseismic Observations of Hydraulic Fracturing in Barre Granite and Opalinus Clayshale. J. Geophys. Res. Solid Earth 2019, 124, 11900–11916. [Google Scholar] [CrossRef]
  4. Dyer, B.C.; Schanz, U.; Ladner, F.; Häring, M.O.; Spillman, T. Microseismic imaging of a geothermal reservoir stimulation. Lead. Edge 2008, 27, 856–869. [Google Scholar] [CrossRef]
  5. Reshetnikov, A.; Kummerow, J.; Asanuma, H.; Häring, M.; Shapiro, S.A. Microseismic reflection imaging and its application to the Basel geothermal reservoir. Geophysics 2015, 80, WC39–WC49. [Google Scholar] [CrossRef]
  6. Kamei, R.; Miyoshi, T.; Pratt, R.G.; Takanashi, M.; Masaya, S. Application of waveform tomography to a crooked-line 2D land seismic data set. Geophysics 2015, 80, B115–B129. [Google Scholar] [CrossRef]
  7. Shemeta, J.; Anderson, P. It’s a matter of size: Magnitude and moment estimates for microseismic data. Lead. Edge 2010, 29, 296–302. [Google Scholar] [CrossRef]
  8. Eisner, L.; Abbott, D.; Barker, W.B.; Lakings, J.; Thornton, M.P. Noise suppression for detection and location of microseismic events using a matched filter. In SEG Technical Program Expanded Abstracts 2008, Proceedings of the 78th Society of Exploration Geophysicists International Exposition and Annual Meeting (SEG LAS VEGAS 2008), Las Vegas, NV, USA, 9–14 November 2008; Society of Exploration Geophysicists: Tulsa, OK, USA, 2008; pp. 1431–1435. [Google Scholar] [CrossRef]
  9. Sabbione, J.I.; Sacchi, M.D.; Velis, D.R. Radon transform-based microseismic event detection and signal-to-noise ratio enhancement. J. Appl. Geophys. 2015, 113, 51–63. [Google Scholar] [CrossRef]
  10. Li, H.; Wang, R.; Cao, S.; Chen, Y.; Huang, W. A method for low-frequency noise suppression based on mathematical morphology in microseismic monitoring. Geophysics 2016, 81, V159–V167. [Google Scholar] [CrossRef]
  11. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  12. Zhang, C.; van der Baan, M. Microseismic and seismic noise attenuation by supervised deep learning with noisy natural images. In SEG Technical Program Expanded Abstracts 2019, Proceedings of the 89th Society of Exploration Geophysicists International Exposition and Annual Meeting 2019, San Antonio, TX, USA, 15–20 September 2019; Society of Exploration Geophysicists: Tulsa, OK, USA, 2019; pp. 4485–4489. [Google Scholar] [CrossRef]
  13. Saad, O.M.; Bai, M.; Chen, Y. Uncovering the microseismic signals from noisy data for high-fidelity 3D source-location imaging using deep learning. Geophysics 2021, 86, KS161–KS173. [Google Scholar] [CrossRef]
  14. Othman, A.; Iqbal, N.; Hanafy, S.M.; Waheed, U.B. Automated Event Detection and Denoising Method for Passive Seismic Data Using Residual Deep Convolutional Neural Networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5900711. [Google Scholar] [CrossRef]
  15. Dong, X.T.; Li, Y. Denoising the Optical Fiber Seismic Data by Using Convolutional Adversarial Network Based on Loss Balance. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10544–10554. [Google Scholar] [CrossRef]
  16. Dong, X.T.; Lin, J.; Lu, S.P.; Huang, X.G.; Wang, H.Z.; Li, Y. Seismic Shot Gather Denoising by Using Supervised-Deep-Learning Method with Weak Dependence on Real Noise Data: A Solution to the Lack of Real Noise Data. Surv. Geophys. 2022, 43, 1363–1394. [Google Scholar] [CrossRef]
  17. Perol, T.; Gharbi, M.; Denolle, M. Convolutional neural network for earthquake detection and location. Sci. Adv. 2018, 4, e1700578. [Google Scholar] [CrossRef]
  18. Zhang, Q.; Zhang, W.; Wu, X.; Zhang, J.; Kuang, W.; Si, X. Deep Learning for Efficient Microseismic Location Using Source Migration-Based Imaging. J. Geophys. Res. Solid Earth 2022, 127, e2021JB022649. [Google Scholar] [CrossRef]
  19. Wang, H.; Alkhalifah, T. Direct microseismic event location and characterization from passive seismic data using convolutional neural networks. Geophysics 2021, 86, KS109–KS121. [Google Scholar] [CrossRef]
  20. Vinard, N.A.; Drijkoningen, G.G.; Verschuur, D.J. Localizing microseismic events on field data using a U-Net-based convolutional neural network trained on synthetic data. Geophysics 2022, 87, KS33–KS43. [Google Scholar] [CrossRef]
  21. Feng, Q.; Han, L.; Pan, B.; Zhao, B. Microseismic Source Location Using Deep Reinforcement Learning. IEEE Trans. Geosci. Remote Sens. 2022, 60, 4510209. [Google Scholar] [CrossRef]
  22. Zhang, Y.; Yang, Q. An overview of multi-task learning. Natl. Sci. Rev. 2018, 5, 30–43. [Google Scholar] [CrossRef]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015, Proceedings of the 18th International Conference, Munich, Germany, 5–9 October 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Lecture Notes in Computer Science; Springer International Publishing: Cham, Switzerland, 2015; Volume 9351, pp. 234–241. [Google Scholar] [CrossRef]
  24. Yang, F.; Ma, J. Deep-learning inversion: A next-generation seismic velocity model building method. Geophysics 2019, 84, R583–R599. [Google Scholar] [CrossRef]
  25. Alfarhan, M.; Deriche, M.; Maalej, A.; AlRegib, G.; Al-Marzouqi, H. Multiple Events Detection In Seismic Structures Using a Novel U-Net Variant. In Proceedings of the 2020 IEEE International Conference on Image Processing (ICIP), Abu Dhabi, United Arab Emirates, 25–28 October 2020; pp. 2900–2904. [Google Scholar] [CrossRef]
  26. Liu, N.; He, T.; Tian, Y.; Wu, B.; Gao, J.; Xu, Z. Common-azimuth seismic data fault analysis using residual UNet. Interpretation 2020, 8, SM25–SM37. [Google Scholar] [CrossRef]
  27. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Module Attention Module. In Computer Vision—ECCV 2018, Proceedings of the 15th European Conference, Munich, Germany, 8–14 September 2018; Ferrari, V., Hebert, M., Sminchisescu, C., Weiss, Y., Eds.; Springer International Publishing: Cham, Switzerland, 2018; pp. 3–19. [Google Scholar] [CrossRef]
  28. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13708–13717. [Google Scholar] [CrossRef]
  29. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar] [CrossRef]
  30. Graham, B. Fractional Max-Pooling. ar5iv. Available online: https://ar5iv.labs.arxiv.org/html/1412.6071 (accessed on 3 April 2024).
  31. Yin, X.; Goudriaan, J.; Lantinga, E.A.; Vos, J.; Spiertz, H.J. A Flexible Sigmoid Function of Determinate Growth. Ann. Bot. 2003, 91, 361–371. [Google Scholar] [CrossRef]
  32. Martin, G.S.; Wiley, R.; Marfurt, K.J. Marmousi2: An elastic upgrade for Marmousi. Lead. Edge 2006, 25, 156–166. [Google Scholar] [CrossRef]
  33. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv, 29 January 2017. Available online: http://arxiv.org/abs/1412.6980 (accessed on 2 April 2024).
  34. Zhang, X.; Zhang, J.; Yuan, C.; Liu, S.; Chen, Z.; Li, W. Locating induced earthquakes with a network of seismic stations in Oklahoma via a deep learning method. Sci. Rep. 2020, 10, 1941. [Google Scholar] [CrossRef]
  35. Lehtinen, J.; Munkberg, J.; Hasselgren, J.; Laine, S.; Karras, T.; Aittala, M.; Aila, T. Noise2Noise: Learning Image Restoration without Clean Data. In Proceedings of the 35th International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
Figure 1. The architecture of the proposed RES module.
Figure 1. The architecture of the proposed RES module.
Applsci 14 05668 g001
Figure 2. The architecture of the CBAM attention module.
Figure 2. The architecture of the CBAM attention module.
Applsci 14 05668 g002
Figure 3. The architecture of the (a) CAM and (b) SAM attention modules.
Figure 3. The architecture of the (a) CAM and (b) SAM attention modules.
Applsci 14 05668 g003
Figure 4. The architecture of the CCA module.
Figure 4. The architecture of the CCA module.
Applsci 14 05668 g004
Figure 5. The architecture of the CA module.
Figure 5. The architecture of the CA module.
Applsci 14 05668 g005
Figure 6. The architecture of the CSEA module.
Figure 6. The architecture of the CSEA module.
Applsci 14 05668 g006
Figure 7. The architecture of the SE Attention Module.
Figure 7. The architecture of the SE Attention Module.
Applsci 14 05668 g007
Figure 8. The architecture of the proposed MA_W-Net network.
Figure 8. The architecture of the proposed MA_W-Net network.
Applsci 14 05668 g008
Figure 9. Velocity model.
Figure 9. Velocity model.
Applsci 14 05668 g009
Figure 10. Microseismic recordings under different noises: (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Figure 10. Microseismic recordings under different noises: (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Applsci 14 05668 g010
Figure 11. Training loss curve.
Figure 11. Training loss curve.
Applsci 14 05668 g011
Figure 12. Localization effects of microseismic recordings under different noises (black crosses indicate the true source locations): (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Figure 12. Localization effects of microseismic recordings under different noises (black crosses indicate the true source locations): (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Applsci 14 05668 g012
Figure 13. Suppression results of microseismic recordings under different noises: (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Figure 13. Suppression results of microseismic recordings under different noises: (a) Original Record; (b) SNR = −5; (c) SNR = −10; (d) SNR = −15.
Applsci 14 05668 g013
Figure 14. Ordinary neural network, black crosses indicate the true source location: (a) original seismic record; (b) seismic record with added noise; (c) true location; (d) localization results.
Figure 14. Ordinary neural network, black crosses indicate the true source location: (a) original seismic record; (b) seismic record with added noise; (c) true location; (d) localization results.
Applsci 14 05668 g014
Figure 15. Noise suppression and localization results under double events, black crosses indicate the real source locations: (a) original seismic record; (b) seismic record with added noise; (c) suppressed seismic record; (d) localization results.
Figure 15. Noise suppression and localization results under double events, black crosses indicate the real source locations: (a) original seismic record; (b) seismic record with added noise; (c) suppressed seismic record; (d) localization results.
Applsci 14 05668 g015
Figure 16. Noise suppression and localization results under three events, black crosses indicate the real source location: (a) original seismic record; (b) seismic record with added noise; (c) suppressed seismic record; (d) localization results.
Figure 16. Noise suppression and localization results under three events, black crosses indicate the real source location: (a) original seismic record; (b) seismic record with added noise; (c) suppressed seismic record; (d) localization results.
Applsci 14 05668 g016
Table 1. Localization effects for different SNRs.
Table 1. Localization effects for different SNRs.
SNRAverage Error DistanceAccuracyAccuracy (<10 m)Accuracy (<20 m)Accuracy (<50 m)
Original Record0.03 m99.75%100.00%100.00%100.00%
SNR = −50.11 m99.00%99.75%100.00%100.00%
SNR = −100.29 m97.50%99.50%99.75%100.00%
SNR = −150.76 m93.50%98.50%99.75%100.00%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, Q.; Zhang, F.; Han, L. MA_W-Net-Based Dual-Output Method for Microseismic Localization in Strong Noise Environments. Appl. Sci. 2024, 14, 5668. https://doi.org/10.3390/app14135668

AMA Style

Li Q, Zhang F, Han L. MA_W-Net-Based Dual-Output Method for Microseismic Localization in Strong Noise Environments. Applied Sciences. 2024; 14(13):5668. https://doi.org/10.3390/app14135668

Chicago/Turabian Style

Li, Qiang, Fengjiao Zhang, and Liguo Han. 2024. "MA_W-Net-Based Dual-Output Method for Microseismic Localization in Strong Noise Environments" Applied Sciences 14, no. 13: 5668. https://doi.org/10.3390/app14135668

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop