1. Introduction
Oceanic temperature and salinity distributions are major drivers of ocean dynamics such as meridional overturning circulation and oceanic turbulence [
1]. These salinity and temperature distributions within the world’s oceans, in conjunction with atmosphere–hydrosphere interactions, have been related to the recent climatic variations [
2,
3] and have been used in developing models for predicting the frequency, intensity, and severity of future hurricanes [
4]. Additionally, the variations in temperature and salinity due to global climate change and their effects in aquatic lives and controlling the production and migration of fish in marine, lake, and river environments have been thoroughly studied [
5,
6,
7,
8]. Thus, accurately estimating the temperature and salinity distributions within oceans, lakes, and rivers is important for global climate change research in a wide variety of disciplines such as physical oceanography, atmospheric science, marine biology, and the fishery industry.
The primary problem in using temperature and salinity is the accuracy of their measurements [
9]. Oceanic temperature and salinity are measured using eXpendable BathyThermograph (XBT) and conductivity-temperature-depth (CTD) sensors, which can provide a regional temperature–salinity (
T-
S) relationship [
10]. These estimates have a high vertical resolution (~1 m). In general, however, these XBT or CTD sensors are sparsely deployed, and thus, their lateral resolutions can be low [
11]. These vertically high-resolution measurements at sparse lateral locations, when interpolated, may not have high lateral resolutions [
11].
Temperature and salinity variations within water-column structures produce visible reflections on the recorded seismic data, which, in turn, allows for imaging of structures like ocean fronts, mesoscale eddies, and Lee waves [
12,
13,
14,
15,
16]. In addition, dynamic processes such as ocean turbulence [
17,
18,
19] and the movement of water masses [
20,
21] have also been recorded from these reflections. These techniques are sensitive to oceanic sound-speed variations [
22] and can be applied to both two- and three-dimensional (2D and 3D) offshore seismic data [
23].
Besides analyzing seismic images, extracting the thermohaline nature of oceans from seismic reflections has been attempted by many, which resulted in the emergence of a new field known as Seismic Oceanography. For example, it has been verified that the temperature contrasts in ocean waters can explain the amplitude-variation-with-angle (AVA) behavior of water-column seismic reflections [
24]. Extracting accurate velocities (sound-speeds) using post-stack, pre-stack, and stochastic inversion methods to estimate temperature and salinity have also been attempted [
25,
26,
27,
28,
29,
30,
31,
32,
33,
34,
35,
36,
37,
38,
39,
40]. Most of these inversions allow for extraction of the high-frequency sound-speed variations within water-columns but require the low-frequency (background) trends to be available from other sources such as XBTs or CTDs. Because it is impossible to deploy many XBTs or CTDs, accurately estimating the background sound-speed variations is difficult. Therefore, estimating the broad-band sound-speed structures of the world’s oceans and relating them to the temperature and salinity using these methods is not an easy task. The main issue with these inversions is that most of these methods use a linearized (gradient-based) local optimization. Padhi et al. (2015) [
11] demonstrated that, when a good background model is available, these local methods work very well. However, using a global optimization method, such as a genetic algorithm (GA), can obtain reliable sound-speed models even when such background trends are unavailable. Additionally, Padhi et al. (2015) [
11] also showed how these estimated sound-speeds can be combined with the equation of state [
41,
42,
43,
44] and make reliable temperature and salinity profiles. These seismically derived temperature and salinity estimates are of lower vertical resolution (~5 m) than those obtainable from XBT or CDT data. However, they are of high lateral resolution (25 m or less) [
11], which can potentially open entirely new avenues for physical oceanography, climate, marine biology, and fishery research.
Although the sound-speed can be accurately estimated from marine seismic data, using which the temperature and salinity can also be estimated, it is difficult to use them on a routine basis. Reasons for this have been thoroughly studied elsewhere [
45]. For completeness, we briefly discuss these reasons and introduce the motivation of our current research.
Figure 1 summarizes one of the current practices of estimating the sound-speed from the pre-stack seismic data [
11]. Because of their inherent non-uniqueness and computationally demanding nature [
46], seismic inversions require an initial model. The primary reason for this requirement is to ensure that the inversion finds a meaningful, not an undesired, solution. For sound-speed estimation, if XBT and CTD data coincide with the seismic data, then an initial sound-speed can be estimated using the equation of state [
41,
42,
43]. However, XBT/CTD may be available only at sparse locations or may not be available at all. To resolve this issue, Padhi et al. (2015) [
11] ran pre-stack GA inversion [
47,
48,
49] at a few sparse locations and estimated the sound-speed for those locations. In the real stacked seismic data from the South China Sea, shown in
Figure 1a, these locations are indicated by arrows. After estimating sound-speeds at sparse locations, their low-frequency estimates were interpolated along three horizons, which Padhi et al. (2015) [
11] manually interpreted from the stacked seismic data. These interpreted horizons are shown as green lines in
Figure 1a and the sound-speed model from interpolation is shown in
Figure 1b. Using this interpolated model as the initial model for each common midpoint (CMP) location, Padhi et al. (2015) [
11] ran prestack waveform inversion using both the GA and nonlinear least square (NLS) methods and estimated the sound-speed for the entire range of the data, and
Figure 1c is the GA inversion result.
Horizon-guided interpolation for initial model generation (
Figure 1b) is a standard practice in seismic inversion. For inversion to produce meaningful results, such an interpolation is considered necessary [
21,
50]. However, if we compare
Figure 1b,c, it is obvious that such horizon-guided interpolation tends to bias the inverted model by leaving the imprint of the originally picked horizons in the final model. If additional horizons are picked, for example, the ones shown as dashed black lines in
Figure 1a, then the initial model and the final model will be different from those shown in
Figure 1b,c. Not only does this bias depend upon the number of horizons, but it is also likely to vary from person to person, based on the way they are being picked from the stacked seismic data. In addition, although many sophisticated automated horizon interpretation procedures are available, including the machine learning-based methods, none of them are error-free and any picking error is likely to be propagated in the inverted sound-speed.
In
Figure 1, we show a 2D case where horizon-picking is straightforward. For 3D seismic data, manually picking horizons for the entire data volume is impossible. Therefore, the horizons are picked at regular inline or crossline intervals and then spatially interpolated to obtain horizon surfaces [
51]. This spatial interpolation is not error-free and any error is likely to be propagated in the inversion results. Consequently, finding a way to obtain a bias-free (and error-free) inversion where no horizon picking is necessary is desirable [
52].
Besides horizon-guided initial model generation, running GA inversion at sparse locations is computationally challenging. To generate a consistent model for the entire data range, these inversions must be carefully run, which involves a lot of expertise and human labor. In the example shown in
Figure 1, these sparse inversions were run only at five locations. For large 3D data volumes, a lot more such inversions involving additional human labor and computation time would be necessary, the avoidance of which is desirable.
Irrespective of the method used for seismic inversion, the goal of seismic oceanography is to map the inverted sound-speeds onto the temperature and salinity. To simultaneously estimate the temperature (
T) and salinity (
S) from the inverted sound-speeds (
V) requires calculating pressure as a function of depth and latitude [
44]. Then, starting with an initial salinity model, it must be iteratively modified using an inversion (optimization) until it is reasonably matched with the regional
T-
S relationship of the CTD data from the Levitus database [
10]. This step is also another computationally challenging task and requires a regional
T-
S relationship, which may not always be available. Therefore, it is preferable that we find an alternate, automated
V to
T-S mapping methodology.
For estimating the sound-speed, a data-driven, attribute-guided method to generate the initial model has been recently developed, which does not require any separate inversions at sparse locations or horizon interpretation [
45]. This method is therefore neither subject to the human errors and biases involved in horizon-picking, nor it is subject to the additional computational expense and human labor caused by sparse inversion runs [
45]. In this work, we outline a two-step approach for estimating
and
from seismic data. In the first step, we estimate
using the data-driven, attribute-guided inversion of the marine seismic data [
45]. In step 2, we develop an inversion method using the Adaptive Moment (ADAM) optimization [
53] and use it to estimate
and
from
.
In the following, we first outline our two-step approach along with the results from the South China Sea seismic data. Next, we discuss results and the possibilities for future improvements by converting our two-step approach into a single step using machine learning. Finally, we end with some concluding remarks.
2. Methodology and Results
Our two-step approach consists of (1) data-driven and attribute-guided inversion of seismic data for estimating the sound-speed and (2) estimation of the temperature and salinity from the inverted sound-speed from step 1. Below, we separately discuss each of them.
Step-1, data-driven, attribute-guided inversion for sound-speed estimation: Details of the data-driven and attribute-guided inversion can be found elsewhere [
45]. Here, we briefly summarize the fundamental steps along with the results from the South China Sea marine seismic data.
The first step in the data-driven, attribute-guided approach is estimating the initial velocity (sound-speed). Instead of running sparse inversions, we perform velocity analysis (VELAN) at every 50th CMP location and then laterally interpolate them on a sample-by-sample basis, not guided via any interpreted horizons as an initial estimate of the sound-speed.
Figure 2a is the sound-speed estimated from VELAN and interpolation. Comparing
Figure 2a with the initial sound-speed model shown in
Figure 1b, we can see that the general values in both are similar, with high values (white-to-red zone with 1.55–1.51 km/s sound-speed) for ~0–0.6 s, moderate values (yellow-green zone with 1.51–1.48 km/s sound-speed) for 0.6–1s, and low values (blue zone with sound-speed less than 1.48 km/s) below 1s. However, there is no apparent horizon-induced bias in the sound-speed shown in
Figure 2a like that shown in
Figure 1b. Following VELAN and lateral interpolation, we apply the geometrical spreading compensation (GSC) to the pre-stack gathers, correct them for the normal moveout (NMO), and stack them.
Figure 2b shows the stacked data. Because VELAN, GSC, NMO, and stacking are standard and automated procedures in seismic data processing [
54], this step is fully automated and not prone to any bias or errors due to horizon-picking. In addition, because no individual GA inversions at sparse locations are required, the process does not involve much computational expense or human labor, and is therefore applicable to 3D applications.
Following stacking, we compute the time-domain attribute
where
is the time from the stacked seismic data as
In Equation (1), is the stacked seismic data and is computed from using the following steps:
Forward Fourier transform to where is the frequency;
Interchange the real and imaginary parts of and obtain . Thus, for any sample , if of is , then the corresponding sample of is , where ;
Inverse Fourier transform and take its real part as .
Thus, the attribute
is directly computable from the stacked data shown in
Figure 2b.
Figure 3 is the attribute-guided approach to generate the initial model for inversion.
Figure 3a is the cross-plot between the attribute
on the horizontal axis and the VELAN and interpolation-estimated sound-speed on the vertical axis for the same five locations, indicated by arrows in
Figure 1a, where GA inversions were run to generate the initial model of
Figure 1b with the plotted points shown as black dots (circles). The red curve shown in this figure is the best-fit polynomial between the points, which gives the sound-speed
as a function of
, as
in which
is in m/s. The
value in the
fit given by Equation (2) is 0.8741. Using Equation (2), we compute
from
for the entire range of the seismic data, and it is shown in
Figure 3b. Like
Figure 1b,c and
Figure 2a, the attribute-guided sound-speed model shown in
Figure 3b is also represented by the high (white to red), moderate (yellow to green) and low (blue) sound-speed zones at the similar time intervals.
Using the sound-speed shown in
Figure 3b as the initial model, we run the GA-based pre-stack waveform inversion and show the inverted model in
Figure 4 in both the time (
Figure 4a) and depth (
Figure 4b) domains. Notice that, although we did not provide any horizon information in the initial model, the inverted sound-speed model (
Figure 4) is in perfect agreement with the reflection boundaries in the stacked seismic data (
Figure 1a and
Figure 2b), as if each of these reflection boundaries were separately picked as horizons and the initial model was guided through them. By computing wave equation-based pre-stack synthetic seismic data using the depth-domain model shown in
Figure 4b and for the same range of source-to-receiver offsets as they were in real data, followed by GSC and NMO correction and stacking, it was further verified that the synthetic stacked data matched almost perfectly with the real stacked data, giving additional confidence in the reliability of the estimated sound-speed model shown in
Figure 4 [
45]. Although the inverted sound-speed model obtained earlier by the horizon-guided approach [
11], shown in
Figure 1, produced synthetic stacked data which also matched real stacked data with equal accuracy, we must note that the latter results are biased by the picked horizons while the attribute-guided results are not. We must also note here that both horizon- and attribute-guided inversions require an initial density model in addition to the sound-speed. Density is an important parameter in physical oceanography. However, the fundamental basis for our analysis is seismic inversion, or, to be specific, quantifying the variations in the seismic reflection amplitudes as a function of the angle-of-incidence. In an earlier work, Mallick and Chakraborty (2022) [
45] carried out a detailed investigation on the sensitivity of these water-column reflection amplitudes to the variations in the sound-speed and density. Their analysis revealed that the relative variations in these amplitudes are primarily controlled by the sound-speed, with density providing a bulk shift to the amplitudes without affecting their relative values. Considering that all seismic inversion methods are based on quantifying the relative amplitude variations, not their absolute values, we thus conclude that the density cannot be accurately estimated from the seismic data. Therefore, within the observed range of the water-column density variations which is approximately 1030–1070 kg/m
3 [
11,
45], we assumed a constant density of 1050 kg/m
3 for both inversions shown in
Figure 1 and
Figure 4, and in all our subsequent analyses.
Step-2, simultaneous estimation of the temperature and salinity from the sound-speed: To estimate the temperature and salinity from the sound-speed, we first define our observed data as an
-component sound-speed vector
, given as
and the model to be estimated as
-component model vector
, given as
Clearly, in Equations (3) and (4) is the number of depth samples at each CMP location, with , , and being, respectively, the sound-speed, temperature, and salinity for the depth sample. Given the sound-speed vector from the seismic inversion for each CMP location (Equation (3)), our goal is to estimate the -component model vector (Equation (4)), in which the first -components are the temperatures and the next -components are the salinity for those CMP locations.
Estimating the
-component model vector
of the temperature and salinity from the observed
-component data
obtained from seismic inversion is an optimization problem. To solve this problem, we used the Adaptive Moment Estimation method (ADAM) [
53]. The ADAM is an iterative optimization algorithm used to minimize the loss function during the training of neural networks and became popular in the use of deep learning for adjusting the learning rate for each parameter in a model [
55]. Although primarily used in machine learning, the ADAM is well-suited for any problem where the data and model sizes are large and the objectives are non-stationary with sparse or noisy gradients [
55]. Because our problem of mapping the seismic sound-speed onto temperature and salinity falls in the same category, we chose the ADAM for optimization.
In our implementation, we first define our forward problem that each
of Equation (3) is related to the
and
of Equation (4) and pressure
where
via the equation of state, given as
Exact expressions for
,
,
, and
in Equation (5) can be found elsewhere [
41,
42,
43]. For completeness, we also provide them below in
Appendix A. Note that our forward problem not only requires the temperature (
) and salinity (
) but also the pressure (
) for each depth sample
. To iteratively update the temperature (
) and salinity (
), it is first necessary to discuss how to calculate the pressure (
) in Equation (5).
The pressure as a function of depth (
) and latitude (
) is given as
In Equation (6),
is the mean atmospheric pressure at sea level (101,325 Pa or 1.01325 bars),
is the acceleration due to gravity at latitude
and depth
in m/s
2, and
is the density of seawater in kg/m
3. The exact value of
in Equation (6) can be computed using the equations for the free-air and Bouguer corrections used in the gravity exploration [
56] and can be written as
in which
is the acceleration due to gravity at the sea surface (
) and latitude
, given by the international gravity formula [
57]:
where
is the sea-surface acceleration due to gravity at the equator (
) and is given as 9.7803185 m/s
2. Equations (6)–(8) are the exact equations to calculate depth- and latitude-dependent pressure. An earlier analysis, however, has shown that assuming a constant
m/s
2 and a constant
kg/m
3 in Equation (6) is adequate for computing
and using depth-varying
and both latitude- and depth-varying
(Equations (7) and (8)) is not necessary [
45].
Now that we know how to calculate pressure, the next step is to define an initial model of temperature and salinity and implement the iterative model-updating procedure based on the ADAM. Each iteration step of the ADAM is denoted as the time step , and below we describe how we generate the initial model and the model updating procedure from time step to time step .
Ideally, an initial model of temperature and salinity should be obtained from the XBT and CTD data. As previously discussed, the traditional approach to simultaneously estimating
T and
S from the inverted sound-speeds (
V) requires the following steps [
11]:
Although the traditional approach outlined above worked well on the South China Sea data [
11], we chose to use an initial model that would not require matching with a regional CTD database. We have already demonstrated that the depth-dependent pressure can be calculated from Equation (6). And
Figure 5,
Figure 6 and
Figure 7 illustrate our approach to obtaining an initial model of
and
.
Figure 5a,b shows the
(in
) and
measurements from XBT and CTD deployments in the Gulf of Mexio area that are available in the public domain. Using these measurements, calculating
from Equation (6) with
kg/m
3 and
m/s
2, we then used Equation (5) to calculate
, which we show in
Figure 5c (in km/s), illustrating that reasonable values of
can be computed from
and
. Next, we attempted to express
and
, shown in
Figure 5a,b as functions of
, which is shown in
Figure 5c. To do so, we cross-plotted both
and
as the vertical axes and
as the horizontal axis and obtained best-fit polynomials through the plotted points. These cross-plots (in black dots) and best-fit polynomials (in red curves) are shown in
Figure 6a,b, expressing
and
as
It must be noted from
Figure 6 that
is more accurate than
. In addition, while the prediction of
is unambiguous, there is an ambiguity in the prediction of
, as the same value of
can be predicted from two different values of
(see
Figure 6). This ambiguity is due to the nature of the temperature and salinity data that we obtained from the Gulf of Mexico. Both
and
decrease with depth (see
Figure 1,
Figure 2,
Figure 3,
Figure 4 and
Figure 5). On the other hand,
increases from the surface to a depth of about 200 m and then decreases (see
Figure 5). Although these variations are uncommon, they were the only data we could find for our initial model. We will further discuss this issue in a later part of this paper.
To further illustrate the accuracy of
and
, shown by Equation (9), in
Figure 7, we show the quality of our prediction.
Figure 7a,b compares the predicted
and
in red with the original (observed)
and
of
Figure 5a,b in black. And
Figure 7c compares the sound-speeds. The black curve of
Figure 7c is the same sound-speed shown in
Figure 5c that we computed from the observed
and
(
Figure 5a,b). On the other hand, we computed the red curve of
Figure 7c from the predicted
and
(red curves of
Figure 7a,b). Notice that, despite less accuracy and ambiguity in the prediction of
, the sound-speeds can be predicted with reasonable accuracy. Therefore, from the seismically inverted sound-speeds, using Equation (9) will provide a reasonably good initial model of
and
for the ADAM optimization.
In our implementation of ADAM, we first define our model at any time step
as
Clearly, for
, our model
is the initial model, which we compute from the inverted sound-speeds using Equation (9) for each depth sample
. Having defined
as above, we then use Equation (5) to compute the predicted sound-speed
for time step
, given as
Having computed
, we then compute the objective as
In Equation (12), is the true sound-speed which we obtain from seismic inversion. In general, additional regularization parameters or weighting matrices are used in between the first and second terms on the right-hand side of Equation (12). The purpose of using these weighting matrices is to stabilize the optimization. In our implementation, however, we did not use them because these regularization parameters are built in to the ADAM optimization method, which we discuss below.
Finding the optimal model
, with its components being the temperature and salinity values (Equation (4)) that predict the observed set of sound-speeds
(Equation (3)), is a classical least square minimization problem that requires iteratively minimizing the objective
. The ADAM algorithm achieves this minimization by using both the gradient of
and its square. Although the details of the methodology can be found elsewhere [
53], for completeness, we outline below the ADAM methodology for estimating
and
from
.
We first compute the gradient of the objective, given as
Note that
is a vector of
-components containing the partial derivatives of each component of
(Equation (11)) with respect to each component of
(Equation (10)). For general optimization problems, these partial derivatives are numerically computed. However, for our problem of mapping
onto
and
, these derivatives can be analytically computed, and in
Appendix B, we derive the explicit forms of these derivatives and provide the analytical expression of
.
After obtaining
, we update the biased first order moment
for the next time step from that of the current time step (
) as
Similarly, we update the biased second order moment
from
as
Like
,
,
,
, and
are
-component vectors,
in Equation (15) is a component-wise square of
. The parameters
and
are regularization parameters. For machine learning-related applications, well-tested default settings for these parameters are
and
[
53]. For our implementation also, these default settings worked very well. Also, note from Equations (14) and (15) that both
and
are updated from
and
; their corresponding values at the previous time step. In practice, for time step zero,
and
are set to null (zero) vectors so that Equations (14) and (15) can be used for subsequent updating. After updating the biased first- and second-order moments
and
, we then compute the bias-corrected first- and second-order moments
and
as
and
Finally, we obtain the updated model
from
The parameter
above is the step size and
is another regularization term. For machine learning applications, the well-tested default values are
and
[
53], and we used these default values in our implementation. Based on this discussion,
Figure 8 is our implementation of the ADAM optimization to estimate
and
from
, which is estimated from seismic inversion.
As shown in
Figure 8, we start with the loop over the CMP location. For each CMP, we first read the inverted sound-speed
for all
depth samples. We then set the time step
and initialize the
-component first- and second-order moments
and
to null vectors. We also use Equation (9) to compute the
-component initial model vector
. Next, we compute the objective
. To compute
, we first compute
from
using Equations (5) and (6), and additional equations provided in
Appendix A. After computing
, for time step zero, we enter the main ADAM optimization modules, shaded in light gray in
Figure 8. First, we use Equation (13) and additional equations provided in
Appendix B to compute
, the gradient of the objective. Next, we use Equations (14) and (15) to update the biased first- and second-order moments
and
for the next time step
from their values at the current time step. Following this, we compute the unbiased moments
and
from Equations (16) and (17) and use Equation (18) to update the model
from
. After obtaining the updated model for the next time step, we must decide whether to continue iterating or whether the updated model is sufficiently close to the true model so that we can exit, output the estimated model, and move onto the next CMP location. To do so, we compute the objective
using the same procedure outlined above and check for convergence. The primary criterion for this convergence check is to find out whether the sound-speed
, predicted from
, is sufficiently close to
. If it is, we exit iterating, output the model
to be our best estimate of the temperature and salinity values, and move onto the next CMP location. On the other hand, if
, predicted from
, is not sufficiently close to
, we continue iterating by setting
as our current time step
. In our implementation, we set this convergence check using two criteria: (1) the maximum number of iterations allowed and (2) the minimum root mean square error (RMSE) between
and
to be achieved. If either one of these two criteria is satisfied, then the method exits iterating, and otherwise, it continues iterating.
We first tested the validity of our method for a single location. We extracted the sound-speed values for CMP 9350, situated in the middle of
Figure 4b, and used the methodology described by
Figure 8 to estimate the temperature and salinity from the extracted sound-speed values.
Figure 9 and
Figure 10 show the results of this estimation. Starting from the initial values computed using Equation (9) (black curves in
Figure 9a,b), we ran the ADAM optimization, and the final values of our estimation of temperature and salinity are shown as red curves in
Figure 9a,b. To validate the accuracy of our estimation, in
Figure 10 we compare the true sound-speed at CMP 9350 with the sound-speeds computed from Equation (5), using both the initial model and the final model. The true sound-speed is shown in red, and those computed from the initial and final models are shown, respectively, in black and green in
Figure 10. Notice that the sound-speed computed from the final model falls right on top of the true sound-speed so closely that the true sound-speed in red is not visible in
Figure 10. This comparison validates that our method can find the temperature and salinity from the sound-speed with a very good accuracy.
After validating the method at a single location, we applied our method to the entire domain of the inverted sound-speed data, shown in
Figure 4b and
Figure 11, which show the results of our estimation.
We further validate our estimation by computing the RMSE between the true sound-speed model of
Figure 4b and the predicted sound-speed model computed from the initial and final estimated models. While the average RMSE between the true and initial models was about 12%, that between the true and inverted models was less than 0.001%. We also computed the sound-speed model using the final estimated model shown in
Figure 11 and the equation of state (Equation (5)), and compared that with the true sound-speed model of
Figure 4b. This predicted sound-speed was almost identical to the sound-speed model of
Figure 4b, with a maximum difference of 0.0025 m/s between them, and is therefore not shown.
3. Discussion and Future Possibilities of Improvement
Our two-step approach to estimate the temperature and salinity from seismic data worked very well. The predicted sound-speeds computed from our estimated values of temperature and salinity match almost perfectly with their true values. We also compared our estimates with previous estimates [
11] and verified that they reasonably match with one another. Primary advantages of our approach compared with the method used in these previous estimates [
11] are its ease of use and computational efficiency. The attribute-guided approach to generate the initial model is fully automated and can even be obtained via shipboard processing. Although GA inversion is computer-intense, if shipboard computers are connected to a high-performance computing facility, then estimating sound-speeds is also feasible. Our proposed ADAM optimization is also computationally efficient. Estimating the temperature and salinity for the entire range of the 992 CMP locations shown in
Figure 11 took about one hour to complete in a Linux workstation. Additionally, we did not make any attempt to parallelize our ADAM methodology. It is, however, clear from
Figure 8 that the estimations of temperature and salinity for each CMP location are independent of one another. Consequently, it is not difficult to modify the method to simultaneously run on different CMP locations in a parallel computing environment, which can further enhance its computational efficiency. Therefore, with network connectivity to a high-performance computing facility, it may be possible to obtain real-time shipboard estimates of temperature and salinity using our method.
It is also important to point out the derivation of the initial models that we used in our estimation of temperature and salinity. From a few measured values from the Gulf of Mexico, we derived the
and
relations (Equation (9)) and used them to obtain our initial temperature and salinity models from the seismic sound-speeds from an entirely different region of the world. And yet, starting from this initial model, our method successfully obtained reasonable estimates of temperature and salinity. We previously pointed out, while discussing these initial models, that there is an ambiguity in our
relation because the same value of
is obtainable from two different values of
(see the red curve in
Figure 6b). To investigate how this ambiguity could affect our estimates, we carried out additional sensitivity analyses and show them in
Figure 12. As shown in this figure, we computed the six sets of sound-speed models using Equations (5) and (6) as follows:
Using the measured
and
values of
Figure 5a,b (solid black curve). Note that this is same as the sound-speed model shown in
Figure 5c;
Using the measured
values of
Figure 5a and a constant value of
(solid red curve);
Using the measured
values of
Figure 5a and a constant value of
(solid green curve);
Using the measured
values of
Figure 5b and a constant
(dashed blue curve);
Using the measured
values of
Figure 5b and a constant
(dashed black curve);
Using the measured
values of
Figure 5b and a constant
.
Several inferences can be drawn from the results shown in
Figure 12, such as:
is more sensitive to the variations in than ;
For a fixed , increases with ;
For a fixed , increases with ;
It is that primarily controls the shape of how varies with depth, and only provides bulk-shifts to their estimates.
Because
controls the shape of
and
only provides a bulk-shift, despite the ambiguity in the estimation of
shown in Equation (9) and
Figure 6b, we believe that our results are accurate. We must, however, note that, because
is more sensitive to the variations in
than those in
, the uncertainty in our estimation of
is likely to be higher than that of
. Additionally, although using an initial model derived from data from the Gulf of Mexico worked well on the data from the South China Sea, this does not necessarily imply that this model can be used as the initial model to invert any marine seismic data. The primary motivation of our current work is to demonstrate the applicability of our two-step method. In practice, the initial model must be site-specific and should be preferably derived from the XBT and CTD. Quantifying the uncertainties of our
and
estimations and the applicability of our two-step process to other marine seismic data is currently being investigated.
Finally, above, we argued that it is feasible to apply our proposed method shipboard such that
and
could be estimated in real time. This possibility, however, requires network connectivity to a high-performance computing facility. Although feasible, we must point out that GA inversion is computationally demanding [
11,
45]. Replacing GA with a more computationally efficient gradient-based local method can improve computational efficiency. However, due to the inherent non-uniqueness of the seismic inverse problems, it has been demonstrated that the GA provides much superior sound-speed estimates to the local methods and should be the preferred choice [
11]. Therefore, seeking means to further improve the computational efficiency of the method is desirable. In recent years, the use of artificial intelligence (AI) and specifically machine learning (ML) methods revolutionized almost all fields of science and engineering. We believe that physics-informed machine learning (PIML) [
58,
59,
60,
61,
62,
63,
64,
65,
66,
67,
68,
69], a recent advancement in supervised learning for ML, can modify our two-step approach to a single step and further improve the computational efficiency. The standard practice for the supervised learning for ML is to use the labelled data for training and adjusting the network weights. Instead of using labelled data, PIML utilizes the underlying physics of the problem as the primary network driver. By using the underlying physics, PIML has been demonstrated to work very well for seismic inverse problems, in terms of improving both computational efficiency and the accuracy of the estimated model as compared with a standard ML approach using the labelled data for supervised learning [
62]. In
Figure 13, we provide our vision of how PIML can modify our two-step approach into a single step and improve computational efficiency. By feeding the seismic data at selected locations from the input layer into a deep learning architecture, we will directly predict the temperature and salinity. Then, by computing pressure from Equation (6), we will use Equation (5) to compute the sound-speed model. Using this sound-speed model and assuming a constant density, we will then use the physics of the problem (wave equation-based method) and compute synthetic (predicted) seismic data. For learning, we will then iteratively adjust the network weights by matching these synthetic seismic data with the true seismic data until convergence. Once learning is complete, we can then use our trained network to directly predict the temperature and salinity from the entire seismic data. Although connectivity to a high-performance parallel computing facility would be necessary for training (learning), we expect that this PIML approach can improve the overall computational efficiency and be used for the real-time shipboard prediction of temperature and salinity directly from the acquired seismic data. Implementation of the overall concept of
Figure 13 is currently being investigated and will be discussed in a separate paper.