1. Introduction
Inverse scattering problems arise in many areas such as medical imaging and seismic detection. The scattering from obstacles leads to exterior boundary value problems for the Helmholtz equation. The two-dimensional case is derived from the scattering from finitely long cylinders, and more importantly, it can serve as a model case for testing numerical approximation schemes in direct and inverse scattering [
1]. In this paper, we consider the inverse problem to reconstruct an obstacle, given partial far-field data in
.
Let
be a bounded domain with a
-boundary
. Denoting by
the incident plane wave, the direct scattering problem for a sound-soft obstacle is to find the scattered field
such that
where
is the wavenumber. It is well known that
has an asymptotic expansion
uniformly in all directions
. The function
defined on
is the far-field pattern of
for
D due to the incident field
. The inverse problem of interest is to reconstruct
D from
n far-field patterns
for one incident wave. Note that the unique determination of
D using the far-field pattern due to one incident wave is an open problem.
Classical inverse scattering methods to reconstruct an obstacle mainly fall into two groups. The first group is the optimization methods to minimize certain cost functions [
2,
3]. These methods usually solve a number of direct scattering problems and work well when proper initial guesses are available. The second group is the sampling methods, e.g., the linear sampling method, the extended sampling method, the orthogonality method, the reverse time migration, and the direct sampling method [
4,
5,
6,
7,
8]. These methods often construct some indicator functions based on the analysis of the partial differential equations to determine if a sampling point is inside or outside the obstacle. In contrast to the optimization methods, the sampling methods do not need initial guesses and forward solvers but usually provide less information of the unknown obstacle.
Although successful for many cases, classical methods have difficulties for problems with partial data. For example, the linear sampling method cannot obtain a reasonable reconstruction if the data are collected on a small aperture. Recently, there has been an increasing interest in solving inverse problems using deep neural networks (DNNs). Various DNNs have been proposed to reconstruct the obstacle from the scattered field [
9,
10,
11,
12]. Although the deep learning methods can process partial data, the performance is highly problem-dependent and the network structures are often sophisticated. Combinations of deep learning and sampling methods for the inverse scattering problems have also been investigated recently [
13,
14,
15,
16]. These methods usually construct a DNN to learn the reconstruction of some sampling method.
In this paper, we propose a DNN-oriented indicator method. The idea is to design an indicator function which is simple enough for standard DNNs to learn it effectively. We borrow from the sampling methods the concept of the indicator. What sets our method apart from the sampling methods is that the indicator is defined as a signed distance function other than using the inner product related to Green’s function or solving the linear ill-posed integral equations. The indicator is then learned by a DNN with partial scattering data as input. This DNN-oriented indicator method inherits the advantage of deep learning methods by compensating for partial information through empirical data. Meanwhile, it retains the simplicity and flexibility of sampling methods. Numerical experiments demonstrate its effectiveness for obstacle reconstructions with a few far-field data due to one incident wave. Note that the idea is to find a formulation for the inverse problem that is easy to learn instead of the design of sophisticated neural networks (see [
17]).
The rest of this paper is organized as follows. In
Section 2, the DNN-oriented indicator method is proposed, and the data structure is specified. In
Section 3, numerical experiments are carried out to demonstrate the performance of the proposed method. Conclusions are given in
Section 4.
2. Dnn-Oriented Indicator Method
We propose a DNN-oriented indicator method for the inverse obstacle scattering problem in
. The framework can be directly applied to other inverse problems and extended to
. For an obstacle
D, we define a signed distance function
It is clear that
is a continuous function with respect to
and can be used to reconstruct the obstacle since
Let
be the vector of far-field data due to one incident wave in
n observation directions. Let
be a (large) domain, known as a prior, containing
D. For a point
, we define the indicator function as
Our goal is to design a DNN to learn
for input
. More specifically, we expect to build a neural network
to approximate the indicator function
I. The input of
is a vector in
consisting of the real and imaginary parts of
and the coordinates of a point
. The output is the corresponding signed distance. Assume that there are
L hidden layers for
. Each hidden layer is defined as
where
and
are the input and output vectors of the
lth layer, respectively,
is the output weight matrix, and
is the bias vector. The non-linear function
is the rectified linear unit (ReLU) defined as
. The input layer with output weight matrix
and bias vector
is connected to the first hidden layer.
A set of
N data is used to train the fully connected network
. The set of parameters is denoted by
and is updated by the backpropagation algorithm, minimizing the half quadratic error of the predicted responses between the set of training mapped measurements
and exact values
. The loss function is the half mean-squared error given by
The algorithm for the DNN-oriented indicator method is illustrated in Algorithm 1. It contains an offline phase and an online phase. In the offline phase, the training data are used to learn the network parameters. Then, the network
is used in the online phase to output the indicator for each
. Using the same notation
for the function produced by the trained network
,
indicates that
, whereas
indicates that
. The approximate support of
D can be reconstructed accordingly.
Algorithm 1 DNN-oriented indicator method. |
offline phase - 1:
Generate far-field vectors for random objects in - 2:
Generate points for - 3:
Collect training data and - 4:
Construct the DNN - 5:
Train using
online phase - 6:
Measure far-field data for unknow object D - 7:
Generate a set T of uniformly distributed points for any subset of which contains D - 8:
For each , use as the input for to predict the indicator - 9:
Approximate D using
|
Remark 1. Other indicator functions, e.g., the characteristic function for D, can be used and have similar performance. The advantage of the indicator function (3) is that it provides information on the distance of a point to the boundary of the obstacle. Remark 2. The set Ω in the offline phase and online phase can be different, which allows additional flexibility.
3. Numerical Experiments
We present some numerical examples to demonstrate the proposed DNN-oriented indicator method. All the experiments are conducted on a laptop with an Intel(R) Core(TM) i7-10510U CPU @ 1.80 GHz (up to 2.30 GHz) and 16 GB of RAM.
The wave number is
, and the incident plane wave is
with
. The training data are generated using a boundary integral method for (
1) [
1]. The observation data
contain the far-field patterns of 6 observation directions
uniformly distributed on one-quarter of the unit circle.
For the offline phase, we use
random obstacles
which are star-shaped domains [
18]:
where
In the experiments, and the coefficients , , are uniformly distributed random numbers. We use points uniformly distributed in the region . The training data consist of for each obstacle and each sampling point as input and the signed distance function at for as output. The size of the training set is .
The neural network is fully connected feed-forward with
layers and the numbers of nodes in each layer are
, respectively, where the number of hidden layers
L and the number of neurons in each hidden layer
are chosen using the rule of thumb [
19] and a trial-and-error method. The network is trained for up to
iterations using a mini-batch size of 300 and 4 epochs. We normalize the input per feature to zero mean and standard deviation one, and use the adaptive moment estimation (Adam) with an initial learning rate of
, gradually reducing the learning rate by a factor of
every 3 epochs. When
and
, the training duration is approximately 13 min and 57 s, while the testing duration is less than 1 s.
To evaluate the effectiveness of the DNN
, we generate a test set of size
. The following relative error for the test set is used:
We test the network with different
L and
and show the relative errors utilizing noiseless far-field data and far-field data with
random noises in
Table 1. For each
L, the upper row is for the noiseless data and the lower row is for noisy data. Since the errors are smaller for
and
, we use them for
in the online phase.
Remark 3. The results indicate that the inverse operator only needs a simple network with reasonably small L and . More sophisticated networks negatively affect the performance.
Remark 4. The complexity of the neural network depends on various elements, including the number and locations of far-field data, the wavenumber k, etc. One would expect different and L for different settings.
To visualize the reconstructions, we use three obstacles: a triangle with boundary given by
a peanut with boundary given by
a kite with boundary given by
and a square whose vertices are
Let the uniformly distributed point set
T for
be
For each , the trained DNN is used to predict . We plot the contours for with respect to , and obtain the reconstructions by finding all such that .
For the triangle obstacle,
Figure 1 shows the contour plot of the indicator function
using 6 far-field data with
random noise, and the reconstruction of
D formed by points
satisfying
. Similar results for the peanut, kite and square are shown in
Figure 2,
Figure 3 and
Figure 4, respectively. The location and size of the obstacle can be reconstructed well, taking the amount of data used into account.
4. Conclusions
We consider the reconstruction of a sound-soft obstacle using only a few far-field data due to a single incident wave. In addition to the inherent nonlinearity and ill-posedness of inverse problems, the presence of partial data introduces additional challenges for classical methods. In this paper, we propose a simple and novel DNN-oriented indicator method for this inverse problem, utilizing an indicator function to determine whether any chosen sample point is inside the scatterer, thus identifying the scatterer. What distinguishes this method from existing sampling methods is that our indicator function no longer relies on solving integrals or integral equations. Instead, it is defined as a signed distance function approximated using standard deep neural networks.
Rather than developing sophisticated networks, we focus on the formulation of suitable inverse operators and data structures such that standard DNNs can be employed. This method maintains the simplicity and flexibility of using indicators in reconstructing scatterers. In comparison to the existing sampling methods, it leverages the advantages of deep learning methods by compensating for insufficient known information through empirical data, allowing it to work effectively even with limited far-field data. Numerical experiments demonstrate that the location and size of the obstacle are reconstructed effectively. Such a reconstruction can serve as a starting point for other methods.
Since the performance of a DNN is highly dependent on the training data, data tailored for single scatterers in the numerical implementation can only train a deep neural network suitable for the reconstruction of single scatterers. In the future, we plan to introduce training data for multiple scatterers and extend the idea of the DNN-oriented indicator method to other inverse problems involving partial data.