Next Article in Journal
k-Nearest Neighbour Estimation of the Conditional Set-Indexed Empirical Process for Functional Data: Asymptotic Properties
Previous Article in Journal
Application of the Averaging Method to the Optimal Control of Parabolic Differential Inclusions on the Semi-Axis
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Novel Neural Network-Based Approach Comparable to High-Precision Finite Difference Methods

1
School of Mathematical Statistics, Ningxia University, Yinchuan 750021, China
2
School of Science, Inner Mongolia University of Science and Technology, Baotou 014010, China
3
School of Science, Dalian Minzu University, Dalian 116600, China
*
Authors to whom correspondence should be addressed.
Axioms 2025, 14(1), 75; https://doi.org/10.3390/axioms14010075
Submission received: 12 December 2024 / Revised: 16 January 2025 / Accepted: 17 January 2025 / Published: 20 January 2025

Abstract

:
Deep learning methods using neural networks for solving partial differential equations (PDEs) have emerged as a new paradigm. However, many of these methods approximate solutions by optimizing loss functions, often encountering convergence issues and accuracy limitations. In this paper, we propose a novel deep learning approach that leverages the expressive power of neural networks to generate basis functions. These basis functions are then used to create trial solutions, which are optimized using the least-squares method to solve for coefficients in a system of linear equations. This method integrates the strengths of streaming PINNs and the traditional least-squares method, offering both flexibility and a high accuracy. We conducted numerical experiments to compare our method with the results of high-order finite difference schemes and several commonly used neural network methods (PINNs, lbPINNs, ELMs, and PIELMs). Thanks to the mesh-less feature of the neural network, it is particularly effective for complex geometries. The numerical results demonstrate that our method significantly enhances the accuracy of deep learning in solving PDEs, achieving error levels comparable to high-accuracy finite difference methods.

1. Introduction

Partial differential equations (PDEs) play a central role in the natural sciences, engineering, mathematics, and related fields, providing an effective tool for describing and analyzing complex phenomena and systems. These equations not only help us understand natural laws, but they also lay the foundation for scientific research and engineering practices [1,2]. Since many PDEs do not have analytical solutions, traditional numerical methods, such as the finite element method (FEM), finite difference method (FDM), finite volume method (FVM), etc., are widely used to compute approximate solutions, assisting researchers and engineers in obtaining meaningful results for various problems.
High-order compact FDMs are widely used for solving PDEs due to their high accuracy and effectiveness. These schemes typically achieve a fourth-order or higher spatial accuracy with smaller truncation errors compared to traditional methods. They require fewer mesh points to achieve the same accuracy, which is particularly beneficial for high-dimensional problems and helps reduce computational costs [3,4,5,6,7]. However, the construction and implementation of compact difference formats are complex, especially when dealing with intricate boundary conditions. They often assume that the solution is smooth, which may not be valid for problems with sharp gradients or discontinuities, potentially leading to unstable or inaccurate results. Additionally, many compact methods are implicit, requiring the solution of larger systems of linear equations to be solved, which increases the implementation complexity and computation time, particularly for large-scale problems.
In recent years, deep learning has gained traction in solving PDEs, leading to the development of various innovative methods, such as physically informed neural networks (PINNs) [8], the deep Ritz method (DRM) [9], deep Galerkin methods (DGMs) [10], and extreme learning machines (ELMs) [11], Self-adaptive loss balanced Physics-informed neural networks (lbPINN) [12]. These approaches demonstrate significant advantages in addressing complex PDE problems. PINNs integrate physical laws into the neural network’s loss function, enabling the model to learn from both data and the governing PDE during training. This method leverages the neural network’s strong fitting capabilities to represent the PDE solution, minimizing the error between the predicted and actual solutions. PINNs are particularly effective for problems with complex boundary conditions and multi-scale features, often yielding good results even with limited training data [13,14]. DGMs combine deep learning with Galerkin techniques to approximate PDE solutions through neural networks that satisfy the weak form of the equations. ELMs are single hidden-layer feedforward neural networks known for their simple and rapid training process, ideal for large-scale data. Unlike traditional neural networks, ELMs randomly select hidden layer weights and determine output weights using least-squares, significantly reducing the training time. They can quickly adapt to varying boundary and initial conditions when solving PDEs.
In addition, the extrapolation—driven network architecture—proposed by Wang, Y. [15] optimizes the physics—informed deep—learning model by combining extrapolation techniques, further enhancing the model’s accuracy and generalization ability. Zeng, B. [16] developed a method of integrating physical encoding into the message—the passing graph network (PhyMPGN), which is specifically designed to solve the spatiotemporal PDE system. These methods demonstrate the great potential of deep learning in the field of numerically solving PDEs, especially when traditional numerical methods encounter bottlenecks.
These deep learning methods offer fresh perspectives and tools for tackling PDEs, especially in scenarios where traditional numerical methods struggle. They effectively handle complex boundary conditions and nonlinear characteristics while minimizing mesh dependency. However, these approaches often require substantial data and computational resources, raising concerns about their feasibility and efficiency in practical applications [17,18,19,20]. Moreover, most deep learning-based methods necessitate optimizing the loss function to approximate the solutions accurately. The complexity of the networks and the number of hyperparameters can lead to optimization challenges, potentially resulting in local optima and an insufficient accuracy. To address these issues, researchers need to employ refined strategies in model design and training, such as advanced optimization algorithms, network structure adjustments, and regularization techniques, to enhance the generalization and accuracy [21,22,23,24].
Compared to the optimization loss function-based approach of PINNs, the ELM-based basis function method effectively avoids convergence issues in the optimization process and demonstrates a significant improvement in accuracy. This strategy reduces the reliance on traditional optimization algorithms by utilizing randomly generated weights and biases, resulting in faster computations and an enhanced model stability. Additionally, the ELM-based model exhibits a superior adaptability in managing complex boundary and initial conditions, making it more advantageous for addressing practical problems. Dwivedi [25] developed physics-informed extreme learning machines (PIELMs), which are fast versions of PINNs that can be applied to static and time-varying linear partial differential equations to meet or exceed the accuracy of PINNs on a range of problems. Quan [26] proposed a novel learning method based on the extreme learning machine algorithm, which utilizes the ELM algorithm to solve a system of linear equations, thereby determining the network parameters for solving PDEs. Wang et al. [27] presented an efficient extension of the extreme learning machine (ELM) method, moving from low dimensions to solving high-dimensional partial differential equations (PDEs). A neural network-based solution for linear and nonlinear partial differential equations has been proposed by integrating concepts from extreme learning machines (ELMs), domain decomposition, and local neural networks [28]. A high-precision subspace method based on neural networks has been proposed for solving partial differential equations. The core idea involves using neural network-based functions as the basis functions to span a subspace, allowing for the approximation of solutions within that subspace [29]. While neural networks hold significant promise for numerically solving PDEs, challenges like model generalization remain. Integrating neural networks with traditional numerical methods offers new avenues for achieving high-precision solutions, likely resulting in more efficient and accurate computations across various fields.
In light of the findings in the literature, we propose a novel deep learning strategy that integrates the ELM with the basis function space method. This approach leverages the advantages of the ELM to enhance the solution process and improve the accuracy of solving PDEs. Firstly, the ELM can be seen as selecting specific basis functions in the function space through randomly generated weights and biases. This not only enriches the expressiveness of the function space, but also significantly boosts the problem-solving efficiency, allowing for the better capture of the PDE solution characteristics and an improved model accuracy. Secondly, the ELM demonstrates an excellent adaptability, enabling flexible adjustments to various boundary and initial conditions. This adaptability is crucial in basis function space methods, as many real-world problems often involve complex and variable conditions. By incorporating the ELM, our model can maintain a strong performance across different physical scenarios, broadening its applicability. Finally, the ELM’s training time is substantially shorter than that of traditional neural networks, greatly enhancing the computational efficiency when addressing complex PDEs. This rapid training process reduces the computational costs and facilitates real-time or near real-time applications, particularly in engineering and scientific computing contexts that demand quick responses. In summary, our proposed deep learning strategy that combines an ELM with the basis function space method not only enhances the efficiency and accuracy of PDE solving, but also offers new insights and tools for tackling complex physical problems.
This paper is organized as follows: Section 2 introduces the model for partial differential equations and the fundamental principles of PINNs. Section 3 explores a high-precision basis function method based on neural networks. Section 4 evaluates the accuracy and effectiveness of the numerical validation method. The conclusion is presented in Section 5.

2. Mathematical Model and Physics-Informed Neural Network

2.1. Mathematical Model

Consider the following generalized form of PDE, which can be expressed as
f ( x ; u x 1 , u x d , 2 u x 1 x 1 , 2 u x 1 x d ) = 0 , x Ω
where x = x 1 , x 2 , x d are the independent variables, u ( x ) is the unknown partial function to be solved, and u x i is the partial derivative of u with respect to x i . PDEs can be classified into different types, such as elliptic, hyperbolic, parabolic, etc., depending on their nature. The exact form and type depend on the structure and boundary conditions of the equation.
To simplify the discussion, the PDEs and their boundary conditions can be transformed into the operator form.
L u ( x ) = f ( x ) , x Ω B u ( x ) = g ( x ) , x Ω
where x = ( x 1 , x 2 , , x d ) T , Ω is a bounded domain in d , Ω is the Ω boundary, L is a differential operator describing the physical process, and B is a boundary operator that specifies the conditions required at the boundary. f and g are the given functions.

2.2. Physics-Informed Neural Networks (PINNs)

PINNs integrate the laws of physics into neural network training, combining data-driven and physical models. They are particularly effective for optimizing the learning of complex systems through strong constraints and have recently emerged as a novel extensible approach to solving PDEs. The process of PINNs for solving PDEs involves three key elements, as shown in Figure 1: (1) the neural network construction trial function; (2) the design of the loss function, which includes PDE residuals, boundary conditions, and labeled data; and (3) the optimization of the training process. These elements interact to create a comprehensive framework for solving PDEs using PINNs.
First, a feedforward neural network (FNN) is defined with N hidden layers to construct the trial function. The trial function u ( x , Θ ) is used as a hypothetical solution to approximate the solution of a partial differential equation. The accuracy of the solution is optimized by tuning the network parameters Θ (i.e., all weights and biases) through training.
u ( x , Θ ) : = W N σ ( W N 1 σ ( W 2 σ ( W 1 x + b 1 ) + b 2 ) + b N 1 ) + b N
where x is the input data, usually containing independent variables (e.g., location and time); W i and b i are the weight matrix and bias vector of i-th layer, with i = 1 , 2 , , N respectively; Θ = { k = 1 N ( W k , b k ) , k = 1 , 2 , , N } is the set of tunable parameters of the neural network; and σ is the activation function.
This trial function (2) constructed from neural networks was used in the PINN framework to solve the corresponding partial differential equations, combining the loss function and the training optimization process to achieve a sufficient accuracy. To constrain the optimization direction of the trial function, PINNs define an objective loss function composed of four components: (1) PDE residual loss, which ensures that the trial function u ( x , Θ ) minimizes the error in satisfying the given PDE. (2) Boundary condition loss, which ensures that the trial function meets practical requirements under the boundary conditions, such as Dirichlet or Neumann conditions. (3) Initial condition loss (if applicable), which ensures that the trial function meets the specified initial conditions. (4) Data loss term, which accounts for errors relative to the provided data, if labeled data are available. Together, these components guide the optimization process effectively.
L ( x , u , Θ , ω ) = ω f L P D E ( x , u , Θ ) + ω i L I C ( x , u , Θ ) + ω b L B C ( x , u , Θ ) + ω d L D a t a ( x , u , Θ )
where L P D E , L B C , L I C , and L d a t a are defined as follows:
L P D E ( x , u , Θ ) = 1 M Ω i = 1 M Ω L u ( x i , Θ ) f ( x i ) 2 , x i Ω ^ L B C ( x , u , Θ ) = 1 M Ω i = 1 M Ω B u ( x i , Θ ) g ( x i ) 2 , x i Ω ^ L I C ( x , u , Θ ) = 1 M I i = 1 M I I u ( x i , Θ ) u 0 ( x i ) 2 , x i Ω ^ L D a t a ( x , u , Θ ) = 1 M d a t a i = 1 M d a t a u ( x i , Θ ) u d ( x i ) 2 , x i D a t a
and ω = { ω f , ω i , ω b , ω d } are weighting factors to balance the contribution of each component to the total loss. M Ω ,   M Ω ,   M I , and M d a t a represent the number of collocations in the interior of the domain, on the boundary, at the initial condition, and the number of labeled data points, respectively.
Finally, the network weight and bias parameters are updated using an optimization algorithm (e.g., Adam, L-BFGS, a combination of Adam and L—BFGS, or stochastic gradient descent) to minimize the overall loss function (3). The training process concludes when the loss reaches a specified threshold, it shows minimal decrease, or the maximum number of iterations is reached.
Θ * = arg min Θ L ( x , u , Θ , ω ) ,   u ( x ) u ( x , Θ * )
Although PINNs offer a versatile framework for solving partial differential equations (PDEs), they face significant challenges in optimizing the loss function to accurately approximate the PDE solutions. This optimization process often requires substantial CPU time and can lead to issues such as convergence to local optima due to the non-convex nature of the loss function. These challenges not only hinder the convergence of the training process, but also compromise the accuracy of the approximations produced by the network.

3. The Neural Network-Based Basis Function Space Method

The universal approximation theorem demonstrates the remarkable expressive power of neural networks, indicating that even a simple two-layer neural network can approximate complex functions to an arbitrary accuracy. Building on this foundation, we utilized the concept of extreme learning machines (ELMs) to generate basis functions through a neural network framework. These basis functions served as the building blocks for constructing our trial solution. Once the trial solution was formulated from the basis functions, we enhanced its accuracy by incorporating traditional numerical methods, such as least-squares optimization, as shown in Figure 2. This hybrid approach allows us to leverage the strengths of both machine learning and numerical analyses, resulting in a more robust and accurate solution to the problem at hand. By effectively combining neural network-based basis functions with established numerical techniques, we can achieve significant improvements in accuracy while maintaining computational efficiency. This methodology not only streamlines the process of finding solutions to complex equations, but also expands the potential applications of neural networks in solving various mathematical and engineering problems. Through this integration, we aimed to harness the full potential of neural networks and traditional numerical methods, paving the way for the more effective and accurate approximation of complex functions.
A feedforward neural network is defined with N hidden layers, where the j -th hidden layer contains N j neurons, the mathematics of which can be expressed by the following equation:
h 0 = x , h k = σ k ( W k h k 1 + b k ) , k = 1 , 2 , N 1 , Φ = h N 1 , u ( x , Θ ) = Φ W N ,
Given an input x d , the output of the network u ( x , Θ ) is computed through a series of weights and activation functions. W k and b k represent the weight and bias, respectively. σ ( ) is the activation function, θ = W 1 , W 2 , , W N , b 1 , b 2 , , b N 1 is a set of parameters in the neural network, and u ( x , θ ) is the output with parameters θ . The initial set of parameters for a neural network are usually randomly generated.
From Equation (6), it is evident that the neural network’s output can be viewed as a combination function of the last layer’s weight coefficients and the output of the penultimate hidden layer’s neurons. This combination allows the network to capture complex input–output relationships and perform nonlinear mapping. Consequently, this multi-layer structure facilitates layer-wise feature extraction, enhancing the model’s expressive power and learning efficiency.
Assuming that the number of neurons in the penultimate hidden layer is M, then the output of M neurons can be viewed as a set of basis functions and denoted as Φ
Φ = { φ 1 ( x , Θ ) , φ 2 ( x , Θ ) , , φ M ( x , Θ ) }
and the function space V can be written as
V = s p a n { φ 1 ( x , Θ ) , φ 2 ( x , Θ ) , , φ M ( x , Θ ) }
The weights of the last hidden layer can be expressed as
ω = { ω 1 , ω 2 , , ω M }
Thus, the output of the neural network can be understood as a function formed by a linear combination of the basis functions, represented as
u ( x , Θ , ω ) = ω Φ = k = 1 M ω k φ ( x , Θ )
Given the diverse types of PDEs and their corresponding solutions, it is essential to train the basis functions appropriately to enable them to express the solutions of PDEs more effectively. This training process focuses on optimizing the parameters of the basis functions, allowing them to more accurately capture the unique characteristics of the solutions for specific PDEs. We used the PINN approach to construct the loss function for pre-training the neural network. Due to the complexity of the PINN optimization process, we established a stopping condition: training was halted when the loss function decreased by two orders of magnitude. This strategy ensured that the network achieved a higher accuracy while minimizing the risk of overfitting, thereby enhancing the model’s generalization ability. Generally, the loss function is as follows:
L D = 1 N Ω j = 1 N Ω ( L u ( x j , Θ ) f ( x j ) ) 2 + 1 N Ω j = 1 N Ω ( B u ( x j , Θ ) g ( x j ) ) 2
Sufficient training was conducted on the loss function (11) until the maximum number of training steps was reached or the following condition was met:
L D L D 0 ε .
where L D 0 is the initial loss error. We adopted ε = 10 3 and the maximum number of training steps was N max = 1000 .
Θ * = arg min Θ L D
After completing the pre-training of the network (12), we obtained an initial set of optimal weights that satisfactorily met the requirements of the partial differential control equations and their boundary conditions (13). Using these weights, we constructed customized basis functions tailored for specific boundary value problems involving partial differential equations.
Φ * = { φ 1 ( x , Θ * ) , φ 2 ( x , Θ * ) , , φ M ( x , Θ * ) }
These basis functions are crucial for their adaptability, as they are specifically designed to effectively address particular problems and accurately capture the solution characteristics. Accordingly, the trial solution output by the neural network can be expressed as
u ( x , ω ) = ω Φ * = k = 1 M ω k φ ( x , Θ * )
A set of neural network-based basis functions was obtained by optimizing the loss function, as shown in Equation (14), and then the approximation solution could be expressed as a linear combination of this set of basis functions, as shown in Equation (15). Next, the combination coefficients were optimized to give a higher approximation accuracy.
Substituting the trial solution in Equation (15) into the governing equation, Equation (1), and making the control equation satisfied at each collocation point yielded the following system of equations:
j = 1 M ω j A φ j ( x i ) = f ( x i ) , x i Ω ^ , i = 1 , 2 , , N Ω , j = 1 M ω j B φ j ( x i ) = g ( x i ) , x i Ω ^ , i = 1 , 2 , , N Ω .
This is a system of linear equations for the combined weight coefficients ω = [ ω 1 , ω 2 , , ω M ] T , which can be written in matrix form as
A φ 1 ( x 1 ) A φ 2 ( x 1 ) A φ 3 ( x 1 ) A φ M ( x 1 ) A φ 1 ( x 2 ) A φ 2 ( x 2 ) A φ 3 ( x 2 ) A φ M ( x 2 ) A φ 1 ( x N Ω ) A φ 2 ( x N Ω ) A φ 3 ( x N Ω ) A φ M ( x N Ω ) B φ 1 ( x 1 ) B φ 2 ( x 1 ) B φ 3 ( x 1 ) B φ M ( x 1 ) B φ 1 ( x 2 ) B φ 2 ( x 2 ) B φ 3 ( x 3 ) B φ M ( x 2 ) B φ 1 ( x N Ω ) B φ 2 ( x N Ω ) B φ 3 ( x N Ω ) B φ M ( x N Ω ) ω 1 ω 2 ω 3 ω 4 ω 5 ω 6 ω M = f ( x 1 ) f ( x 2 ) f ( x N Ω ) g ( x 1 ) g ( x 2 ) g ( x N Ω )
It can be further abbreviated as
H ω = R
where H ( N Ω + N Ω ) × M , R ( N Ω + N Ω ) × 1 , and ω M × 1 .
This is a system of hyper-deterministic linear equations, and we can solve this system using the least-squares method to obtain the optimal combination of coefficients.
ω * = ( H T H ) 1 H T R
The optimized trial solution will approximate the analytical solution of the PDEs with a high accuracy.
u ( x , ω * ) u * ( x ) < ε

4. Numerical Examples

In this section, we demonstrate the high-precision characteristics of the present method through a series of numerical experiments.
The definition of error was as follows:
e L 2 = i = 1 n ( u θ ( x i ) u ^ θ ( x i ) ) 2 i = 1 n u i θ ( x i ) ,   e L = max 0 i n u θ ( x i ) u ^ θ ( x i )
The activation function was set to the sin function in the numerical examples. The parameter updates were based on the Adams method, with a standard learning rate of 0.001. This network model contained 100 neurons in each hidden layer, and the number of subspace layer basis functions was M. The stopping condition for the training basis function was 1 × 10 3 , and training points were sampled uniformly for a fixed random seed at 1.

4.1. Reaction–Diffusion Equation

Consider the reaction–diffusion equation in the spatio-temporal domain Ω = [ a , b ] × ( 0 , T ] ,
u t = u x x + f ( x , t ) , u ( a , t ) = g 1 ( t ) , u ( b , t ) = g 2 ( t ) , u ( x , 0 ) = h ( x ) .
where g 1 ( t ) ,   g 2 ( t ) is the boundary condition and h ( x ) is the initial condition. When a = 0 , b = 1 , T = 1 , the exact solution is u ( x ) = 1 1 + e θ ( x θ t ) , θ = 1 2 . .
In Table 1, the depth of the four hidden layers remains fixed. By varying the number of training points and the quantity M of basis functions within the subspace layer, the optimal error value is achieved when M = 600 , with the error reaching as low as 10 14 . Table 2 presents the error under different conditions of hidden layer depths and training points. It can be observed that the error of the present method ranged from 10 12 to 10 13 , and when there were four hidden layers and 600 basis functions, the error reached as low as 1.75 × 10 14 . On the whole, regardless of changing the number of basis functions or the depth of the hidden layer, the error for the present method remained stable. Therefore, the optimal neural network structure obtained by this method in this example was 2 , 100 , 100 , 100 , 100 , M = 600 , 1 .
Table 3 compares the performance of different activation functions. It can be seen that the present method was able to stably achieve a high-precision solution under different activation functions. In the following tests, we used sin as the activation function.
Table 4 shows that different initial weights, biases, and so on are obtained by changing the random seeds. It was found that the present method is not sensitive to these changes and the error remains stable.
Table 5 compares the errors of PINNs, lbPINNs, ELMs, PIELMs, the traditional FDM, and the present method. The FDM had errors at τ = 1 / 128 and h = 1 / 128 , the errors of the other five methods with the same internal training points of 20 × 20 .
Among them, PINNs, lbPINNs, and the existing method are under a network structure with four hidden layers and a subspace layer M = 600 , while ELMs and PIELMs have only one hidden layer with a value of 600. It can be seen that this method achieves results comparable to the traditional FDM method and other neural network methods with fewer training points and iteration times, further demonstrating the advantages of this method in solving general problems.
Figure 3 shows the loss function evolution curves of different methods. It can be seen that PINNs and lbPINNs basically reached convergence and stability at 6000 steps. In contrast, the present method only needed to drop the loss function by three orders of magnitude, so iterations of tens of steps can achieve the purpose, avoiding the inefficiency problem caused by long iterations. The method in this paper spent far fewer iteration steps and much less CPU time than the PINN and lbPINN methods, but it achieved a much higher accuracy than them.
Figure 4 presents the exact solutions, predicted solutions, and error diagrams of the present method under different values of θ . It can be observed that, when θ increases, the solutions will become steeper, which makes the function u ( x ) change more drastically in the region close to x = θ t . Meanwhile, the difficulty of solving will also increase. Nevertheless, our method can still maintain a good accuracy during the solving process.

4.2. Damped Wave Equation

This example aimed to demonstrate the effect of damping on the propagation of waves.
2 u t 2 + α u t = 2 u x 2 + ( 1 + α + k 2 π 2 ) e t sin ( k π x ) , x , t 0 , 1 × 0 , T
The exact solution to this problem is u x , t = e t sin ( k π x ) .
Table 6 shows the error values under different numbers of internal training points and M , from which it can be seen that the minimum error corresponds to M = 600 , consistent with the optimal network structure obtained in Example 1. Therefore, in Table 5, Table 6, Table 7 and Table 8, only the optimal structure conditions are changed with different internal training points to obtain the optimal number of training points and the corresponding losses and errors. Table 6 and Table 7 show the e L errors when k = 1 , T = 1 , α takes different values. It can be seen that the error at k = 1 , α = 0.2 , T = 1 remains at 10 6 10 8 , Figure 5 shows the exact solution, the predicted solution, and the error plot with 50 × 50 internal training points. And the e L , e L 2 error at internal training points 40 × 40 remains at 10 7 10 9 , when k = 1   , α = 0.4 ,   T = 1 .
Table 8 shows that the error is optimal at the internal training points of 60 × 60 , and is 1.89 × 10 7 when k = 1 , α = 0.2 , T = 5 , Figure 6 shows the exact solution, predicted solution, and error plot under this parameter. Table 9 shows that the error is optimal at the internal training points of 100 × 100 , and is 8.32 × 10 7 when k = 1 , α = 0.4 , T = 5 . By changing the number of internal training points, the e L error is very stable at 10 6 10 7 .
Table 10 depicts the error at different numbers of internal points when k = 1 , T = 5 , α = 30, 40, 50. We can see that the optimal internal points are 100 × 100 , 100 × 100 , and 150 × 150 when α = 30, 40, 50. The number of required training points increases as the value of α increases.
Table 11 compares the error values of six methods when α , T are taken with different values. The FDM method takes the spatiotemporal grid as h = 1 / 50 , τ = h when α = 0.2 , T = 1 . When α = 0.2 , T = 5 , the grid is h = 1 / 60 , τ = 3.2 h 2 . When α = 0.4 and T = 1 , the spatiotemporal grid is h = 1 / 40 , τ = h ; when T = 5 , the grid is h = 1 / 100 , τ = 3.2 h 2 . For PINNs, lbPINNs, and the present method, when the parameters are different, the internal training points are 50 × 50 ,   60 × 60 ,   40 × 40 ,   100 × 100 and the network structure is [ 2 ,   100 ,   100 ,   100 ,   100 ,   M = 600 ,   1 ] ; for the ELM and PIELM network structure of [ 2 , M = 600 , 1 ] . It can be seen that, under the same parameter conditions, the method proposed in this paper has a better error performance compared with those of other methods, and its precision is also very stable, truly reaching a level comparable to that of traditional methods.
Figure 7 shows the exact solution for different k values, the numerical solution of the method in this paper, and the corresponding error. It can be seen that the frequency of vibration of the function values on the x-axis accelerated as k increased and the solution became more complex. However, the present method was still able to solve the case of complex solutions well and with a relatively high accuracy.

4.3. High-Wave-Number Helmholtz Equation

u x x + k 2 u = f ( x ) , x 0 , 1
Exact   solution : u ( x ) = sin ( m π x )
Table 12 compares the errors under four hidden layers with 100 neurons per hidden layer at different training points and M. From the results, it seems that the present method was able to achieve a stable and high accuracy under different numbers of training points and different numbers of subspace neurons.
Table 13 shows the e L error for different values of k in the optimal network structure ( 1 , 100 , 100 , 100 , 100 , M = 600 , 1 ) when m = 2 . It can be seen from Table 13 that, under the conditions of different wave numbers, when the number of sampling points is changed, the performance of the method proposed in this study was quite stable and had a very high precision, with an e L error within 10 13 .
Table 14, Table 15 and Table 16 show the error values for different values of m . As m increased, the function graph was compressed in the x -axis direction and became steeper, and the difficulty of solving also increased. It can be seen that the method in this paper still maintained a high level of accuracy as m increased.
Table 17 compares the errors of PINNs, lbPINNs, ELM, PIELM, FDM, and the present method under the same parameters at different values of m . The PINNs, lbPINNs, and present method were under the optimal network structure of 1 , 100 , 100 , 100 , 100 , M = 600 , 1 ; the ELM and PIELM contained only one hidden layer and took 600. By presenting the error values under different wave numbers, it can be seen that, regardless of how m changed, the present method performed well, with a high precision. The e L 2 error values of the FDM at different m values and the spatial grids h = 1 / 30 , 1 / 100 , 1 / 500 , 1 / 2000 show that, compared with the present method, our method had a significantly better accuracy when the m value was small. As the m value increased, the accuracy of the two methods was basically the same. This also indicates that the present method performed better than traditional methods and other neural network methods.
Figure 8 shows the prediction results of the exact solution sums with different parameters and the corresponding error plots. The results show that the solution exhibited high-frequency characteristics when the parameter m increased; when m = 50, a significant jump occurred. The reason for this is that, due to complex geometric and physical phenomena, the neural network may fail to capture all the key features of the problem during the training process (underfitting). In this case, it can be eliminated simply by increasing the dataset. And the present method was able to perfectly match and capture the characteristics of the solution in the case of different forms of solutions and exhibited a superb accuracy.

4.4. The Poisson Equation Within the Complex Region

This example used the pigeon-shape interface, and had a very complex parametric interface. The interface equation is
x ( θ ) = r cos ( θ )   y ( θ ) = r sin ( θ )
where r   =   0.6   +   0.216 sin ( 3 θ )   +   0.096 cos ( 2 θ )   +   0.24 cos ( 5 θ ) ,   θ 0 ,   2 π .
The Poisson equation for this example is provided below:
Δ u = f ( x , y ) , ( x , y ) Ω , u ( x , y ) = g ( x , y ) , ( x , y ) Ω .
And f ( x , y ) , g ( x , y ) satisfies certain conditions within the above parameterized interface, whose exact solution is
u ( x , y ) = sin π x sin π y .
Figure 9 illustrates the non-regular computation region and the distribution of the training points used. There were 3000 sample points on the boundary and 30,000 sample points inside the region.
Table 18 shows the error and CPU time for different numbers of training points. The results under different training points achieved a very high accuracy, and the CPU time did not increase significantly with an increase in the number of training points, which shows a very high accuracy and efficiency.
Figure 10 illustrates the exact solution, the numerical solution, and the corresponding error for a complex regional scenario. It can be seen that the method in this paper demonstrated an excellent capability on the basis of complex regions and achieved an impressive accuracy. This synergy between efficiency and accuracy gives the method great potential for tackling complex regional challenges, paving the way for future widespread applications and heralding a new era of problem solving in related fields.

5. Discussion

The neural network-based basis function method proposed in this paper employs a two-step approach to achieve a high-accuracy approximation of PDE solutions. In the first step, the loss function is constructed based on the PDE control equations. The neural network is then trained, and the basis functions are generated through a small number of iterative steps. By excluding additional solution information, such as boundary conditions, the complexity and difficulty of the optimization problem are reduced, improving the efficiency of the basis function training process. In the second step, the base functions are used to construct the trial solution, which satisfies the control equations and solution conditions at the training points. This yields a linear system of equations for the combined coefficients of the trial solution. Highly accurate approximate solutions can then be obtained by optimizing the combination coefficients using the least-squares method. The proposed method combines the neural network approach and the traditional least-squares method, leveraging the advantages of both. It does not require mesh generation or extensive training, and it solves a system of algebraic equations to obtain the continuous function solution, rather than just discrete points. Compared to traditional methods, this approach achieves a high accuracy and efficiency.

6. Conclusions

Neural networks have achieved significant breakthroughs in various fields, particularly in solving complex PDEs. However, deep learning methods that optimize loss functions often face challenges like long training times, local optima, and a limited accuracy. This paper introduces a novel strategy inspired by extreme learning machines that uses neural networks to generate basis functions and optimize trial solutions. Our method demonstrates substantial improvements in both efficiency and accuracy compared to conventional approaches like finite difference methods and other neural network techniques such as PINNs, lbPINNs, ELM, and PIELM. The results indicate that our approach matches the accuracy of traditional methods while significantly reducing the number of iterative steps required. This advancement not only enhances the efficiency of solving PDEs, including complex domain problems, but it also challenges preconceived notions about the accuracy of neural networks. Overall, this breakthrough opens new avenues for research and applications in using neural networks for PDE solutions.
Though the proposed method has demonstrated advantages in terms of accuracy and efficiency in solving smooth PDE problems with single equations, further exploration is needed to address the challenges posed by a broader range of PDE problems with discontinuous solutions, as well as those exhibiting strongly nonlinear characteristics.

Author Contributions

Conceptualization, F.P., F.C. and Y.G.; methodology, F.P., F.C. and Y.G.; software, F.P. and F.C.; validation, F.P., F.C. and Y.G.; formal analysis, F.P. and F.C.; writing—original draft preparation, F.P. and F.C.; writing—review and editing, F.C. and Y.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The authors declare the availability of the data.

Acknowledgments

This work was partially supported by the National Natural Science Foundation of China (12161067), the National Natural Science Foundation of China Project (12261067), and the National Natural Science Foundation of Ningxia (2022AAC02023, 2023AAC03002).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Ames, W.F. Numerical Methods for Partial Differential Equations; Academic Press: Cambridge, MA, USA, 2014. [Google Scholar]
  2. Morton, K.W.; Mayers, D.F. Numerical Solution of Partial Differential Equations; Cambridge University Press: Cambridge, UK, 1995. [Google Scholar]
  3. Spotz, W.F.; Carey, G.F. High-order compact finite difference methods. In Proceedings of the Preliminary Proceedings International Conference on Spectral and High Order Methods, Houston, TX, USA, 5–9 June 1995; pp. 397–408. [Google Scholar]
  4. Liu, S.E.; Ge, Y.; Wang, S. Conservative higher-order finite difference scheme for the coupled nonlinear Schrödinger equations. Commun. Nonlinear Sci. Numer. Simul. 2024, 131, 107797. [Google Scholar] [CrossRef]
  5. Wang, S.; Ge, Y.; Liu, S.E. Numerical solutions of the nonlinear wave equations with energy-preserving sixth-order finite difference schemes. Comput. Math. Appl. 2024, 168, 100–119. [Google Scholar] [CrossRef]
  6. Liu, S.E.; Ge, Y.; Qi, Y. A sixth-order finite difference method for the two-dimensional nonlinear advection diffusion reaction equation. J. Comput. Phys. 2024, 517, 113361. [Google Scholar] [CrossRef]
  7. Tian, Z.F.; Dai, S.Q. High-order compact exponential finite difference methods for convection–diffusion type problems. J. Comput. Phys. 2007, 220, 952–974. [Google Scholar] [CrossRef]
  8. Raissi, M.; Perdikaris, P.; Karniadakis, G.E. Physics-informed neural networks: A deep learning framework for solving forward and inverse problems involving nonlinear partial differential equations. J. Comput. Phys. 2019, 378, 686–707. [Google Scholar] [CrossRef]
  9. Yu, B. The deep Ritz method: A deep learning-based numerical algorithm for solving variational problems. Commun. Math. Stat. 2018, 6, 1–12. [Google Scholar]
  10. Sirignano, J.; Spiliopoulos, K. DGM: A deep learning algorithm for solving partial differential equations. J. Comput. Phys. 2018, 375, 1339–1364. [Google Scholar] [CrossRef]
  11. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: A new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; Volume 2, pp. 985–990. [Google Scholar]
  12. Xiang, Z.; Peng, W.; Liu, X. Self-adaptive loss balanced Physics-informed neural networks. Neurocomputing 2022, 496, 11–34. [Google Scholar] [CrossRef]
  13. Cuomo, S.; Di Cola, V.S.; Giampaolo, F.; Rozza, G.; Raissi, M.; Piccialli, F. Scientific machine learning through physics–informed neural networks: Where we are and what’s next. J. Sci. Comput. 2022, 92, 88. [Google Scholar] [CrossRef]
  14. Markidis, S. The old and the new: Can physics-informed deep-learning replace traditional linear solvers? Front. Big Data 2021, 4, 669097. [Google Scholar] [CrossRef]
  15. Wang, Y.; Yao, Y.; Gao, Z. An extrapolation-driven network architecture for physics-informed deep learning. Neural Netw. 2025, 183, 106998. [Google Scholar] [CrossRef] [PubMed]
  16. Zeng, B.; Wang, Q.; Yan, M.; Liu, Y.; Chengze, R.; Zhang, Y.; Liu, H.; Wang, Z.; Sun, H. PhyMPGN: Physics-encoded Message Passing Graph Network for spatiotemporal PDE systems. arXiv 2024, arXiv:2410.01337. [Google Scholar]
  17. Wang, S.; Yu, X.; Perdikaris, P. When and why PINNs fail to train: A neural tangent kernel perspective. J. Comput. Phys. 2022, 449, 110768. [Google Scholar] [CrossRef]
  18. Urbán, J.F.; Stefanou, P.; Pons, J.A. Unveiling the optimization process of Physics Informed Neural Networks: How accurate and competitive can PINNs be? arXiv 2024, arXiv:2405.04230. [Google Scholar] [CrossRef]
  19. Rathore, P.; Lei, W.; Frangella, Z.; Lu, L.; Udell, M. Challenges in training PINNs: A loss landscape perspective. arXiv 2024, arXiv:2402.01868. [Google Scholar]
  20. Yan, L.; Zhou, Y.; Liu, H.; Liu, L. An improved method for Physics-Informed Neural Networks that accelerates convergence. IEEE Access 2024, 12, 23943–23953. [Google Scholar] [CrossRef]
  21. Zou, Z.; Meng, X.; Karniadakis, G.E. Correcting model misspecification in physics-informed neural networks (PINNs). J. Comput. Phys. 2024, 505, 112918. [Google Scholar] [CrossRef]
  22. Müller, J.; Zeinhofer, M. Achieving high accuracy with PINNs via energy natural gradient descent. In Proceedings of the International Conference on Machine Learning, Budapest, Hungary, 25–29 July 2004; pp. 25471–25485. [Google Scholar]
  23. Hua, J.; Li, Y.; Liu, C.; Wan, P.; Liu, X. Physics-informed neural networks with weighted losses by uncertainty evaluation for accurate and stable prediction of manufacturing systems. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 11064–11076. [Google Scholar] [CrossRef]
  24. Mishra, S.; Molinaro, R. Estimates on the generalization error of physics-informed neural networks for approximating PDEs. IMA J. Numer. Anal. 2023, 43, 1–43. [Google Scholar] [CrossRef]
  25. Dwivedi, V.; Srinivasan, B. Physics informed extreme learning machine (pielm)–a rapid method for the numerical solution of partial differential equations. Neurocomputing 2020, 391, 96–118. [Google Scholar] [CrossRef]
  26. Quan, H.D.; Huynh, H.T. Solving partial differential equation based on extreme learning machine. Math. Comput. Simul. 2023, 205, 697–708. [Google Scholar] [CrossRef]
  27. Wang, Y.; Dong, S. An extreme learning machine-based method for computational PDEs in higher dimensions. Comput. Methods Appl. Mech. Eng. 2024, 418, 116578. [Google Scholar] [CrossRef]
  28. Dong, S.; Li, Z. Local extreme learning machines and domain decomposition for solving linear and nonlinear partial differential equations. Comput. Methods Appl. Mech. Eng. 2021, 387, 114129. [Google Scholar] [CrossRef]
  29. Xu, Z.; Sheng, Z. Subspace method based on neural networks for solving the partial differential equation. arXiv 2024, arXiv:2404.08223. [Google Scholar]
  30. Sheng, L.; Yong, G.; Ying, Q. High-precision finite-difference method for the nonlinear reaction-diffusion equation. Appl. Math. 2023, 36, 747–755. [Google Scholar]
  31. Meng, W.; Yong, G.; Zhi, W. Explicit high-order compact difference method for solving nonlinear hyperbolic equations with three types of boundary conditions. Wave Motion 2023, 118, 103120. [Google Scholar]
  32. Zhi, W.; Yong, G. A high precision compact difference method for solving the Helmholtz equation of variable coefficient. J. Chongqing Norm. Univ. 2019, 36, 92–97. [Google Scholar]
Figure 1. The process framework for PINNs solving PDEs.
Figure 1. The process framework for PINNs solving PDEs.
Axioms 14 00075 g001
Figure 2. Neural network-based basis function construction process.
Figure 2. Neural network-based basis function construction process.
Axioms 14 00075 g002
Figure 3. The loss decrease curves of PINNs, lbPINNs, and the present method.
Figure 3. The loss decrease curves of PINNs, lbPINNs, and the present method.
Axioms 14 00075 g003
Figure 4. Exact solution, predicted solution, and error plot under different θ values.
Figure 4. Exact solution, predicted solution, and error plot under different θ values.
Axioms 14 00075 g004
Figure 5. The (left) predicted solution, (middle) exact solution, and (right) error at k = 1 ,   α = 0.2 ,   T = 1 .
Figure 5. The (left) predicted solution, (middle) exact solution, and (right) error at k = 1 ,   α = 0.2 ,   T = 1 .
Axioms 14 00075 g005
Figure 6. The predicted solution (a), exact solution (b), and error (c) at k = 1 , α = 0.2 , T = 5 .
Figure 6. The predicted solution (a), exact solution (b), and error (c) at k = 1 , α = 0.2 , T = 5 .
Axioms 14 00075 g006
Figure 7. The exact solutions, predicted solutions, and error graphs corresponding to different values of k when α = 0.2 and T = 1 .
Figure 7. The exact solutions, predicted solutions, and error graphs corresponding to different values of k when α = 0.2 and T = 1 .
Axioms 14 00075 g007
Figure 8. The exact solution, predicted solution, and error plot for different values of m at k = 1 , α = 0.2 .
Figure 8. The exact solution, predicted solution, and error plot for different values of m at k = 1 , α = 0.2 .
Axioms 14 00075 g008aAxioms 14 00075 g008b
Figure 9. Computational domain and scatter plot.
Figure 9. Computational domain and scatter plot.
Axioms 14 00075 g009
Figure 10. Predicted solution (left), exact solution (middle), and error (right) for N i n = N b c = 500 .
Figure 10. Predicted solution (left), exact solution (middle), and error (right) for N i n = N b c = 500 .
Axioms 14 00075 g010
Table 1. Error with fixed hidden layer depth.
Table 1. Error with fixed hidden layer depth.
Points
M
400500600700800
20 × 20 5.94 × 10 14 3.70 × 10 14 1.75 × 10 14 5.84 × 10 14 3.13 × 10 14
25 × 25 9.03 × 10 14 4.47 × 10 14 7.76 × 10 14 4.71 × 10 14 3.60 × 10 14
30 × 30 1.77 × 10 13 9.80 × 10 14 7.00 × 10 14 3.77 × 10 14 3.57 × 10 14
35 × 35 2.57 × 10 13 8.69 × 10 14 7.11 × 10 14 6.11 × 10 14 5.16 × 10 14
40 × 40 1.35 × 10 13 6.30 × 10 14 9.79 × 10 14 7.73 × 10 14 6.94 × 10 14
60 × 60 1.28 × 10 13 8.09 × 10 14 1.81 × 10 13 1.05 × 10 13 7.94 × 10 14
Table 2. Error for different hidden layer depths.
Table 2. Error for different hidden layer depths.
Hidden Layer
M
400500600800
3 4.10 × 10 14 1.83 × 10 14 1.80 × 10 14 1.77 × 10 14
4 5.94 × 10 14 3.70 × 10 14 1.75 × 10 14 3.13 × 10 14
5 4.43 × 10 13 4.91 × 10 13 4.39 × 10 13 2.47 × 10 13
6 2.20 × 10 12 6.93 × 10 13 3.89 × 10 12 6.91 × 10 13
Table 3. Errors corresponding to different activation functions.
Table 3. Errors corresponding to different activation functions.
Activation FunctionSinCosTanhSigmoidSwish
e L 2 7.69 × 10 15 1.63 × 10 11 3.86 × 10 13 2.70 × 10 10 1.86 × 10 13
e L 1.75 × 10 14 5.14 × 10 11 7.56 × 10 13 4.35 × 10 10 2.43 × 10 13
Table 4. Errors corresponding to different random seeds.
Table 4. Errors corresponding to different random seeds.
Random Seed12571020
e L 2 7.69 × 10 15 1.29 × 10 14 2.25 × 10 14 2.12 × 10 14 4.10 × 10 14 2.31 × 10 14
e L 1.75 × 10 14 2.35 × 10 14 3.47 × 10 14 3.42 × 10 14 7.20 × 10 14 3.84 × 10 14
Table 5. Comparison of the errors and CPU time of different methods.
Table 5. Comparison of the errors and CPU time of different methods.
MethodNetwork Structure e L e L 2 EpocshCPU
PINNs 2 , 100 , 100 , 100 , 100 , M = 600 , 1 3.04 × 10 4 2.44 × 10 4 6000476.84s
lbPINNs 2 , 100 , 100 , 100 , 100 , M = 600 , 1 2.09 × 10 4 1.98 × 10 4 6000496.44s
ELMs 2 , M = 600 , 1 7.18 × 10 13 4.54 × 10 13 09.34s
PIELMs 2 , M = 600 , 1 1.47 × 10 13 2.88 × 10 13 32716.53s
FDM [30]-6.70 × 10 14 4.10 × 10 14 32030.16s
Present 2 , 100 , 100 , 100 , 100 , M = 600 , 1 1.75 × 10 14 7.69 × 10 15 8116.36s
Table 6. Error at k = 1 ,   α = 0.2 ,   T = 1 with different M and training points.
Table 6. Error at k = 1 ,   α = 0.2 ,   T = 1 with different M and training points.
Points
M
2004006008001000
20 × 20 9.15 × 10 7 2.39 × 10 7 7.31 × 10 8 1.84 × 10 7 1.38 × 10 7
30 × 30 1.31 × 10 6 1.58 × 10 7 4.77 × 10 8 1.88 × 10 7 2.11 × 10 7
40 × 40 1.19 × 10 6 1.21 × 10 7 1.05 × 10 7 2.08 × 10 7 1.90 × 10 7
50 × 50 7.15 × 10 7 4.73 × 10 8 1.05 × 10 8 1.81 × 10 7 1.08 × 10 7
Table 7. Error at k = 1 ,   α = 0.4 ,   T = 1 under different training points.
Table 7. Error at k = 1 ,   α = 0.4 ,   T = 1 under different training points.
Points 20 × 20 30 × 30 40 × 40 50 × 50 60 × 60
e L 2 5.47 × 10 9 9.56 × 10 9 2.09  × 10 9 1.29 × 10 8 5.38 × 10 8
e L 1.78 × 10 8 2.71 × 10 8 6.72  × 10 9 3.87 × 10 8 1.61 × 10 7
Table 8. Error at different numbers of training points and subspace dimensions when k = 1 ,   α = 0.2 ,   T = 5 .
Table 8. Error at different numbers of training points and subspace dimensions when k = 1 ,   α = 0.2 ,   T = 5 .
Points 50 × 50 60 × 60 80 × 80 100 × 100 120 × 120
e L 1.33 × 10 5 1.43  × 10 6 6.75 × 10 6 8.16 × 10 6 8.52 × 10 6
Table 9. Error at different numbers of training points and subspace dimensions when k = 1 ,   α = 0.4 ,   T = 5 .
Table 9. Error at different numbers of training points and subspace dimensions when k = 1 ,   α = 0.4 ,   T = 5 .
Points 50 × 50 60 × 60 80 × 80 100 × 100 120 × 120
e L 3.31 × 10 6 5.44 × 10 6 3.44 × 10 6 8.32 × 10 7 7.11 × 10 6
Table 10. Error in the number of training points and subspace dimension at k = 1   , T = 5 .
Table 10. Error in the number of training points and subspace dimension at k = 1   , T = 5 .
α
Points
80 × 80 100 × 100 120 × 120 150 × 150
e L 2 e L e L 2 e L e L 2 e L e L 2 e L
301.07 × 10 9 3.55 × 10 7 5.42  × 10 10 1.80  × 10 7 8.25 × 10 10 3.12 × 10 7 1.23 × 10 9 2.82 × 10 7
401.28 × 10 9 4.11 × 10 7 1.23  × 10 9 3.81  × 10 7 1.39 × 10 9 3.16 × 10 7 2.33 × 10 9 6.27 × 10 7
501.11 × 10 9 2.38 × 10 7 1.15 × 10 9 3.56 × 10 7 1.39 × 10 9 2.89 × 10 7 9.38  × 10 10 2.13  × 10 7
Table 11. Comparison of the errors of different methods.
Table 11. Comparison of the errors of different methods.
MethodNetwork Structure α = 0.2 α = 0.4
T = 1 T = 5 T = 1 T = 5
e L 2 e L e L 2 e L e L 2 e L e L 2 e L
PINNs 2 , 100 , 100 , 100 , 100 , M = 600 , 1 1.02 × 10 2 2.46 × 10 2 1.17 × 10 2 2.30 × 10 + 0 1.91 × 10 2 4.60 × 10 2 8.91 × 10 3 2.31 × 10 + 0
lbPINNs 2 , 100 , 100 , 100 , 100 , M = 600 , 1 3.68 × 10 2 9.33 × 10 2 3.95 × 10 3 5.11 × 10 1 4.78 × 10 3 1.12 × 10 2 7.46 × 10 3 7.98 × 10 1
ELM 2 , M = 600 , 1 1.31 × 10 5 3.56 × 10 5 8.24 × 10 6 7.46 × 10 4 3.98 × 10 6 1.22 × 10 5 1.98 × 10 5 2.27 × 10 3
PIELM 2 , M = 600 , 1 2.53 × 10 6 6.60 × 10 6 1.11 × 10 6 9.06 × 10 5 1.67 × 10 6 4.47 × 10 6 1.04 × 10 6 9.69 × 10 5
FDM [31]-1.82 × 10 7 2.58 × 10 7 2.94 × 10 6 4.16 × 10 6 4.41 × 10 7 6.23 × 10 7 3.75 × 10 7 5.31 × 10 7
Present 2 , 100 , 100 , 100 , 100 , M = 600 , 1 3.09  × 10 9 1.05 × 10 8 1.85  × 10 8 1.43  × 10 6 2.09  × 10 9 6.72  × 10 9 8.86  × 10 9 8.32 × 10 7
Table 12. Error at k = 1 ,   m = 2 for different numbers of training points and M.
Table 12. Error at k = 1 ,   m = 2 for different numbers of training points and M.
Points
M
300400500600
305.34 × 10 13 5.82 × 10 13 4.44 × 10 13 3.30 × 10 13
407.53 × 10 13 6.23 × 10 13 7.53 × 10 13 6.35 × 10 13
504.41 × 10 13 5.87 × 10 13 5.55 × 10 13 8.11 × 10 13
604.03 × 10 13 6.60 × 10 13 6.04 × 10 13 5.18 × 10 13
804.44 × 10 13 6.21 × 10 13 5.54 × 10 13 3.64 × 10 13
Table 13. Error of m = 2 at different values of k .
Table 13. Error of m = 2 at different values of k .
Points608090100140180
k = 3 1.38 × 10 12 2.79 × 10 13 1.76  × 10 13 5.63 × 10 13 4.46 × 10 13 4.12 × 10 13
k = 5 × 10 2 3.28 × 10 13 2.87 × 10 13 3.97 × 10 13 2.59  × 10 13 3.57 × 10 13 3.35 × 10 13
k = 10 3 4.31 × 10 13 3.19 × 10 13 3.82 × 10 13 3.35 × 10 13 3.15 × 10 13 2.33  × 10 13
Table 14. Errors under different training points when m = 10 ,   k = 1 .
Table 14. Errors under different training points when m = 10 ,   k = 1 .
Points80100150200
e L 2 4.60 × 10 11 1.07  × 10 12 4.21 × 10 12 5.57 × 10 12
e L 7.68 × 10 11 2.34  × 10 12 6.64 × 10 12 6.43 × 10 12
Table 15. Errors under different training points when m = 50 ,   k = 1 .
Table 15. Errors under different training points when m = 50 ,   k = 1 .
Points3005008001000
e L 2 1.63 × 10 4 2.61 × 10 8 4.22 × 10 7 3.34 × 10 8
e L 2.39 × 10 4 3.16 × 10 8 6.30 × 10 7 3.59 × 10 8
Table 16. Errors under different training points when m = 100 ,   k = 1 .
Table 16. Errors under different training points when m = 100 ,   k = 1 .
Points1000150020005000
e L 2 2.05 × 10 7 2.54 × 10 7 8.14 × 10 8 2.93 × 10 7
e L 2.85 × 10 7 3.29 × 10 7 8.52 × 10 8 2.46 × 10 7
Table 17. Comparison of the errors and CPU time of different methods.
Table 17. Comparison of the errors and CPU time of different methods.
MethodNetwork Structure m = 2 m = 10 m = 50 m = 100
e L e L 2 e L e L 2 e L e L 2 e L e L 2
PINNs 1 , 100 , 100 , 100 , 100 , M = 600 , 1 3.37 × 10 4 3.13 × 10 4 4.37 × 10 1 5.65 × 10 1 2.42 × 10 + 1 2.73 × 10 + 1 1.63 × 10 + 0 1.27 × 10 + 0
lbPINNs 1 , 100 , 100 , 100 , 100 , M = 600 , 1 6.89 × 10 4 3.54 × 10 4 2.00 × 10 3 1.40 × 10 3 2.64 × 10 1 1.42 × 10 1 1.01 × 10 + 0 7.15 × 10 1
ELM 1 , M = 600 , 1 4.83 × 10 7 2.69 × 10 7 6.28 × 10 + 1 4.50 × 10 + 1 4.80 × 10 + 3 4.33 × 10 + 3 2.06 × 10 + 3 2.06 × 10 + 3
PIELM 1 , M = 600 , 1 1.92 × 10 9 8.42 × 10 10 4.69 × 10 + 1 4.16 × 10 + 1 1.99 × 10 + 2 1.35 × 10 + 2 1.99 × 10 + 2 1.35 × 10 + 2
FDM [32]-8.20 × 10 7 4.63 × 10 7 1.22 × 10 5 6.97 × 10 6 1.24 × 10 5 7.07 × 10 6 1.23 × 10 7 7.03 × 10 8
Present 1 , 100 , 100 , 100 , 100 , M = 600 , 1 3.30 × 10 13 1.82 × 10 13 2.34 × 10 12 1.07 × 10 12 3.16 × 10 8 2.61 × 10 8 8.52 × 10 8 8.14 × 10 8
Table 18. Comparison of the errors and CPU time at the different collocation points.
Table 18. Comparison of the errors and CPU time at the different collocation points.
N i n / N b c 100/100200/200300/300400/400500/500600/600
e L 2 1.09 × 10 8 7.90 × 10 10 9.84 × 10 10 1.09 × 10 9 6.39 × 10 10 1.15 × 10 9
e L 3.28 × 10 8 1.63 × 10 9 1.78 × 10 9 2.44 × 10 9 1.37 × 10 9 1.92 × 10 9
epochs792817740892733719
CPU(s)13.2218.0219.1831.9729.4431.76
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pei, F.; Cao, F.; Ge, Y. A Novel Neural Network-Based Approach Comparable to High-Precision Finite Difference Methods. Axioms 2025, 14, 75. https://doi.org/10.3390/axioms14010075

AMA Style

Pei F, Cao F, Ge Y. A Novel Neural Network-Based Approach Comparable to High-Precision Finite Difference Methods. Axioms. 2025; 14(1):75. https://doi.org/10.3390/axioms14010075

Chicago/Turabian Style

Pei, Fanghua, Fujun Cao, and Yongbin Ge. 2025. "A Novel Neural Network-Based Approach Comparable to High-Precision Finite Difference Methods" Axioms 14, no. 1: 75. https://doi.org/10.3390/axioms14010075

APA Style

Pei, F., Cao, F., & Ge, Y. (2025). A Novel Neural Network-Based Approach Comparable to High-Precision Finite Difference Methods. Axioms, 14(1), 75. https://doi.org/10.3390/axioms14010075

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop