1. Introduction
Partial differential equations (PDEs) play a central role in the natural sciences, engineering, mathematics, and related fields, providing an effective tool for describing and analyzing complex phenomena and systems. These equations not only help us understand natural laws, but they also lay the foundation for scientific research and engineering practices [
1,
2]. Since many PDEs do not have analytical solutions, traditional numerical methods, such as the finite element method (FEM), finite difference method (FDM), finite volume method (FVM), etc., are widely used to compute approximate solutions, assisting researchers and engineers in obtaining meaningful results for various problems.
High-order compact FDMs are widely used for solving PDEs due to their high accuracy and effectiveness. These schemes typically achieve a fourth-order or higher spatial accuracy with smaller truncation errors compared to traditional methods. They require fewer mesh points to achieve the same accuracy, which is particularly beneficial for high-dimensional problems and helps reduce computational costs [
3,
4,
5,
6,
7]. However, the construction and implementation of compact difference formats are complex, especially when dealing with intricate boundary conditions. They often assume that the solution is smooth, which may not be valid for problems with sharp gradients or discontinuities, potentially leading to unstable or inaccurate results. Additionally, many compact methods are implicit, requiring the solution of larger systems of linear equations to be solved, which increases the implementation complexity and computation time, particularly for large-scale problems.
In recent years, deep learning has gained traction in solving PDEs, leading to the development of various innovative methods, such as physically informed neural networks (PINNs) [
8], the deep Ritz method (DRM) [
9], deep Galerkin methods (DGMs) [
10], and extreme learning machines (ELMs) [
11], Self-adaptive loss balanced Physics-informed neural networks (lbPINN) [
12]. These approaches demonstrate significant advantages in addressing complex PDE problems. PINNs integrate physical laws into the neural network’s loss function, enabling the model to learn from both data and the governing PDE during training. This method leverages the neural network’s strong fitting capabilities to represent the PDE solution, minimizing the error between the predicted and actual solutions. PINNs are particularly effective for problems with complex boundary conditions and multi-scale features, often yielding good results even with limited training data [
13,
14]. DGMs combine deep learning with Galerkin techniques to approximate PDE solutions through neural networks that satisfy the weak form of the equations. ELMs are single hidden-layer feedforward neural networks known for their simple and rapid training process, ideal for large-scale data. Unlike traditional neural networks, ELMs randomly select hidden layer weights and determine output weights using least-squares, significantly reducing the training time. They can quickly adapt to varying boundary and initial conditions when solving PDEs.
In addition, the extrapolation—driven network architecture—proposed by Wang, Y. [
15] optimizes the physics—informed deep—learning model by combining extrapolation techniques, further enhancing the model’s accuracy and generalization ability. Zeng, B. [
16] developed a method of integrating physical encoding into the message—the passing graph network (PhyMPGN), which is specifically designed to solve the spatiotemporal PDE system. These methods demonstrate the great potential of deep learning in the field of numerically solving PDEs, especially when traditional numerical methods encounter bottlenecks.
These deep learning methods offer fresh perspectives and tools for tackling PDEs, especially in scenarios where traditional numerical methods struggle. They effectively handle complex boundary conditions and nonlinear characteristics while minimizing mesh dependency. However, these approaches often require substantial data and computational resources, raising concerns about their feasibility and efficiency in practical applications [
17,
18,
19,
20]. Moreover, most deep learning-based methods necessitate optimizing the loss function to approximate the solutions accurately. The complexity of the networks and the number of hyperparameters can lead to optimization challenges, potentially resulting in local optima and an insufficient accuracy. To address these issues, researchers need to employ refined strategies in model design and training, such as advanced optimization algorithms, network structure adjustments, and regularization techniques, to enhance the generalization and accuracy [
21,
22,
23,
24].
Compared to the optimization loss function-based approach of PINNs, the ELM-based basis function method effectively avoids convergence issues in the optimization process and demonstrates a significant improvement in accuracy. This strategy reduces the reliance on traditional optimization algorithms by utilizing randomly generated weights and biases, resulting in faster computations and an enhanced model stability. Additionally, the ELM-based model exhibits a superior adaptability in managing complex boundary and initial conditions, making it more advantageous for addressing practical problems. Dwivedi [
25] developed physics-informed extreme learning machines (PIELMs), which are fast versions of PINNs that can be applied to static and time-varying linear partial differential equations to meet or exceed the accuracy of PINNs on a range of problems. Quan [
26] proposed a novel learning method based on the extreme learning machine algorithm, which utilizes the ELM algorithm to solve a system of linear equations, thereby determining the network parameters for solving PDEs. Wang et al. [
27] presented an efficient extension of the extreme learning machine (ELM) method, moving from low dimensions to solving high-dimensional partial differential equations (PDEs). A neural network-based solution for linear and nonlinear partial differential equations has been proposed by integrating concepts from extreme learning machines (ELMs), domain decomposition, and local neural networks [
28]. A high-precision subspace method based on neural networks has been proposed for solving partial differential equations. The core idea involves using neural network-based functions as the basis functions to span a subspace, allowing for the approximation of solutions within that subspace [
29]. While neural networks hold significant promise for numerically solving PDEs, challenges like model generalization remain. Integrating neural networks with traditional numerical methods offers new avenues for achieving high-precision solutions, likely resulting in more efficient and accurate computations across various fields.
In light of the findings in the literature, we propose a novel deep learning strategy that integrates the ELM with the basis function space method. This approach leverages the advantages of the ELM to enhance the solution process and improve the accuracy of solving PDEs. Firstly, the ELM can be seen as selecting specific basis functions in the function space through randomly generated weights and biases. This not only enriches the expressiveness of the function space, but also significantly boosts the problem-solving efficiency, allowing for the better capture of the PDE solution characteristics and an improved model accuracy. Secondly, the ELM demonstrates an excellent adaptability, enabling flexible adjustments to various boundary and initial conditions. This adaptability is crucial in basis function space methods, as many real-world problems often involve complex and variable conditions. By incorporating the ELM, our model can maintain a strong performance across different physical scenarios, broadening its applicability. Finally, the ELM’s training time is substantially shorter than that of traditional neural networks, greatly enhancing the computational efficiency when addressing complex PDEs. This rapid training process reduces the computational costs and facilitates real-time or near real-time applications, particularly in engineering and scientific computing contexts that demand quick responses. In summary, our proposed deep learning strategy that combines an ELM with the basis function space method not only enhances the efficiency and accuracy of PDE solving, but also offers new insights and tools for tackling complex physical problems.
This paper is organized as follows:
Section 2 introduces the model for partial differential equations and the fundamental principles of PINNs.
Section 3 explores a high-precision basis function method based on neural networks.
Section 4 evaluates the accuracy and effectiveness of the numerical validation method. The conclusion is presented in
Section 5.
3. The Neural Network-Based Basis Function Space Method
The universal approximation theorem demonstrates the remarkable expressive power of neural networks, indicating that even a simple two-layer neural network can approximate complex functions to an arbitrary accuracy. Building on this foundation, we utilized the concept of extreme learning machines (ELMs) to generate basis functions through a neural network framework. These basis functions served as the building blocks for constructing our trial solution. Once the trial solution was formulated from the basis functions, we enhanced its accuracy by incorporating traditional numerical methods, such as least-squares optimization, as shown in
Figure 2. This hybrid approach allows us to leverage the strengths of both machine learning and numerical analyses, resulting in a more robust and accurate solution to the problem at hand. By effectively combining neural network-based basis functions with established numerical techniques, we can achieve significant improvements in accuracy while maintaining computational efficiency. This methodology not only streamlines the process of finding solutions to complex equations, but also expands the potential applications of neural networks in solving various mathematical and engineering problems. Through this integration, we aimed to harness the full potential of neural networks and traditional numerical methods, paving the way for the more effective and accurate approximation of complex functions.
A feedforward neural network is defined with
hidden layers, where the
-th hidden layer contains
neurons, the mathematics of which can be expressed by the following equation:
Given an input , the output of the network is computed through a series of weights and activation functions. and represent the weight and bias, respectively. is the activation function, is a set of parameters in the neural network, and is the output with parameters . The initial set of parameters for a neural network are usually randomly generated.
From Equation (6), it is evident that the neural network’s output can be viewed as a combination function of the last layer’s weight coefficients and the output of the penultimate hidden layer’s neurons. This combination allows the network to capture complex input–output relationships and perform nonlinear mapping. Consequently, this multi-layer structure facilitates layer-wise feature extraction, enhancing the model’s expressive power and learning efficiency.
Assuming that the number of neurons in the penultimate hidden layer is M, then the output of M neurons can be viewed as a set of basis functions and denoted as
and the function space
V can be written as
The weights of the last hidden layer can be expressed as
Thus, the output of the neural network can be understood as a function formed by a linear combination of the basis functions, represented as
Given the diverse types of PDEs and their corresponding solutions, it is essential to train the basis functions appropriately to enable them to express the solutions of PDEs more effectively. This training process focuses on optimizing the parameters of the basis functions, allowing them to more accurately capture the unique characteristics of the solutions for specific PDEs. We used the PINN approach to construct the loss function for pre-training the neural network. Due to the complexity of the PINN optimization process, we established a stopping condition: training was halted when the loss function decreased by two orders of magnitude. This strategy ensured that the network achieved a higher accuracy while minimizing the risk of overfitting, thereby enhancing the model’s generalization ability. Generally, the loss function is as follows:
Sufficient training was conducted on the loss function (11) until the maximum number of training steps was reached or the following condition was met:
where
is the initial loss error. We adopted
and the maximum number of training steps was
.
After completing the pre-training of the network (12), we obtained an initial set of optimal weights that satisfactorily met the requirements of the partial differential control equations and their boundary conditions (13). Using these weights, we constructed customized basis functions tailored for specific boundary value problems involving partial differential equations.
These basis functions are crucial for their adaptability, as they are specifically designed to effectively address particular problems and accurately capture the solution characteristics. Accordingly, the trial solution output by the neural network can be expressed as
A set of neural network-based basis functions was obtained by optimizing the loss function, as shown in Equation (14), and then the approximation solution could be expressed as a linear combination of this set of basis functions, as shown in Equation (15). Next, the combination coefficients were optimized to give a higher approximation accuracy.
Substituting the trial solution in Equation (15) into the governing equation, Equation (1), and making the control equation satisfied at each collocation point yielded the following system of equations:
This is a system of linear equations for the combined weight coefficients
, which can be written in matrix form as
It can be further abbreviated as
where
,
, and
.
This is a system of hyper-deterministic linear equations, and we can solve this system using the least-squares method to obtain the optimal combination of coefficients.
The optimized trial solution will approximate the analytical solution of the PDEs with a high accuracy.
6. Conclusions
Neural networks have achieved significant breakthroughs in various fields, particularly in solving complex PDEs. However, deep learning methods that optimize loss functions often face challenges like long training times, local optima, and a limited accuracy. This paper introduces a novel strategy inspired by extreme learning machines that uses neural networks to generate basis functions and optimize trial solutions. Our method demonstrates substantial improvements in both efficiency and accuracy compared to conventional approaches like finite difference methods and other neural network techniques such as PINNs, lbPINNs, ELM, and PIELM. The results indicate that our approach matches the accuracy of traditional methods while significantly reducing the number of iterative steps required. This advancement not only enhances the efficiency of solving PDEs, including complex domain problems, but it also challenges preconceived notions about the accuracy of neural networks. Overall, this breakthrough opens new avenues for research and applications in using neural networks for PDE solutions.
Though the proposed method has demonstrated advantages in terms of accuracy and efficiency in solving smooth PDE problems with single equations, further exploration is needed to address the challenges posed by a broader range of PDE problems with discontinuous solutions, as well as those exhibiting strongly nonlinear characteristics.