1. Introduction
Differential equations find extensive applications in diverse fields such as fluid dynamics, biology, and finance. For instance, in fluid dynamics, the Navier–Stokes equations serve as fundamental tools for modeling fluid behavior [
1]. In epidemiology, ordinary differential equations, such as those based on the Susceptible–Infectious–Recovered (SIR) model, are indispensable for analyzing the spread of infectious diseases [
2]. Additionally, in finance, the Black–Scholes model relies on differential equations to facilitate option pricing [
3]. However, computing solutions for high-dimensional differential equations presents a significant challenge. This challenge arises from the curse of dimensionality, wherein traditional numerical methods entail discretizing equations at grid points. As the dimension of equations increases, the number of grid points grows exponentially. Consequently, even basic linear algebra computations within these numerical algorithms become prohibitively complex.
A promising strategy for mitigating this complexity involves using quantum computing techniques since quantum computer leverages unique features such as quantum entanglement and superposition, enabling exponential speedups in specific computational tasks. As previous research has pointed out, replacing the linear systems solvers in classical numerical methods with their quantum counterparts, the Quantum Linear Systems Algorithm (QLSA), [
4] can yield provably efficient quantum algorithms for solving linear ordinary differential equations (ODEs) [
5,
6,
7,
8] and linear Partial Differential Equations (PDEs) [
7,
9,
10,
11]. Additionally, quantum algorithms based on Hamiltonian simulation, rather than the QLSA, have also been theoretically shown to efficiently solve computational problems involving linear differential equations, whether for the standard Hamiltonian simulation or non-Hermitian Hamiltonian simulation [
12].
However, despite a series of theoretical successes in developing quantum algorithms for linear differential equations, addressing nonlinear cases remains an open problem. Early attempts [
13] to directly combine QLSA with nonlinear ODEs proved to have poor scaling with evolution time. Recently, Liu et al. [
14] demonstrated significant improvements by introducing Carleman Linearization to certain nonlinear ODEs. Using QLSA on the linearized differential equations greatly enhanced the complexity bound with respect to the evolution time. This approach has motivated many researchers to consider first employing linearization techniques, such as Carleman Linearization [
15,
16,
17], Koopman von Neumann linearization [
18,
19,
20] and the level set method [
21], to approximate or represent low-dimensional nonlinear differential equations with high-dimensional linear ones, and then using quantum algorithms designed for linear differential equations to solve these problems.
Figure 1 illustrates the work flow of these various quantum algorithms for solving nonlinear ODEs.
As highlighted in reference [
22], the Carleman matrix is, in fact, a transposed finite-section matrix approximation of the Koopman Operator’s generator on a polynomial basis. This specific connection and the weaker performance of Carleman Linearization in nonpolynomial cases motivate our research. Despite this connection, modern research on the Koopman Operator and Carleman Linearization follows distinct paths. Investigations into the Koopman Operator predominantly adopt a data-driven approach, as they use spatiotemporal data without requiring the explicit equation to find an approximation of the principal Koopman modes and, correspondingly, Koopman eigenpairs. In particular, numerical schemes such as dynamic mode decomposition (DMD) [
23,
24,
25] and its diverse variations, including Extended Dynamic Mode Decomposition (EDMD) [
26], make use of the data collected from time snapshots of a specified dynamical system. These approaches employ Singular Value Decomposition (SVD) and other matrix decomposition techniques to reveal the temporal and spatial–spectral attributes of the system. Expanding beyond DMD, EDMD, and its variants, methods for obtaining tractable representations have been further connected with deep neural networks [
27,
28]. Two specific neural network architectures have been explored in this context. One is based on the autoencoder principle, employing a low-dimensional latent space to enhance the interpretability of outcomes. The other architecture involves a higher-dimensional latent space, which often yields improved results when dealing with dynamical systems featuring continuous spectra. A comprehensive review of the theoretical framework and algorithms related to Data-Driven Koopman Operators is provided by [
29].
Diverging from the previously discussed data-driven methods, our approach constitutes a progressive evolution of the linearization technique. This distinction mainly arises from our method of constructing the infinitesimal generator for the approximated Koopman operator. Rather than relying on data, we utilize the values from the right-hand side function of the ODEs at various discrete spatial points, thereby necessitating an explicit equation. More precisely, our method integrates the differentiation matrix in the spectral method [
30,
31], yielding an explicit matrix approximation of the Koopman Operator’s generator. Subsequently, this matrix generates an explicit higher-dimensional linear equation that closely approximates the original nonlinear system. In this context, the identical approximation matrix was employed across various scalar observable configurations, with only the initial value being altered. This innovative approach was motivated by the spectral method for differential equations discussed in [
32,
33]. Therefore, like other spectral methods, our approach’s effectiveness and convergence rate are influenced by the choice of the basis function. Potential options for the elements of D include polynomials [
34], Fourier modes [
32], radial basis functions [
35], and spectral elements [
36]. The best choice of basis functions is likely dependent on the specific characteristics of the underlying dynamical system and the data sampling strategy. Specifically, we adapt the Chebyshev Differentiation Matrix for demonstration purposes. This study focuses on autonomous systems and potentially serves as an alternative approach applicable in situations where Carleman Linearization has been conventionally utilized.
In addition to applications in quantum computing, our methods also hold potential for applications in traditional computational contexts. The most conventional method of linearization involves using a first-order Taylor expansion of the system dynamics. This technique, while proven effective and widely utilized in areas such as Model Predictive Control (MPC) and related domains, is limited to applications within a very narrow vicinity around the expansion point. To address these limitations more comprehensively, a range of more sophisticated linearization techniques such as Carleman Linearization [
37] and the Koopman Operator [
38] have been utilized. These methods were originally applied in the fields of dynamical systems and control [
22].
The paper is structured as follows:
Section 2 presents the background knowledge and discusses the Koopman Spectral Linearization Method in
Section 3. In
Section 4, we provide the numerical results, and the subsequent sections delve into the discussion and conclusions in
Section 5.
4. Numerical Result
This section presents the performance comparison of Koopman Spectral Linearization against Carleman Linearization on five nonlinear dynamic models in
Section 4.1. In each example, we visualized the curves of specific solution components and investigated the influence of truncation order
N on accuracy. As
N increases, both the precision of linearization and the computational effort increase correspondingly. In practical applications, the selection of parameter
N should be based on the required accuracy and available computational resources. Since Carleman Linearization can only be applied to polynomial cases, all the following nonpolynomial examples are converted to polynomial form using higher-order Taylor expansions before computation. The reference solution is generated by The 4th order Runge–Kutta method (RK4) is used if no closed-form solution is available. We also aim to compare the loss of accuracy in the linearization step under the assumption of an ideal quantum computing environment, where a high-precision quantum linear solver or other quantum algorithms for linear ODEs are available. Hence, the solutions of the resulting linear ODEs are directly computed using the
exp function in MATLAB 2022a. Next, in
Section 4.2, we further compare the matrix size and computational cost (all the MATLAB codes can be downloaded at
https://github.com/DongDongShiShi/Koopman-Comparison (accessed on 30 May 2024).
4.1. Linearize Dynamics with Koopman Spectral Linearization
4.1.1. Quadratic Model
We start with a simple nonlinear ODE, and the governing ODE is given by
We set
and T = 10 in this example. This quadratic model has a closed form solution
. In the tests, we fix
.
Figure 2 summarizes the results of the first comparison.
Figure 2b shows the exponential convergence of our approach concerning the truncation order, which is similar to the conclusion in the conventional Carleman Linearization method. Further, our approach shows a faster convergence rate.
4.1.2. Cosine Square Model
This cosine square model is a synthetic model invented as a nonpolynomial case for our demonstrative purposes. The governing equation is given as
And we set
,
. Despite the nonlinear nature, we still have a closed-form solution
. We fix
in the tests. It is important to note that as the system evolves over time, errors gradually accumulate. This is particularly evident when using lower-order polynomial interpolation (i.e.,
). Specifically, in
Figure 3a, beyond
, the curve begins to exhibit a more pronounced deviation from the standard solution.
Figure 3b shows the exponential convergence of our approach concerning the truncation order, which is still similar to the conclusion in the previous example. Conversely, for this nonpolynomial model, the accuracy of Carleman Linearization is very low, and we did not observe any significant changes in accuracy across the different values of
N.
4.1.3. Simple Pendulum Model
The simple pendulum is a well-studied model in science and engineering. The movement of the pendulum is described by the following second-order ordinary different equation:
where L is the length of the pendulum,
is the displacement angle, and the parameter g is the gravity acceleration. This second-order equation can be converted to a two-dimensional first-order ODE system. As a notation, we define
and
. Also, we set
for simplicity. Correspondingly,
and we set
,
. We fix
in the tests.
Figure 4 presents the numerical result of a simple pendulum. Similar to the previous result, the exponential convergence of error concerning truncation order
N is observed from
Figure 4b. Although the accuracy of Carleman Linearization for this nonpolynomial model is significantly higher compared to its application on the cosine square model and even more accurate than Koopman Spectral Linearization at
and
, the accuracy does not significantly change as
N increases.
4.1.4. Lotka–Volterra Model
The Lotka–Volterra model, also known as the predator–prey model [
40], is a mathematical model used to describe the interactions between predators and prey in an ecosystem. The model describes the interaction between two fundamental populations: prey and predator populations. Here, we define as below:
We set , and .
Figure 5 presents the results of these tests for the Lotka–Volterra model. Again, errors are decreased exponentially with respect to the truncation order. For this polynomial model, the accuracy of Carleman Linearization improves with increasing N. However, regarding both numerical accuracy and convergence speed, Koopman Spectral Linearization shows a significant advantage.
4.1.5. Kraichnan–Orszag Model
The Kraichnan–Orszag model was presented in [
41] for modeling fluid dynamics. This three-dimensional nonlinear model is given by
We set −, and .
Figure 6 presents the results of the Kraichnan–Orszag model. Different from the previous example, a shorter time period is set due to the strong oscillations of the Kraichnan–Orszag model.
The truncation order testing figures illustrate the error as a function of the truncation order
N and provide a comparison between Carleman Linearization and our method. The mistake of Carleman Linearization in polynomial cases (i.e.,
Section 4.1.1,
Section 4.1.4 and
Section 4.1.5) exhibits exponential decay, a phenomenon that has been substantiated by recent theoretical research [
42,
43]. Similarly, our method leverages the Chebyshev node interpolation approach, which also enjoys exponential convergence guarantees and demonstrates similar exponential convergence in numerical experiments.
For all the polynomial examples in
Section 4.1.1,
Section 4.1.4 and
Section 4.1.5, it is evident that our approach exhibits higher accuracy at the same truncation order, and the error decreases more rapidly as the truncation order is increasing. Moreover, when applied to nonpolynomial dynamical systems, our method showcases superior numerical accuracy and faster convergence rates. In
Section 4.1.2 and
Section 4.1.3, we present a comparative analysis of the error–truncation order relationship in nonpolynomial dynamical systems. Both nonpolynomial dynamics are subjected to 12th-order Taylor expansions to transform them into polynomial forms. The Carleman method exhibits significantly lower accuracy in this scenario, with notably slower convergence. This limitation arises because in nonpolynomial cases, the relevant part of the Taylor expansion is confined to orders lower than the truncation order, resulting in substantial errors when employing relatively small expansion orders as demonstrated in our study. Consequently, our method demonstrates heightened precision in nonpolynomial cases. Even in the case of the ‘simple pendulum’ where
N = 3, our method initially exhibits lower accuracy. However, as
N increases to 9, the errors obtained by our method rapidly become dramatically smaller than those produced by the Carleman method.
4.2. Matrix Size and Computational Cost Comparison
Following the construction procedure introduced in
Section 3, we apply the tensor product rule to construct collocation points. Consequently, the approximation matrix of the Koopman generator has a size of
. However, we can reuse the lifted matrix with different initial values for multi-dimensional dynamics. In the case of the Carleman Linearization procedure, the matrix size will be
, and we can approximate the size of the Carleman lifted matrix as
. Therefore, when considering the matrix size of the derived linear system with a relatively small truncation order
N, the matrix in our approach can be significantly smaller than the matrix in Carleman Linearization. To further illustrate this point,
Figure 7 compares our approach with the size of the Carleman Linearization matrix.
While our approach generates
d linear systems with the same matrix but different initial values, which may not be as convenient as Carleman Linearization, where only one linear system is produced, the significant difference in matrix sizes results in a notable increase in computational cost. As an illustrative example, we present the running time and error in
Table 1 for a truncation order of
N = 9. The computational time required for Carleman Linearization is slower than our approach, especially when
d = 2 or 3. In particular, for the Kraichnan–Orszag model (
d = 3), the Carleman Linearization procedure took 32.0546 s, being considerably slower than our approach.
On the other hand, our method’s capacity to yield numerical results with smaller matrices, reduced computational time, and increased accuracy, all under the same truncation order, hinges upon a critical prerequisite—namely, the selection of an appropriate parameter r, a requirement not present in the Carleman Linearization method. Furthermore, when only the time-domain evolution differential equations of the dynamical system are available in advance, without access to spatial evolution information of the state, our method cannot capture the state’s spatial evolution. Therefore, unless supplementary information is accessible regarding the spatial evolution of the dynamical system, allowing for a reasonable estimation of r and surpassing the precision of Carleman Linearization remains unattainable.
5. Discussion and Conclusions
The Koopman Spectral Linearization method explicitly uses the differentiation matrix to derive the lifted matrix of nonlinear autonomous dynamical systems. It provides a finite-dimensional representation for the generator of the Koopman Operator with a polynomial basis. Therefore, similar to Carleman Linearization, our approach offers a linearization method for nonlinear autonomous ordinary differential equations. In each numerical experiment presented, our approach exhibits exponential convergence with respect to truncation order N, just like Carleman Linearization. Under the same truncation order, Koopman Spectral Linearization tends to exhibit significantly higher accuracy associated with lower time cost compared to Carleman Linearization, especially when the dynamics are not polynomial. Therefore, it is more suitable as an alternative linearization method where high-accuracy approximations are needed and f is nonpolynomial. Different from Carleman Linearization, which describes the evolution of the state itself, our approach is used to describe a scalar smooth observable, which is more flexible.
However, despite its advantages, the Koopman Spectral Linearization method has some limitations. It is essential to consider certain drawbacks when evaluating its applicability since a reasonable estimation of radius r is necessary to obtain an accurate approximation of the eigenfunctions. We employ interpolation within a local space of radius r around the initial point to approximate the eigenfunctions across in a local domain. However, since the spatial range of the system evolution during the process is not known in advance, it is very challenging to select the value of r without additional information; when theoretical guidance is unavailable, r is chosen empirically as a relatively small value (i.e., 0.1–0.5) to ensure a higher precision of approximation. If the spatial range of the system’s evolution is known beforehand, such as the case of fluid moving in a pipeline, r can then be selected based on spatial constraints. To further understand how this parameter r influences the accuracy, a comprehensive numerical analysis will be included in future work. In addition, the global interpolation method might overcome the limitation of the estimate r. Besides concerns regarding the parameter r, the degree of nonlinearity in physical systems can vary significantly, which might also influence the performance of our linearization method. Particularly for highly nonlinear systems such as turbulence, no linearization technique is capable of providing high-precision linear approximations over extended periods and large areas. In such instances, data-driven approaches may offer more satisfactory solutions, addressing the complexities inherent in these systems more effectively.
Regarding the matrix size as indicated in
Section 4.2, the Koopman Spectral Method obtains a much smaller matrix with even higher accuracy compared with Carleman Linearization. However, the matrix size is exponential increased with respect to dimension
d (i.e.,
) since the tensor product rule is applied. Therefore, both Carleman Linearization (i.e.,
) and our approach might not be efficient for the relatively high-dimensional problems on classic computers. A possible way to overcome this limitation is to adapt sparse grid methods to construct collocation points, which has been found successful in [
30]. As pointed out in the Koopman Operator Theory, Carleman Linearization and our approach rely on tensor product rule; therefore, they need further modification to deal with high-dimensional problems on classical computers. However, on quantum computers, tensor products may not be a challenge, which may lead to innovative designs of algorithms that are completely different from their counterpart on classical computers. This is the main reason we conduct these comparisons.
As previous sections noted, the underlying principle (Schrödinger equation) of quantum computers is linear. Consequently, there are fundamental difficulties in using quantum computing to solve nonlinear ODEs/PDEs. Currently, the linearization of differential equations appears to be an unavoidable step in addressing this challenge. In particular, Carleman Linearization as a solution strategy has garnered significant attention in recent research [
17]. By reviewing the underlying theory of Carleman Linearization—Koopman Operator Theory—and integrating it with the concept of differentiation matrices in spectral methods, we propose the Koopman Spectral Linearization method. Our novel linearization approach might be useful when designing quantum algorithms for nonlinear ODEs/PDEs under certain situations.