Next Article in Journal
Bounded Perturbation Resilience of Two Modified Relaxed CQ Algorithms for the Multiple-Sets Split Feasibility Problem
Next Article in Special Issue
Improved Block-Pulse Functions for Numerical Solution of Mixed Volterra-Fredholm Integral Equations
Previous Article in Journal
Global Directed Dynamic Behaviors of a Lotka-Volterra Competition-Diffusion-Advection System
Previous Article in Special Issue
Periodic Property and Instability of a Rotating Pendulum System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Tikhonov-Regularized Variable Projection Algorithm for Separable Nonlinear Least Squares

1
College of Geodesy and Geomatics, Shandong University of Science and Technology, Qingdao 266590, China
2
College of Mathematics and Systems Science, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
Axioms 2021, 10(3), 196; https://doi.org/10.3390/axioms10030196
Submission received: 17 July 2021 / Revised: 18 August 2021 / Accepted: 19 August 2021 / Published: 22 August 2021

Abstract

:
In this work, we investigate the ill-conditioned problem of a separable, nonlinear least squares model by using the variable projection method. Based on the truncated singular value decomposition method and the Tikhonov regularization method, we propose an improved Tikhonov regularization method, which neither discards small singular values, nor treats all singular value corrections. By fitting the Mackey–Glass time series in an exponential model, we compare the three regularization methods, and the numerically simulated results indicate that the improved regularization method is more effective at reducing the mean square error of the solution and increasing the accuracy of unknowns.

1. Introduction

Many problems in physics, chemistry, machine learning, computer vision, signal processing and mechanical engineering can only be described by a specialized type of nonlinear regression model, which is a linear combination of nonlinear functions. In particular, given the sequence of the observed values { ( t i , y i ) , i = 1 , 2 , , m } , a fitting model can be established as follows:
η ( θ L , θ N ; t i ) : = j = 1 n a j φ j ( θ N ; t i )
where θ L = ( a 1 , a 2 , , a n ) T is the linear parameter, θ N = ( b 1 , b 2 , , b k ) T is the nonlinear parameter and φ j ( θ N ; t i ) is a quadratic differentiable nonlinear function, which depends on θ N and t i . With the least squares method, it reduces to the following nonlinear function.
v ( θ L , θ N ) = i = 1 m ( y i η ( θ L , θ N ; t i ) ) 2
which can be rewritten in the matrix form
v ( θ L , θ N ) = y Φ ( θ N ) θ L 2
where y = ( y 1 , y 2 , , y m ) T , t = ( t 1 , t 2 , , t m ) T and ϕ j ( θ N ; t ) = ( φ j ( θ N ; t 1 ) , φ j ( θ N ; t m ) ) T . Additionally, Φ ( θ N ) = ( ϕ 1 ( θ N ; t ) , ϕ n ( θ N ; t ) ) is a matrix function, and 2 denotes the Euclidean norm.
The minimization problem (3) is a nonlinear least squares problem that does not consider the properties of the parameters. The Gauss–Newton (GN) algorithm [1], the Levenberg–Marquardt (LM) algorithm [2,3] and the iterative embedding points algorithm [4] are commonly used to solve such problems. In fact, there are two sets of mutually dependent parameters in this model: the linear parameter and the nonlinear parameter. Golub and Pereyra [5,6] refer to this type of data-fitting problem as a separable nonlinear least squares (SNLLS) problem. In consideration of the specialized separable structure of this type of model, they proposed the variable projection (VP) algorithm. The general aim of implementing this algorithm is to eliminate linear parameters and obtain a simplified problem with only nonlinear parameters. For any given nonlinear parameter θ N , θ L can be obtained by solving the following linear least squares problem.
θ L = arg min θ L y Φ ( θ N ) θ L 2
The least squares solution of Equation (4) is
θ ^ L = ( Φ ( θ N ) T Φ ( θ N ) ) 1 Φ ( θ N ) T y = Φ ( θ N ) + y
where Φ ( θ N ) + is the pseudo-inverse of Φ ( θ N ) . By substituting θ ^ L into (3), we obtain
v ( θ N ) = y Φ ( θ N ) Φ ( θ N ) + y 2 = ( I Φ ( θ N ) Φ ( θ N ) + ) y 2 = ( I P Φ ( θ N ) ) y 2 = P Φ ( θ N ) y 2
where P Φ ( θ N ) = Φ ( θ N ) Φ ( θ N ) + is the orthogonal projection on the linear space spanned by the column vectors of Φ ( θ N ) , and P Φ ( θ N ) = I P Φ ( θ N ) is the projector on the orthogonal complement of the column space of Φ ( θ N ) .
Equation (6) is the revised residual function, in which the linear parameter is eliminated. The VP algorithm reduces the dimensionality of the parameter space and increases the probability of finding the global optimal solution. It is an effective method for solving the SNLLS problem.
The VP algorithm has been significantly improved and widely applied since it was proposed. For example, Kaufman [7] proposed an improved VP algorithm based on the trapezoidal decomposition of a matrix, and provided a simplified Jacobian matrix of the VP method. Ruhe et al. [8] analyzed the asymptotic convergence of the simplified VP algorithm. O’Leary and Rust [9] subsequently discovered that a VP algorithm with a complete Jacobian matrix requires fewer iterations than Kaufman’s simplified algorithm. Alternatively, Ruano et al. [10] proposed a more simplified Jacobian matrix for the VP algorithm, and an improved VP algorithm that entailed applying QR decomposition to the sparse nonlinear function matrix. They found that their method effectively increased computational efficiency. Gan et al. [11] compared the separation algorithms for different Jacobian matrices and concluded that a VP algorithm with the complete Jacobian matrix is more stable than the simplified algorithm proposed by Kaufman. Gan and Li [12] proposed a VP algorithm that utilizes the classic Gram–Schmidt (GS) matrix decomposition method to treat cases in which the number of observations is significantly larger than the number of linear parameters; their algorithm was found to reduce the computational cost. Alternatively, Chen et al. [13] employed a modified GS method in their development of a more robust VP algorithm for SNLLS problems; they reported that the combination of a VP algorithm and the L–M method is more effective than a combination of the VP algorithm and the G–N method.
Although many considerable efforts have been dedicated to developing methods to solve SNLLS problems, there are few studies on ill-conditioned problems with iterative processes. It can be seen that the least squares solution given as Equation (5) is conditioned on the fact that ( Φ ( θ N ) T Φ ( θ N ) ) 1 Φ ( θ N ) T is computable—that is, the matrix Φ ( θ N ) T Φ ( θ N ) is invertible. However, when the equation is ill-conditioned, the smallest characteristic root of the matrix Φ ( θ N ) T Φ ( θ N ) is near zero; therefore, the elements of ( Φ ( θ N ) T Φ ( θ N ) ) 1 will become significantly large. Small changes in the observation value will significantly affect the least squares solution, causing it to become unstable. Thus, a regularization method needs to be introduced to convert the ill-posed problem into a relatively mild or benign problem [14,15,16] before proceeding. There are many regularization methods, including Tikhonov regularization (TR) [17,18], truncated singular value decomposition (TSVD) regularization [19,20], kernel-based regularization [21,22] and l 1 -regularization [23]. Generally, a regularization method is applied to improve the condition of the ill-conditioned matrix by introducing regularization parameters. Common methods for determining regularization parameters include ridge-tracing, generalized cross-check and L-curve-based methods. The TSVD method is significantly popular because it is relatively well developed, and can be applied to solve computationally complex problems. Zeng et al. [24] estimated the linear parameters in a separable least squares parameter optimization process by implementing regularization parameter detection techniques in the TR and TSVD methods. In their attempt to regularize SNLLS ill-posed problems using a VP algorithm, Chen et al. [25] developed a weighted generalized cross-validation method to determine TR parameters. The effectiveness of the algorithm was verified through experimentation. Wang et al. [26] separated linear and nonlinear regularization parameters using a singular value decomposition (SVD)-based VP algorithm. They also estimated the linear parameters by using both the LS method and an L-curve-based spectral correction method. Their experiments confirmed that the algorithm can effectively solve ill-conditioned problems.
In this paper, an improved TR optimization method is proposed for the parameter estimation problem of SNLLS models. With the VP method as the basic framework, the algorithm was developed to take into account the specialized structure of the linear combination of nonlinear functions and separate the linear and nonlinear parameters. The linear parameters are estimated by using an improved TR method, whereas the nonlinear parameters are optimized by using the LM method. Regarding the iterative process, the simplified Jacobian matrix proposed by Kaufman is implemented in the VP algorithm. Numerical simulations were performed to compare the improved TR method to the original TR and TSVD regularization methods. The effectiveness of the VP algorithm with the improved TR method was verified, and the accuracies of different linear parameter estimation methods were evaluated.
We begin this paper by summarizing the VP method and the existing problems related to solve SNLLS problems. The methods of nonlinear parameter estimation and linear parameter estimation are explained based on an improved VP algorithm derived from SVD, and the steps to solve SNLLS problems are then detailed. Thereafter, we compare and evaluate the performances of the method proposed in this paper and the conventional regularized VP algorithm by numerical simulations.

2. Parameter Estimation Method

2.1. Regularization Algorithm for Linear Parameter Estimation

With no loss of generality, the model is abbreviated as
y = Φ θ L
Then, as a result of applying SVD, Φ can be rewritten as
Φ = U S V T = i = 1 l u i σ i v i T
where U = ( u 1 , u 2 , , u m ) and V = ( v 1 , v 2 , , v n ) are composed of the eigenvectors Φ T Φ and Φ Φ T respectively; and U R m × m , V = R n × n , U T U = I and V T V = I . S = diag ( σ 1 , , σ l , σ l + 1 , , σ n ) is a diagonal matrix with diagonal elements that are singular values of Φ , where σ 1 σ 2 σ l > 0 and σ l + 1 = σ l + 2 = = σ n = 0 . l is the rank of Φ .
The Moore–Penrose inverse of Φ is given as Φ + = i = 1 l u i T v i σ i . The least squares solution of the linear parameter, according to Equation (5), is given as
θ ^ L = i = 1 l u i T y v i σ i
Clearly, if some small singular values are significantly close to zero, even a small error will produce a significantly large discrepancy between the least square solution and the true value, making the solution unstable. To avoid this problem, a filter factor was introduced to suppress the error term in the ill-conditioned solution. Thus, the approximate solution can be obtained as
θ ^ L = i = 1 l φ u i T y v i σ i

2.1.1. TSVD Method

The TSVD method removes the disturbing influence of small singular values on the LS solution, thereby allowing a stable solution to be achieved. This is achieved by setting a threshold and then setting any singular value smaller than this threshold to equal zero. Generally, an appropriate threshold can be determined by setting a cutoff ratio coefficient α , such that, if σ i < α σ 1 , then σ i is set to zero. The corresponding TSVD filter factor is
φ α = { 1 σ i α σ 1 0 σ i < α σ 1
At this point, the TSVD regularization solution is given as
θ ^ L = i = 1 l φ α u i T y v i σ i = i = 1 k u i T y v i σ i
where k l is the cutoff point for the singular value, which is typically implemented as the regularization parameter in the TSVD method.

2.1.2. TR Method

The TR method is currently the preferred method for solving ill-conditioned problems. The method entails the use of a regularization parameter μ to constrain the solution and construct a secondary optimization problem as follows:
θ L = arg min θ L { y Φ ( θ N ) θ L 2 + 1 2 μ 2 θ L 2 }
where μ is the regularization parameter. When θ N is fixed, the solution of Equation (13) becomes
θ ^ L = ( Φ ( θ N ) T Φ ( θ N ) + μ 2 I ) 1 Φ ( θ N ) T y
Subsequently, applying SVD to the matrix yields
θ ^ L = i = 1 l φ μ u i T y v i σ i = i = 1 l σ i 2 σ i 2 + μ 2 u i T y v i σ i
Thus, the TR filter factor can be expressed as φ μ = σ i 2 σ i 2 + μ 2 .

2.1.3. Improved TR Method

The TSVD and TR methods are essentially the same. However, there is a difference in the extent to which either method can reduce the influence of small singular values on the solution. Specifically, the TR method changes all singular values, which may cause the approximated solution to be over-smoothed. Alternatively, the TSVD method sets a small portion of the singular values to zero, which not only reduces the variance, but also affects the resolution of the solution, resulting in low accuracy. However, the improved TR method only changes small singular values by enabling the determination of the regularization matrix [27]. Given that the differences between the larger and smaller singular values of the ill-conditioned matrix are large, the singular values are determined to be small singular values with a significant influence when the sum of the standard components of small singular values accounts for more than 95% of the standard deviation. Consequently, they are regularized to reduce the influence on the standard deviation. This condition can be expressed as follows [28]:
i = k l σ 0 σ i 95 % i = 1 l σ 0 σ i
Thus, the regularization parameter k can be obtained. Then, with the filter factor set as φ k , μ = { 1      1 < i k σ i 2 σ i 2 + μ 2      k < i l , when 1 i k , φ k , μ = φ α ; when k i l , φ k , μ = φ μ . The solution of the improved TR method is
θ ^ L = i = 1 l φ k , μ u i T y v i σ i
The improved TR method is a combination of the TSVD and TR methods. Unlike the TSVD method, the improved TR method does not disregard small singular values; furthermore, it yields a more accurate solution. Additionally, unlike the conventional TR method, the improved TR method only modifies the small singular values, keeping the large singular values unchanged, to ensure high accuracy.
A regularization method is employed to determine the appropriate regularization parameters. The L-curve method was adopted in this work. This method entails selecting different μ values to obtain a set of points ( y Φ θ ^ L 2 , θ ^ L ) ; subsequently, an L-shaped curve is constructed with y Φ θ ^ L 2 as the abscissa and θ ^ L as the ordinate. Finally, the corresponding μ regularization parameter value is determined by locating the point of maximum curvature.

2.2. LM Algorithm for Nonlinear Parameter Estimation

After values for the linear parameters have been determined, Equation (6) contains only nonlinear parameters, the values of which can be determined by using the LM algorithm. The iteration strategy is
θ N k + 1 = θ N k + α k d k
where α k is the step length, which ensures that the objective function is in a descending state, and d k is the search direction. α k can be obtained by employing a line search method for imprecise searches. Let m k be the smallest non-negative integer m satisfying r ( θ N k + ρ m d k ) r ( θ N k ) + σ ρ m g k T d k in the k t h iteration process. Then,
α k = ρ m k
where d k can be determined by solving the following equation:
[ J ( θ N k ) T J ( θ N k ) + γ k I ] d k = J ( θ N k ) T F ( θ N k )
where γ k is the damping parameter that controls the magnitude and direction of d k . When γ k is sufficiently large, the direction of d k is consistent with the negative gradient direction of the objective function. When γ k tends toward zero, d k tends toward the G–N direction. In the LM algorithm implemented here, γ is adjusted by employing a strategy similar to adjusting the radius of the trust region [29]. A quadratic function is defined at the current iteration point as follows:
q ( d ) = r ( θ N k ) + ( J ( θ N k ) F ( θ N k ) ) T d + 1 2 d T ( J ( θ N k ) F ( θ N k ) ) d
The ratio of the objective function to the increment of q ( d ) is denoted by η k as follows:
η k = Δ r ( d k ) Δ q ( d k ) = r ( θ N k + 1 ) r ( θ N k ) ( J ( θ N k ) F ( θ N k ) ) T d k + 1 2 d k T ( J ( θ N k ) F ( θ N k ) ) d k
In each step of the LM algorithm, γ k is first assigned an initial value to calculate d k , such as the corresponding previous iteration value. Then, γ k is adjusted according to the value of η k ; d k is subsequently calculated according to the adjusted γ k , and a line search is performed. Clearly, when η k is close to one, γ should be relatively smaller; this can be achieved by using the LM method to solve the nonlinear problem. When η k is close to zero, the modulus length of d k must be reduced, and γ should be relatively larger. When η k is neither close to one nor zero, the value of the parameter γ k is determined to be suitable. The critical values of η are typically 0.25 and 0.75. Accordingly, the update rule of γ k is given as
γ k + 1 : = { 0.1 γ k ,   η k > 0.75 , γ k ,     0.25 η k 0.75 , 10 γ k ,    η k > 0.25 .  
In Equation (20), J ( θ k ) is the Jacobian matrix of the residual function, which, according to the simplified Jacobian matrix calculation method proposed by Kaufman [7], is given as
J K A U = P Φ D Φ Φ y
where D Φ is the Fréchet derivative of the matrix Φ , and Φ is the symmetric generalized inverse of Φ . In [30], P Φ = Φ Φ , where Φ satisfies:
Φ Φ Φ = Φ ,   ( Φ Φ ) T = Φ Φ
Thus, Φ + is not needed to calculate P Φ .
Applying SVD to decompose matrix Φ yields
Φ = U S V T = [ U 1 , U 2 ] [ Σ 0 0 0 ] [ V 1 , V 2 ] T
where Σ is the l t h diagonal matrix, and the diagonal elements are the singular values of Φ . Note that rank ( U 1 ) = rank ( V 1 ) = l , where l is the rank of Φ . Accordingly, we obtain
Φ = [ V 1 , V 2 ] [ Σ 1 0 0 0 ] [ U 1 , U 2 ] T = V 1 Σ 1 U 1 T
Then
P Φ = Φ Φ = [ U 1 , U 2 ] [ I r 0 0 0 ] [ U 1 , U 2 ] T
Thus, the corresponding residual function is
r ( θ N ) = P Φ y 2 = [ U 1 , U 2 ] [ 0 0 0 I m r ] [ U 1 , U 2 ] T y 2 = U 2 U 2 T y 2
and the Jacobian matrix is
J K A U = P Φ D Φ Φ y = U 2 U 2 T D Φ V 1 Σ 1 U 1 T y

2.3. Algorithm Solution Determination

Here, we describe an improved TR optimization method for the SNLLS problem. This method entails the implementation of the VP algorithm to separate the variables, followed by the use of the improved TR method to update the linear parameters, and the LM method to search for nonlinear parameters. We compared the performance of the improved TR regularization method to those of the conventional TR method and TSVD regularization method to verify the effectiveness of the method. The model is given as
( θ ^ L , θ ^ N ) = arg { min θ N y Φ ( θ N ) θ ˜ L 2 , θ ˜ L = i = 1 l φ u i T y v i σ i }
A summary of the steps used for obtaining the solution is given as follows:
Step 1:
Take the initial value of the nonlinear parameter θ N ( 0 ) , the maximum number of iteration steps M and set k = 0 .
Step 2:
The initial nonlinear parameter value θ N ( 0 ) is used to calculate the initial values of the linear parameters θ L ( 0 ) via the TR, TSVD or improved TR method. Then the residual function r ( θ N ) and approximate Jacobian matrix J K A U are obtained.
Step 3:
The iterative step length α k and search direction d k are determined by solving Equations (19) and (20), respectively; thereafter, the nonlinear parameters are updated according to Equation (18).
Step 4:
The linear and nonlinear parameters are cross-updated until k > M ; then, the calculation is terminated.

3. Numerical Simulation

3.1. Predicting the Mackey-Class Time Series Using an RBF Neural Network

Numerical Mackey–Glass time-series simulations were performed for the validation experiment. In the experiment, the LM algorithm was separately implemented in the TSVD-based VP method (VPTSVD), the TR-based VP method (VPTR) and the improved TR method (VPTSVD-TR) to fit a time-series image. This experiment was also performed to reveal the advantages and disadvantages of the aforementioned VP algorithms. The experimental environment was MATLAB R2016a running on a 1.80 GHz PC with Windows 7.
The exponential model is given as
η ( a , α ; x ) = a 1 + j = 2 n a j e λ j x z j 2
where a = ( a 1 , a 2 , a n ) is the linear parameter set, and α = ( λ 2 , λ 3 , , λ n ; z 2 T , z 3 T , , z n T ) T is the nonlinear parameter set, for the functions φ 1 ( α ; x ) 1 , φ j ( α ; x ) = e λ j x z j 2 , and j = 2 , 3 , , n . This model was used to fit the chaotic Mackey–Glass time series generated by the following delay differential equation:
d y ( t ) d t = a y ( t τ ) 1 + y c ( t τ ) b y ( t )
where a = 0.2 , b = 0.1 , c = 10 , τ = 17 . The initial value condition was set as y ( 0 ) = 1.2 , and the time interval was set to 0.1. The Runge–Kutta method was used to solve the differential equation and generate 500 data points, as shown in Figure 1. Among these 500 data points, 300 were extracted from the generated data as follows:
[ y ( t 24 ) , y ( t 18 ) , y ( t 12 ) , y ( t 6 ) , y ( t ) ] ( t = [ 24 323 ] )
Using these 300 data points, VPTSVD+LM, VPTR+LM and VPTSVD-TR+LM were applied to estimate the parameters for the model; the remaining data were used for prediction.
When n = 2, the exponential model yielded 24 nonlinear parameters and 19 linear parameters. Given the same initial iterative value, the fits of the curves derived from the training and prediction data points output by the three algorithms are shown in Figure 2. The red circles correspond to data points extracted from the time-series images. It can be intuitively determined from the figure that the curve fitted by the VPTSVD-TR+LM algorithm is in good agreement with the data generated.
Table 1 presents the results—the number of iterations, the number of calculated functions, the second norm of the vector formed by the residuals of each point and the root mean square error (RMSE t) of the objective function with respect to the original data—for the three iterative methods. Figure 3 shows the convergences of the corresponding nonlinear parameters for the three iterative algorithms.
From the results in Table 1, we can see that the VPTSVD method required the most iterations, and had the highest RMSE-t value. However, the VPTR and the VPTSVD-TR methods were improved significantly and optimized in the iterative process due to fewer iterations, fewer function calculations and smaller RMSEs. Although the numbers of iterations and function evaluations were similar, the results obtained by using the VPTSVD-TR method to calculate the model parameters were considerably more accurate.
From Figure 3, it is evident that all three methods yielded converging results with the same initial values. However, the results of the VPTSVD+LM method began to approach the optimized solution around the 40th iteration, whereas the VPTSVD-TR+LM algorithm began slowly approaching the optimized parameter values at the third iteration, indicating a higher rate of convergence.
Every 10 points were grouped, beginning at the 400th point in the Mackey–Glass time series, to obtain several time-series predictions. The RMSE results for each group are shown in Figure 4.
It can be seen from Figure 4 that the RMSE values for the VPTSVD+LM method increased with the number of predictions, indicating decreasing prediction accuracy and degrading prediction ability. Conversely, in the case of the VPTSVD-TR+LM method, there was a minimal increase in the RMSE, proving the superior stability of the method. These results also indicate that the proposed method features a higher prediction accuracy.

3.2. Height Anomaly Fitting

A height anomaly is the difference in elevation between the surface of the reference ellipsoid and a quasi-geoid. As we can see in Figure 5, the height anomaly ξ can be expressed as ξ = H H n o r m a l . The geodetic height of a point can be calculated by the GPS positioning technique. The normal height can be calculated as long as the height anomaly is determined in the certain area.
A multi-surface function model was used, and its expression had the characteristic that the linear and nonlinear parameters could be separated. The TSVD method, TR method and improved algorithm were applied to solve the parameters of the model according to the height anomaly values of known control points. The model is best expressed as:
ξ i = i = 1 n a i Q i ( X , Y , X i , Y i )
where Q i ( X , Y , X i , Y i ) = ( X X i ) 2 + ( Y Y i ) 2 + b i 2 . The example data, coordinates, geodetic heights, normal heights and height anomaly values of the known points are shown in Table 2.
D01–D15 were used for fitting, and the rest were used to verify the reliability of the parameters and the prediction abilities of the three algorithms. The iteration initial values of the nonlinear parameters were set to 0.5. Model (34) and the algorithm proposed in this paper were applied to fit the height anomalies, and the results are compared with the results calculated by VPTSVD and VPTR.
The parameter estimations were obtained after alternative calculations between the linear parameters and nonlinear parameters. The fitting images corresponding to the three algorithms are shown in Figure 6. Table 3 shows the residual sum of squares, root mean squared error and coefficient of determination values during the processes of fitting and prediction. It can be concluded from Table 3 that all algorithms enabled the model to get high fitting accuracy. During fitting and prediction, The order of magnitude of the root mean square error for VPTSVD, VPTR and VPTSVD-TR is respectively 10−1, 10−2 and 10−3. The accuracy of VPTSVD algorithm was slightly lower. The fitting accuracies of VPTR and VPTSVD-TR were higher. It can be seen from SSRf and RMSEf that the improved algorithm proposed in this paper had better performance in height anomaly prediction.
The normal height of each point was obtained according to H n o r m a l = H ξ and the height anomaly values fitted by three variable projection algorithms. The parameter estimations were verified by the known normal height. Figure 7 shows the fitting effects, and Table 4 lists the normal height fitting residual and RMSE of each point. It can be concluded from Figure 7 and Table 4 that the differences between the normal heights estimated by the three algorithms and the known example data are very small, and the three broken lines are almost coincident, which indicates that the model well fit the normal heights of known points. This is consistent with the conclusions drawn in Figure 6 and Table 3.

4. Conclusions

Separable nonlinear models have been widely applied in many fields and studied by many scholars all over the world. The VP method is one of the most effective algorithms for solving related problems. However, the classical VP method may not be able to estimate the parameters of ill-conditioned problems. As regularization is the most common method for solving ill-conditioned problems, we proposed an improved regularized VP method for the SNLLS problem. This method entails applying certain rules to determine the filter factors necessary to correct small singular values, and using the L-curve method to determine regularization parameters. The results of numerical simulations revealed that the mean square error of the proposed regularized VP method was less than those of the TSVD and TR methods. Furthermore, we found that combining a regularization method with the LM algorithm-based nonlinear parameter estimation method results in a higher rate of convergence. This improved regularized parameter estimation method was demonstrated to have certain advantages in terms of its ability to mitigate the ill-conditioned problem and improve parameter estimation accuracy. It was also found to be more stable when applied to an SNLLS problem.

Author Contributions

Methodology, H.G.; software, L.W.; writing—original draft, H.G.; writing—review and editing, G.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (Grant no. 42074009).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The authors declare that there is no conflict of interest regarding the publication of this paper.

References

  1. Okatani, T.; Deguchi, K. On the Wiberg algorithm for matrix factorization in the presence of missing components. Int. J. Comput. Vis. 2007, 72, 329–337. [Google Scholar] [CrossRef]
  2. Levenberg, K. A method for the solution of certain non-linear problems in least squares. Q. Appl. Math. 1944, 2, 164–168. [Google Scholar] [CrossRef] [Green Version]
  3. Marquardt, D.W. An Algorithm for Least-Squares Estimation of Nonlinear Parameters. SIAM J. Appl. Math. 1963, 11, 431–441. [Google Scholar] [CrossRef]
  4. Hong, J.H.; Zach, C.; Fitzgibbon, A. Revisiting the variable projection method for separable nonlinear least squares problems. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 5939–5947. [Google Scholar]
  5. Golub, G.H.; Pereyra, V. The Differentiation of Pseudo-Inverses and Nonlinear Least Squares Problems Whose Variables Separate. SIAM J. Numer. Anal. 1973, 10, 413–432. [Google Scholar] [CrossRef]
  6. Golub, G.H.; Pereyra, V. Separable nonlinear least squares: The variable projection method and its applications. Speech Commun. 2003, 45, 63–87. [Google Scholar] [CrossRef]
  7. Kaufman, L. A variable projection method for solving separable nonlinear least squares problems. BIT Numer. Math. 1975, 15, 49–57. [Google Scholar] [CrossRef]
  8. Ruhe, A. Algorithms for separable nonlinear least squares problems. SIAM Rev. 1980, 22, 318–337. [Google Scholar] [CrossRef]
  9. O’Leary, D.P.; Rust, B.W. Variable projection for nonlinear least squares problems. Comput. Optim. Appl. 2013, 54, 579–593. [Google Scholar] [CrossRef]
  10. Ruano, A.E.B.; Jones, D.I.; Fleming, P.J. A new formulation of the learning problem of a neural network controller. In Proceedings of the 30th IEEE Conference on Decision and Control, Brighton, UK, 11–13 December 1991; pp. 865–866. [Google Scholar]
  11. Gan, M.; Chen, C.L.P.; Chen, G.; Chen, L. On some separated algorithms for separable nonlinear least squares problems. IEEE Trans. Cybern. 2018, 48, 2866–2874. [Google Scholar] [CrossRef]
  12. Gan, M.; Li, H. An Efficient Variable Projection Formulation for Separable Nonlinear Least Squares Problems. IEEE Trans. Cybern. 2014, 44, 707–711. [Google Scholar] [CrossRef]
  13. Chen, G.Y.; Gan, M.; Ding, F. Modified Gram-Schmidt Method Based Variable Projection Algorithm for Separable Nonlinear models. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 2410–2418. [Google Scholar] [CrossRef]
  14. Böckmann, C. A modification of the trust-region Gauss-Newton method to solve separable nonlinear least squares problems. J. Math. Syst. Estim. Control 1995, 5, 1–16. [Google Scholar]
  15. Chung, J.J.; Nagy, G. An efficient iterative approach for large-scale separable nonlinear inverse problems. SIAM J. Sci. Comput. 2010, 31, 4654–4674. [Google Scholar] [CrossRef]
  16. Li, X.L.; Liu, K.; Dong, Y.S.; Tao, D.C. Patch alignment manifold matting. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 3214–3226. [Google Scholar] [CrossRef] [Green Version]
  17. Tikhonov, A.N.; Arsenin, V.Y. Solutions of ill-posed problems. SIAM Rev. 1979, 21, 266–267. [Google Scholar]
  18. Park, Y.; Reichel, L.; Rodriguez, G. Parameter determination for Tikhonov regularization problems in general form. J. Comput. Appl. Math. 2018, 343, 12–25. [Google Scholar] [CrossRef]
  19. Hansen, P.C. The Truncated SVD as A Method for Regularization. BIT Numer. Math. 1987, 27, 534–553. [Google Scholar] [CrossRef]
  20. Xu, P.L. Truncated SVD methods for Discrete Linear Ill-posed Problems. Geophys. J. Int. 1998, 1335, 505–514. [Google Scholar] [CrossRef]
  21. Aravkin, A.Y.; Drusvyatskiy, D.; van Leeuwen, T. Efficient quadratic penalization through the partial minimization technique. IEEE Trans. Autom. Control 2018, 63, 2131–2138. [Google Scholar] [CrossRef] [Green Version]
  22. Pang, S.C.; Li, T.; Dai, F.; Yu, M. Particle Swarm Optimization Algorithm for Multi-Salesman Problem with Time and Capacity Constraints. Appl. Math. Inf. Sci. 2013, 7, 2439–2444. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, X.Z.; Liu, D.Y.; Zhang, Q.Y.; Huang, H.L. The iteration by correcting characteristic value and its application in surveying data processing. J. Heilongjiang Inst. Technol. 2001, 15, 3–6. [Google Scholar]
  24. Zeng, X.Y.; Peng, H.; Zhou, F.; Xi, Y.H. Implementation of regularization for separable nonlinear least squares problems. Appl. Soft Comput. 2017, 60, 397–406. [Google Scholar] [CrossRef]
  25. Chen, G.Y.; Gan, M.; Chen CL, P.; Li, H.X. A Regularized Variable Projection Algorithm for Separable Nonlinear Least–Squares Problems. IEEE Trans. Autom. Control 2019, 64, 526–537. [Google Scholar] [CrossRef]
  26. Wang, K.; Liu, G.L.; Tao, Q.X.; Zhai, M. Efficient Parameters Estimation Method for the Separable Nonlinear Least Squares Problem. Complexity 2020, 2020, 9619427. [Google Scholar] [CrossRef] [Green Version]
  27. Fuhry, M.; Reichel, L. A new Tikhonov regularization method. Number Algorithms 2012, 59, 433–445. [Google Scholar] [CrossRef]
  28. Lin, D.f.; Zhu, J.j.; Song, Y.C. Regularized singular value decomposition parameter construction method. J. Surv. Mapp. 2016, 45, 883–889. [Google Scholar]
  29. Byrd, R.H.; Schnabel, R.B.; Shultz, G.A. A Trust Region Algorithm for Nonlinearly Constrained Optimization. SIAM J. Numer. Anal. 1987, 24, 1152–1170. [Google Scholar] [CrossRef] [Green Version]
  30. Golub, G.H.; Styan, G.P.H. Numerical computations for univariate linear models. J. Stat. Comput. Simul. 1973, 2, 253–274. [Google Scholar] [CrossRef]
Figure 1. Chaotic Mackey–Glass time series.
Figure 1. Chaotic Mackey–Glass time series.
Axioms 10 00196 g001
Figure 2. Curve-fitting evaluation results.
Figure 2. Curve-fitting evaluation results.
Axioms 10 00196 g002
Figure 3. The iterative process for nonlinear parameters.
Figure 3. The iterative process for nonlinear parameters.
Axioms 10 00196 g003
Figure 4. RMSE results following multi-step time-series forecasting.
Figure 4. RMSE results following multi-step time-series forecasting.
Axioms 10 00196 g004
Figure 5. A geodetic height, normal height and height anomaly.
Figure 5. A geodetic height, normal height and height anomaly.
Axioms 10 00196 g005
Figure 6. Fitting effects with respect to height anomaly values.
Figure 6. Fitting effects with respect to height anomaly values.
Axioms 10 00196 g006
Figure 7. Normal height fitting effects.
Figure 7. Normal height fitting effects.
Axioms 10 00196 g007
Table 1. Simulated results for a Mackey–Glass time series.
Table 1. Simulated results for a Mackey–Glass time series.
Number of IterationsFunction Evaluation NumberSecond Norm of Residual VectorRMSE-t
VPTSVD7519530.343730.27918
VPTR51350.364270.19919
VPTSVD-TR71850.058180.09653
Table 2. Coordinates, geodetic heights ( H ), normal heights ( H n o r m a l ) and height anomaly values ( ξ ).
Table 2. Coordinates, geodetic heights ( H ), normal heights ( H n o r m a l ) and height anomaly values ( ξ ).
No.LatitudeLongitude H ( m ) H n o r m a l ( m ) ξ ( m )
D01343000.00001120000.00001001.52201056.5490−55.0270
D02343000.00001120230.00001009.12901063.9280−54.7990
D03343000.00001120500.00001012.39501067.2510−54.8560
D04343000.00001120730.00001070.18601125.3030−55.1170
D05343000.00001121000.00001025.23301080.2320−54.9990
D06343000.00001121230.00001019.43101074.4540−55.0230
D07343000.00001121500.00001026.70901081.7560−55.0470
D08343000.00001121730.00001067.99401123.1420−55.1480
D09343000.00001122000.00001157.72201212.5890−54.8670
D10343000.00001122230.00001017.53201072.6350−55.1030
D11343000.00001122500.00001004.06301058.9510−54.8880
D12343000.00001122730.00001004.35001059.1140−54.7640
D13342730.00001120000.0000988.81301043.3570−54.5440
D14342730.00001120230.0000975.30101029.7370−54.4360
D15342730.00001120500.0000983.51301038.1040−54.5910
D16342730.00001120730.0000988.85501043.5930−54.7380
D17342730.00001121000.00001108.94301163.7680−54.8250
D18342730.00001121230.0000980.59401035.2260−54.6320
D19342730.00001121500.0000964.18701018.5260−54.3390
D20342730.00001120000.00001001.52201056.5490−55.0270
Table 3. Residual sum of squares (SSR, SSRf), root mean squared error (RMSE, RMSEf) and coefficient of determination ( R 2 ) values.
Table 3. Residual sum of squares (SSR, SSRf), root mean squared error (RMSE, RMSEf) and coefficient of determination ( R 2 ) values.
AlgorithmsSSR (m2)SSRf (m2)RMSE (m)RMSEf (m) R 2
VPTSVD0.50040.93370.18260.43210.2599
VPTR0.08500.37700.07530.27460.8744
VPTSVD-TR0.07980.35450.07290.26630.8820
Table 4. Fitting residuals (r) and RMSEs of the normal heights.
Table 4. Fitting residuals (r) and RMSEs of the normal heights.
No.Hnormal (m)rTSVD (m)rTR (m)rTSVD+TR (m)
D011056.54900.28710.10410.0902
D021063.9280−0.0457−0.0457−0.0161
D031067.25100.1498−0.0245−0.0742
D041125.30300.17020.00150.0224
D051080.2320−0.2271−0.02660.0108
D061074.4540−0.1912−0.0084−0.0399
D071081.7560−0.0025−0.0036−0.0052
D081123.14200.23120.09420.1011
D091212.58900.0406−0.1688−0.1694
D101072.63500.28500.10920.1078
D111058.9510−0.0049−0.0108−0.0069
D121059.1140−0.2656−0.0069−0.0091
D131043.3570−0.1411−0.0576−0.0701
D141029.7370−0.2104−0.0699−0.0269
D151038.1040−0.07260.11390.0854
D161043.5930−0.12040.18820.0878
D171163.7680−0.26110.20260.1069
D181035.2260−0.4721−0.0523−0.1058
D191018.5260−0.6528−0.4140−0.4319
D201056.5490−0.4494−0.3555−0.3710
RMSE00.26780.15200.1474
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Guo, H.; Liu, G.; Wang, L. An Improved Tikhonov-Regularized Variable Projection Algorithm for Separable Nonlinear Least Squares. Axioms 2021, 10, 196. https://doi.org/10.3390/axioms10030196

AMA Style

Guo H, Liu G, Wang L. An Improved Tikhonov-Regularized Variable Projection Algorithm for Separable Nonlinear Least Squares. Axioms. 2021; 10(3):196. https://doi.org/10.3390/axioms10030196

Chicago/Turabian Style

Guo, Hua, Guolin Liu, and Luyao Wang. 2021. "An Improved Tikhonov-Regularized Variable Projection Algorithm for Separable Nonlinear Least Squares" Axioms 10, no. 3: 196. https://doi.org/10.3390/axioms10030196

APA Style

Guo, H., Liu, G., & Wang, L. (2021). An Improved Tikhonov-Regularized Variable Projection Algorithm for Separable Nonlinear Least Squares. Axioms, 10(3), 196. https://doi.org/10.3390/axioms10030196

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop