Next Article in Journal
Fractional Integral of the Confluent Hypergeometric Function Related to Fuzzy Differential Subordination Theory
Previous Article in Journal
Special Issue: Fractal Functions and Applications
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Active-Set Fischer–Burmeister Trust-Region Algorithm to Solve a Nonlinear Bilevel Optimization Problem

Department of Mathematics, Faculty of Science, Alexandria University, Alexandria 5424041, Egypt
*
Author to whom correspondence should be addressed.
Fractal Fract. 2022, 6(8), 412; https://doi.org/10.3390/fractalfract6080412
Submission received: 18 May 2022 / Revised: 1 July 2022 / Accepted: 18 July 2022 / Published: 27 July 2022
(This article belongs to the Section Numerical and Computational Methods)

Abstract

:
In this paper, the Fischer–Burmeister active-set trust-region (FBACTR) algorithm is introduced to solve the nonlinear bilevel programming problems. In FBACTR algorithm, a Karush–Kuhn–Tucker (KKT) condition is used with the Fischer–Burmeister function to transform a nonlinear bilevel programming (NBLP) problem into an equivalent smooth single objective nonlinear programming problem. To ensure global convergence for the FBACTR algorithm, an active-set strategy is used with a trust-region globalization strategy. The theory of global convergence for the FBACTR algorithm is presented. To clarify the effectiveness of the proposed FBACTR algorithm, applications of mathematical programs with equilibrium constraints are tested.

1. Introduction

The mathematical formulation for NBLP problem which we will consider it is
min v f u ( v , w ) s . t . g u ( v , w ) 0 , min w f l ( v , w ) , s . t . g l ( v , w ) 0 ,
where v n 1 and w n 2 . In our approach, the  functions f u : n 1 + n 2 , f l : n 1 + n 2 , g u : n 1 + n 2 m 1 , and  g l : n 1 + n 2 m 2 must have a twice continuously differentiable function at least.
The NBLP problem (1) is utilized so extensively in transaction network, resource allocation, finance budget, price control, etc., see [1,2,3,4]. The  NBLP problem (1) has two levels of optimization problems, upper and lower levels. A decision maker with the upper level objective function f u ( v , w ) takes the lead, and so he chooses the decision vector v. According to this, the decision maker with lower level objective function f l ( v , w ) , chooses the decision vector w to optimize her objective, parameterized in v.
To obtain the solution of problem (1), number of different approaches have been offered, see (1), see [5,6,7,8,9]. In our method, we utilize one of these approaches to transforme NBLP problem (1) to a single level one by replacing the lower level optimization problem with its KKT conditions, see [10,11].
Utilizing KKT optimality conditions for the lower level problem, the NBLP problem (1) is reduced to the following single-objective optimization problem:
min v , w f u ( v , w ) s . t . g u ( v , w ) 0 , w f l ( v , w ) + w g l ( v , w ) λ = 0 , g l ( v , w ) 0 , λ j g l j ( v , w ) = 0 , j = 1 , , m 2 , λ j 0 , j = 1 , , m 2 ,
where λ m 2 a multiplier vector which is associated with inequality constraint g l ( v , w ) .
Problem (2) is non-differentiable and non-convex. Furthermore, the regularity assumption prerequisites to successfully handle smooth optimization problems are never satisfied. Following the smoothing method which is proposed by [2], we can introduce the FBACTR algorithm to solve the problem (2). Before introducing FBACTR algorithm, we need the following definition.
Definition 1.
A Fischer–Burmeister function is the function Ψ ( e , d ) : 2 and it is defined by Ψ ( e , d ) = e + d e 2 + d 2 . A perturbed Fischer–Burmeister function is the function ψ ( e , d , ε ^ ) : 3 and it is defined by ψ ( e , d , ε ^ ) = e + d e 2 + d 2 + ε ^ .
The Fischer–Burmeister function has the property that Ψ ( e , d ) = 0 if and only if e 0 , d 0 , and  e d = 0 . It is non-differentiable at e = d = 0 . Its perturbed variant satisfies ψ ( e , d , ε ^ ) = 0 if and only if e > 0 , d > 0 , and  e d = ε ^ 2 for ε ^ > 0 . This function is smooth with respect to e, d, for  ε ^ > 0 , and for more details see [12,13,14,15].
The next perturbed Fischer–Burmeister function is used to satisfy the asymptotic stability conditions, and allow the FBACTR algorithm to solve problem (2).
ψ ( e , d , ε ^ ) = e 2 + d 2 + ε ^ e d .
Using the perturbed Fischer–Burmeister function (3), problem (2) can be approximated by:
min v , w f u ( v , w ) s . t . g u ( v , w ) 0 , w f l ( v , w ) + w g l ( v , w ) λ = 0 , g l j 2 + λ j 2 + ε ^ λ j + g l j = 0 , j = 1 , , m 2 .
The following notations are introduced to simplify our discussion. These notations are n = n 1 + n 2 + m 2 , x = ( v , w , λ ) T n and c ( x ) = ( w f l ( v , w ) + w g l ( v , w ) λ , g l j 2 + λ j 2 + ε ^ λ j + g l j ) T , j = 1 , , m 2 . Hence problem (4) can be reduced as follows:
minimize f u ( x ) subject to g u ( x ) 0 , c ( x ) = 0 ,
where f u : n , g u : n m 1 , and  c : n n 2 + m 2 .
A set of indices of binding or violated inequality constraints at x is defined by I ( x ) = { i : g u i ( x ) 0 } . A regular point is the point x * at which the vectors of the set { c i ( x * ) , i = 1 , 2 , , n 2 + m 2 } { g u i ( x * ) , i I ( x * ) } are linearly independent.
A regular point x * is KKT point of problem (5) if there exist Lagrange multiplier vectors μ * n 2 + m 2 and λ * m 1 such that the following KKT conditions hold:
f u ( x * ) + c ( x * ) μ * + g u ( x * ) λ * = 0 ,
c ( x * ) = 0 ,
g u ( x * ) 0 ,
( λ * ) i g u i ( x * ) = 0 , i = 1 , , m 1 ,
( λ * ) i 0 , i = 1 , , m 1 .
To solve the nonlinear single-objective constrained optimization problem (5), various approaches have been proposed; for more details, see [16,17,18,19,20,21,22].
An active-set strategy is utilized to reduce problem (5) to equality constrained optimization problem. The  idea beyond the active-set method is to identify at every iteration, the active inequality constraints and treat them as equalities and this allows to utilize the improved methods which are used to solve the equality constrained problems, see [21,23,24]. Most of the methods that are used to solve the equality constrained problems, may not converge if the starting point is far away from the stationary point, so it is called a local method.
To ensure a convergence to the solution from any starting point, a trust-region strategy which is strongly global convergence can be induced. It is very important strategy to solve a smooth optimization. It is more robust when it deals with rounding errors. It does not require the objective function of the model be convex. For more details see [11,21,22,23,24,25,26,27,28,29,30,31,32].
To treat the difficult of having infeasible trust-region subproblem in FBACTR algorithm, a reduced Hessian technique which is suggested by [33,34] and used by [22,24,35] is utilized.
Under five assumptions, a theory of global convergence for FBACTR algorithm is proved. Moreover, numerical experiments display that FBACTR algorithm performers effectively and efficiently in pursuance.
We shall use the following notation and terminology. We use . to denote the Euclidean norm . 2 . Subscript k refers to iteration indices. For example, f u k f u ( x k ) , g u k g u ( x k ) , c k c ( x k ) , Y k Y ( x k ) , P k P ( x k ) , x k x ( x k , μ k ) , and so on to denote the function value at a particular point.
The rest of the paper is organized as follows. Section 2 is devoted to the description of an active-set trust-region algorithm to solve problem (5) and summarized to FBACTR algorithm to solve NBLP problem (1) is introduced. In Section 3 the analysis of the theory of global convergence of the FBACTR algorithm is presented. Section 4 contains an implementation of the FBACTR algorithm and the results of test problems. Finally, some further remarks are given in Section 5.

2. Active-Set with Trust-Region Technique

A detailed description for active-set with the trust-region strategy to solve problem (5) and summarized to FBACTR algorithm to solve problem (1) are introduced in this section.
Based on the active-set method which is suggested by [36] and used with [21,22,23,24], we define a 0–1 diagonal matrix P ( x ) m 1 × m 1 , whose diagonal entries are:
p i ( x ) = 1 i f g u i ( x ) 0 , 0 i f g u i ( x ) < 0 .
Using the previous definition of the matrix P ( x ) , a smooth and simple function is utilized to replace problem (5) with the following simple problem
minimize f u ( x ) + r 2 P ( x ) g u ( x ) 2 subject to c ( x ) = 0 ,
where r > 0 is a parameter, see [21,22,23]. The Lagrangian function associated with problem (12) is given by:
L ( x , μ ; r ) = ( x , μ ) + r 2 P ( x ) g u ( x ) 2 ,
where
( x , μ ) = f u ( x ) + μ T c ( x ) ,
and μ n 2 + m 2 represents a Lagrange multiplier vector which is associated with the constraint c ( x ) . A KKT point ( x * , μ * ) for problem (12) is the point at which the following conditions are satisfied
( x * , μ * ) + r g u ( x * ) P ( x * ) g u ( x * ) = 0 ,
h ( x * ) = 0 ,
where ( x * , μ * ) = f u ( x * ) + c ( x * ) μ * .
If the KKT point ( x * , μ * ) satisfies conditions (6)–(10), we notice that it is also satisfies conditions (15) and (16), but the converse is not necessarily true. So, we design FBACTR algorithm in a way that, if  ( x * , μ * ) satisfies conditions (15) and (16), then it is also satisfies KKT conditions (6)–(10).
Various approaches which were proposed to solve the equality constrained are local methods. By local method, we mean a method such that if the starting point is sufficiently close to a solution, then under some reasonable assumptions the method is guaranteed by theory to converge to the solution. There is no guarantee that the local method converges starting from the remote. Globalizing a local method means modifying the method in such a way that is guaranteed to converge from any starting point without sacrificing its fast local rate of convergence. To ensure convergence from the remote, the trust-region technique is utilized.

2.1. A Trust-Region Technique

To solve problem (12) and to convergence from remote with any starting point, the trust-region strategy is used. A naive trust-region quadratic subproblem associated with problem (12) is:
minimize q k ( s ) = k + x k T s + 1 2 s T H k s + r 2 P k ( g u k + g u k ) T s ) 2 subject to c k + c k T s = 0 , s δ k ,
where 0 < δ k represents the trust-region radius and H k is the Hessian matrix of the Lagrangian function (14) or an approximation to it.
Subproblem (17) may be infeasible because there may be no intersecting points between hyperplane of the linearized constraints c ( x ) + c ( x ) T s = 0 and the constraint s δ k . Even if they intersect, there is no guarantee that this will keep true if δ k is reduced, see [37]. To overcome this problem, a reduced Hessian technique which was suggested by [33,34] and used by [22,23,35] is used. In this technique, to obtain the trial step s k , it is decomposed into two orthogonal components: the tangential component s k t to improve optimality and the normal component s k n to improve feasibility. To evaluate each of s k n and s k t , two unconstrained trust-region subproblems are solved.
  • To obtain the normal component s n
To evaluate the normal component s k n , the following trust-region subproblem must be solved:
minimize 1 2 c k + c k T s n 2 subject to s n ζ δ k ,
for some ζ ( 0 , 1 ) .
Any method can be used to solve subproblem (18), as long as a fraction of the normal predicted decrease obtained by the Cauchy step s k n c p is less than or equal to the normal predicted decrease obtained by s k n . That is, the following condition must be held:
c k 2 c k + c k T s k n 2 ϑ 1 { c k 2 c k + c k T s k n c p 2 } ,
for some ϑ 1 ( 0 , 1 ] . The normal Cauchy step s k n c p is given by:
s k n c p = τ k n c p c k c k ,
where the parameter τ k n c p is given by:
τ k n c p = c k c k 2 ( c k ) T c k c k 2 i f c k c k 3 c k T c k c k ) 2 δ k and c k T c k c k ) > 0 , δ k c k c k otherwise .
A dogleg method is used to solve subproblem (18). It is very cheap if the Hessian is indefinite. The dogleg algorithm approximates the solution curve to subproblem (18) by piecewise linear function connecting the Newton point to the Cauchy point. For more details, see [35].
Once s k n is estimated, we will compute s k t = Y k s ¯ k t . A matrix Y k is the matrix whose columns form a basis for the null space of ( c k ) T .
  • To obtain the tangential component s k t .
To evaluate the tangential component s k t , the following subproblem is solved by using the dogleg method
minimize ( Y k T q k ( s k n ) ) T s ¯ t + 1 2 s ¯ t T Y k T B k Y k s ¯ t subject to Y k s ¯ t Δ k ,
where q k ( s k n ) = x k + B k s k n + r k g u k P k g u k , B k = H k + r k g u k P k g u k T , and  Δ k = δ k 2 s k n 2 .
Since the dogleg method is used to solve the above subproblem, then a fraction of the tangential predicted decrease obtained by a tangential Cauchy step s ¯ k t c p is less than or equal to the tangential predicted decrease which is obtained by tangential step s ¯ k t . That is, the following conditions hold
q k ( s k n ) q k ( s k n + Y k s ¯ k t ) ϑ 2 [ q k ( s k n ) q k ( s k n + Y k s ¯ k t c p ) ] ,
for some ϑ 2 ( 0 , 1 ] . The tangential Cauchy step s k t c p is defined as follows
s ¯ k t c p = τ k t c p Y k T q k ( s k n ) ,
where the parameter τ k t c p is given by
τ k t c p = Y k T q k ( s k n ) 2 ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) i f Y k T q k ( s k n ) 3 ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Δ k and ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) > 0 , Δ k Y k T q k ( s k n ) otherwise ,
such that B ¯ k = Y k T B k Y k .
To be decided whether the step s k = s k n + s k t will be accepted or not, a merit function is needed to tie the objective function with the constraints in such a way that progress in the merit function means progress in solving the problem. The following augmented Lagrange function is used in FBACTR algorithm as a merit function,
Φ ( x , μ ; r ; σ ) = ( x , μ ) + r 2 P ( x ) g u ( x ) 2 + σ c ( x ) 2 ,
where σ > 0 is a penalty parameter.
To test whether the point ( x k + s k , μ k + 1 ) will be taken in the next iterate, an actual reduction and a predicted reduction are defined.
The actual reduction A r e d k in the merit function in moving from ( x k , μ k ) to ( x k + s k , μ k + 1 ) is defined as follows:
A r e d k = Φ ( x k , μ k ; r k ; σ k ) Φ ( x k + d k , μ k + 1 ; r k ; σ k ) .
A r e d k can also be written as follows:
A r e d k = ( x k , μ k ) ( x k + 1 , μ k ) Δ μ k T c k + 1 + r k 2 [ P k g u ( x k ) 2 P k + 1 g u k + 1 2 ] + σ k [ c k 2 c k + 1 2 ] ,
where Δ μ k = ( μ k + 1 μ k ) .
The predicted reduction in the merit function is defined to be:
P r e d k = ( x ( x k , μ k ) ) T s k 1 2 s k T H k s k Δ μ k T ( c k + c k T s k ) + r k 2 [ P k g u k 2 P k ( g u k + g u k T s k ) 2 ] + σ k [ c k 2 c k + c k T s k 2 ] .
P r e d k can be written as:
P r e d k = q k ( 0 ) q k ( s k ) Δ μ k T ( c k + c k T s k ) + σ k [ c k 2 c k + c k T s k 2 ] .
  • To update the penalty parameter σ k
To update the penalty parameter σ k to ensure that P r e d k 0 , the following schemeis used (see Algorithm 1):
Algorithm 1 To update the penalty parameter σ k
If
P r e d k σ k 2 [ c k 2 c k + c k T μ k s k 2 ] ,
then, set
σ k = 2 [ q k ( s k ) q k ( 0 ) + Δ μ k T ( c k + c k T s k ) ] c k 2 c k + c k T s k 2 + β 0 ,
where β 0 > 0 is a small fixed constant. Else, set
σ k + 1 = max ( σ k , r k 2 ) .
End if.
For more details, see [22].
  • To test the step s k and update δ k
The framework to test the step s k and update δ k is clarified in the Algorithm 2.
Algorithm 2 (To test the step s k and update δ k )
Choose 0 < α 1 < α 2 < 1 , 0 < τ 1 < 1 < τ 2 , and  δ m i n δ 0 δ m a x .
While A r e d k P r e d k ( 0 , α 1 ) or P r e d k 0 .
Set δ k = τ 1 s k .
Evaluate a new trial step s k .
End while.
If A r e d k P r e d k [ α 1 , α 2 ) .
Set x k + 1 = x k + s k and δ k + 1 = max ( δ k , δ m i n ) .
End if. If A r e d k P r e d k [ α 2 , 1 ] .
Set x k + 1 = x k + s k and δ k + 1 = min { δ m a x , max { δ m i n , τ 2 δ k } } .
End if.
  • To update the positive parameter r k
To update the positive parameter r k , we use the following scheme (see Algorithm 3)
Algorithm 3 To update the positive parameter r k
If
1 2 [ q k ( s k n ) q k ( s k ) ] g u ( x k ) P ( x k ) g u ( x k ) min { g u ( x k ) P ( x k ) g u ( x k ) , δ k } ,
Set r k + 1 = r k .
Else, set r k + 1 = 2 r k .
End if.
For more details see, [25].
Finally, the algorithm stopped if the termination criteria Y k T x k + g u k P k g u k + c k ε 1 or s k ε 2 , for some ε 1 , ε 2 > 0 is satisfied.
  • A trust-region algorithm
The framework of the trust-region algorithm to solve subproblem (17) are summarized as follows (see Algorithm 4).
Algorithm 4 Trust-region algorithm
Step 0. (Initialization)
  Starting with x 0 . Evaluate μ 0 and P 0 . Set r 0 = 1 , σ 0 = 1 , and  β 0 = 0.1 .
  Choose ε 1 , ε 2 , τ 1 , τ 2 , α 1 , and  α 2 such that 0 < ε 1 , 0 < ε 2 , 0 < τ 1 < 1 < τ 2 ,
  and  0 < α 1 < α 2 < 1 .
  Choose δ m i n , δ m a x , and  δ 0 such that δ m i n δ 0 δ m a x . Set k = 0 .
Step 1. If Y k T x k + g u k P k g u k + c k ε 1 , then stop.
Step 2. (How to compute  s k )
  (a) Evaluate the normal component s k n by solving subproblem (18).
  (b) Evaluate the tangential component s ¯ k t by solving subproblem (22).
  (c) Set s k = s k n + Y k s ¯ k t .
Step 3. If s k ε 2 , then stop.
Step 4. Set x k + 1 = x k + s k .
Step 5. Compute P k + 1 given by (11).
Step 6. Evaluate μ k + 1 by solving the following subproblem
m i n i m i z e f u k + 1 + c k + 1 μ + r k g u k + 1 P k + 1 g u k + 1 2 .
Step 7. To update the penalty parameter σ k , using Algorithm 1.
Step 8. To test the step s k and update the radius δ k , using Algorithm 2.
Step 9. To update the positive parameter  r k , using Algorithm 3.
Step 10. Set k = k + 1 and go to Step 1.
The main steps for solving the NBLP problem (1) are clarified in the following algorithm.

2.2. Fischer–Burmeister Active-Set Trust-Region Algorithm

The framework to solve NBLP problem (1) is summarized in the Algorithm 5.
Algorithm 5 FBACTR algorithm
Step 1. Use KKT optimality conditions for the lower level of problem (1) and convert it to a single objective constrained optimization problem (2).
Step 2. Using Fischer–Burmeister function (3) with ϵ = 0.001 to obtain the smooth problem (4).
Step 3. Summarize problem (4) to the form of nonlinear optimization problem (5).
Step 4. Use the active set strategy to reduce problem (5) to problem (12).
Step 5. Use trust-region Algorithm 4 to solve problem (12) and obtained approximate solution for problem (5) which is approximate solution for problem (1).
The next section is dedicated to the global convergence analysis for the active-set with the trust-region algorithm.

3. Global Convergence Analysis

Let { ( x k , μ k ) } be the sequence of points generated by FBACTR Algorithm 5. Let Ω n be a convex set which is contained all iterates x k n and x k + s k n .
Standard assumptions which are needed on the set Ω to demonstrate global convergence theory for FBACTR Algorithm 5 are stated in the following section.

3.1. A Standard Assumptions

The next standard assumptions are required to demonstrate the global convergence theory for the FBACTR Algorithm 5.
  • [ SA 1 .] Functions f u : n , g u : n 1 m , f l : n n 2 , and g l : n m 2 are twice continuously differentiable functions for all x Ω .
  • [ SA 2 .] The sequence of the Lagrange multiplier vectors { μ k } is bounded.
  • [ SA 3 .] All of c ( x ) , c ( x ) , 2 c i ( x ) for i = 1 , 2 , , n 2 + m 2 , g u ( x ) , g u ( x ) , 2 g u i ( x ) for i = 1 , 2 , , m 1 , and ( c ( x ) T c ( x ) ) 1 are uniformly bounded on Ω .
  • [ SA 4 .] The matrix c ( x ) has full column rank.
  • [ SA 5 .] The sequence of Hessian matrices { H k } is bounded.
Some fundamental lemmas which are needed in the proof of the main theorem introduced in the following section.

3.2. Main Lemmas

Some basic lemmas which are required to demonstrate the main theorems are presented in this section.
Lemma 1.
Under standard assumption S A 1 S A 5 and at any iteration k, there exists a positive constant K 1 such that:
s k n K 1 c k .
Proof. 
Since the normal component s k n is normal to the tangent space, then we have:
s k n = c k ( c k T c k ) 1 c k T s k = c k ( c k T c k ) 1 [ c k + c k T s k c k ] c k ( c k T c k ) 1 [ c k + c k T s k + c k ] c k ( c k T c k ) 1 c k ,
where c k + c k T s k c k . Using standard assumptions S A 1 S A 5 , we have the desired result. □
Lemma 2.
Under standard assumptions S A 1 and S A 3 , the functions P ( x ) g u ( x ) are Lipschitz continuous in Ω.
Proof. 
See Lemma (4.1) in [36]. □
From Lemma 2, we conclude that g u ( x ) T P ( x ) g u ( x ) is differentiable and g u ( x ) P ( x ) g u ( x ) is Lipschitz continuous in Ω .
Lemma 3.
At any iteration k, let A ( x k ) ( m 1 ) × ( m 1 ) be a diagonal matrix whose diagonal entries are:
( a k ) i = 1 i f ( g u k ) i < 0   a n d   ( g u k + 1 ) i 0 , 1 i f ( g u k ) i 0   a n d   ( g u k + 1 ) i < 0 , 0 o t h e r w i s e ,
where i = 1 , 2 , , m 1 . Then
P k + 1 = P k + A k .
Proof. 
See Lemma (6.2) in [21]. □
Lemma 4.
Under standard assumptions S A 1 and S A 3 , there exists a positive constant K 2 such that
A k g u k K 2 s k .
Proof. 
See Lemma (6.3) in [21]. □
Lemma 5.
Under standard assumptions S A 1 S A 5 , there exists a positive constant K 3 such that:
A r e d k P r e d k K 3 σ k s k 2 .
Proof. 
From (37) and (27) we have:
A r e d k = ( x k , μ k ) ( x k + 1 , μ k ) Δ μ k T c k + 1 + r k 2 [ g u k T P k g u k g u k + 1 T ( P k + A k ) g u k + 1 ] + σ k [ c k 2 c k + 1 2 ] .
From (40), (28), and using Cauchy–Schwarz inequality, we have:
A r e d k P r e d k ( x k , μ k ) + x ( x k , μ k ) T s k ( x k + 1 , μ k ) + Δ μ k T [ c k + c k T s k c k + 1 ] + r k 2 P k ( g u k + g u k T s k ) 2 g u k + 1 T ( P k + A k ) g u k + 1 + σ k c k + c k T s k 2 c k + 1 2 .
Hence,
| A r e d k P r e d k | 1 2 s k T ( H k 2 ( x k + ξ 1 s k , μ k ) ) s k + 1 2 s k T [ 2 c ( x k + ξ 2 s k ) Δ μ k ] s k + r k 2 s k T [ g u k P k g u k T g u ( x k + ξ 4 s k ) P k g u ( x k + ξ 4 s k ) T ] s k + r k 2 s k T 2 g u ( x k + ξ 4 s k ) P k g u ( x k + ξ 4 s k ) s k + r k 2 A k [ g u k + g u ( x k + ξ 5 s k ) T s k ] 2 + σ k s k T [ c k c k T c ( x k + ξ 6 s k ) c ( x k + ξ 6 s k ) T ] s k + σ k s k T 2 c ( x k + ξ 6 s k ) c ( x k + ξ 6 s k ) s k ,
for some ξ 1 , ξ 2 , ξ 3 , ξ 4 , ξ 5 , and ξ 6 ( 0 , 1 ) . Using standard assumptions S A 1 S A 5 , σ k r k , σ k 1 , and inequality (38), we have:
A r e d k P r e d k κ 1 s k 2 + κ 2 σ k s k 2 c k + κ 3 σ k s k 3 ,
where κ 1 > 0 , κ 2 > 0 , and κ 3 > 0 are constants and independent of the iteration k. From inequality (41), σ k 1 , s k , and c k are uniformly bounded, we obtain the desired result. □
Lemma 6.
Under standard assumptions S A 1 S A 5 , there exists a positive constant K 4 such that:
c k 2 c k + c k T s k n 2 K 4 c k min { δ k , c k } .
Proof. 
We consider two cases:
Firstly, from (20), if s k n c p = δ k c k c k ( c k c k ) and δ k c k T c k c k 2 c k c k 3 , then we have:
c k 2 c k + c k T s k n c p 2 = 2 ( c k c k ) T s k n c p s k n c p T c k c k T s k n c p = 2 δ k c k c k δ k 2 c k T c k c k 2 c k c k 2 2 δ k c k c k δ k c k c k δ k c k c k .
Secondly, from (20), if s k n c p = c k c k 2 c k T c k c k 2 ( c k c k ) and δ k c k T c k c k 2 c k c k 3 , then we have:
c k 2 c k + c k T s k n c p 2 = 2 ( c k c k ) T s k n c p s k n c p T c k c k T s k n c p = 2 c k c k 4 c k T c k c k 2 c k c k 4 c k T c k c k 2 = c k c k 4 c k T c k c k 2 c k c k 2 c k T c k c k 2 .
Using standard assumption S A 3 , we have c k c k c k ( c k T c k ) 1 c k . From inequalities (19), (43), (44), and using standard assumption S A 2 , we obtain the desired result.
From Algorithm 1 and Lemma 6, we have, for all k:
P r e d k σ k 2 K 4 c k min { δ k , c k } .
Lemma 7.
Under standard assumptions S A 1 S A 5 , there exists a constant K 5 > 0 , such that:
q k ( s k n ) q k ( s k n + Y k s ¯ k t ) 1 2 K 5 Y k T q k ( s k n ) min { Δ k , Y k T q k ( s k n ) B ¯ k } ,
where B ¯ k = Y k T B k Y k .
Proof. 
We consider two cases:
Firstly, from (24), if s ¯ k t c p = Δ k Y k T q k ( s k n ) Y k T q k ( s k n ) and Δ k ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Y k T q k ( s k n ) 3 , then we have:
q k ( s k n ) q k ( s k n + Y k s ¯ k t c p ) = ( Y k T q k ( s k n ) ) T s ¯ k t c p 1 2 s ¯ k t c p T B ¯ k s ¯ k t c p = Δ k Y k T q k ( s k n ) Δ k 2 2 Y k T q k ( s k n ) 2 [ ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) ] Δ k Y k T q k ( s k n ) 1 2 Δ k Y k T q k ( s k n ) 1 2 Δ k Y k T q k ( s k n ) .
Secondly, from (24), if s ¯ k t c p = Y k T q k ( s k n ) 2 Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Y k T q k ( s k n ) and Δ k ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Y k T q k ( s k n ) 3 , then we have:
q k ( s k n ) q k ( s k n + Y k s ¯ k t c p ) = ( Y k T q k ( s k n ) ) T s ¯ k t c p 1 2 s ¯ k t c p T B ¯ k s ¯ k t c p = Y k T q k ( s k n ) 4 ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Y k T q k ( s k n ) 4 2 ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) = Y k T q k ( s k n ) 4 2 ( Y k T q k ( s k n ) ) T B ¯ k Y k T q k ( s k n ) Y k T q k ( s k n ) 2 2 B ¯ k .
From inequalities (23), (47), (48), and using standard assumptions S A 1 S A 5 , we obtain the desired result.
The next lemma shows that FBACTR algorithm cannot be looped infinitely without finding an acceptable step.
Lemma 8.
Under standard assumptions S A 1 S A 5 , if there exists ε > 0 such that c k ε , then A r e d k j P r e d k j α 1 for some finite j.
Proof. 
Using (39), (45), and from c k ε , we have:
A r e d k P r e d k 1 = A r e d k P r e d k P r e d k 2 K 3 δ k 2 K 4 ε min { ε , δ k } .
δ k j becomes small as s k j gets rejected and eventually we will have:
A r e d k j P r e d k j 1 2 K 3 δ k j K 4 ε .
Then, the acceptance rule will be met for finite j. This completes the proof. □
Lemma 9.
Under standard assumptions S A 1 S A 5 and if jth trial step of iteration k satisfies
s k j min { ( 1 α 1 ) K 4 4 K 3 , 1 } c k ,
then it must be accepted.
Proof. 
This lemma is proved by contradiction. In case of assuming that inequality (49) holds, the trial step s k j is rejected, by using inequalities (39), (45), and (49), then:
( 1 α 1 ) < A r e d k j P r e d k j P r e d k j < 2 K 3 s k j 2 K 4 c k s k j ( 1 α 1 ) 2 .
This contradiction and therefore the lemma is proved.
Lemma 10.
Under standard assumptions S A 1 S A 5 , there exists δ k j satisfies:
δ k j min { δ m i n β 1 , τ 1 ( 1 α 1 ) K 4 4 K 3 , τ 1 } c k ,
for all trial steps j of any iteration k where β 1 is a positive constant independent of k or j.
Proof. 
At any trial iterate k j of any iteration k, we consider two cases:
Firstly, if j = 1 , then the step is accepted. That is δ k 1 δ m i n and take β 1 = sup x Ω c k , we have
δ k δ m i n δ m i n β 1 c k .
Secondly, if j > 1 , then there is at least one trial step which is rejected. From Lemma 9, we have
s k i > min { ( 1 α 1 ) K 4 4 K 3 , 1 } c k ,
for all trial steps i = 1 , 2 , j 1 which are rejected. Since s k i is a trial step which is rejected, then from the previous inequality and Algorithm 2, we have:
δ k j = τ 1 s k j 1 > τ 1 min { ( 1 α 1 ) K 4 4 K 3 , 1 } c k .
From inequality (51) and the above inequality, we obtain the desired result.
The next lemma obviously shows that as long as c k is bounded away from zero, the radius of the trust-region is bounded away from zero. □
Lemma 11.
Under standard assumptions S A 1 S A 5 , if there exists ε > 0 such that c k ε . Then there exists K 6 > 0 such that:
δ k j K 6 .
Proof. 
Let
K 6 = ε min { δ m i n β 1 , τ 1 ( 1 α 1 ) K 4 4 K 3 , τ 1 } ,
and using (50), the proof follows directly. □
In the next section, the iteration sequence convergence is studied when r k .

3.3. Convergence When the Positive Parameter r k

This section is devoted to the convergence of the iteration sequence when the positive parameter r k goes to infinity.
Notice that, we do not require [ g u i ( x ) , i I ( x ) ] has full column rank in standard assumption S A 4 , so, we may have other kinds of stationary points, which are defined in the following definitions.
Definition 2.
A feasible Fritz John (FFJ) point is a point x * that satisfies the following FFJ conditions:
η * f u ( x * ) + c ( x * ) μ * + g u ( x * ) λ * = 0 , c ( x * ) = 0 , P ( x * ) g ( x * ) = 0 , ( λ * ) i g u i ( x * ) = 0 , i = 1 , , m 1 , η * , ( λ * ) i 0 , i = 1 , , m 1 .
where η * , μ * , and λ * are not all zeros. For more details see [18].
If η * 0 , then the point ( x * , 1 , μ * η * , λ * η * ) is called a KKT point and FFJ conditions are called KKT conditions.
Definition 3.
An infeasible Fritz John (IFJ) point is a point x * that satisfies the following IFJ conditions:
η * f u ( x * ) + c ( x * ) μ * + g u ( x * ) λ * = 0 , c ( x * ) = 0 , g u ( x * ) P ( x * ) g u ( x * ) = 0 b u t P ( x * ) g u ( x * ) > 0 , ( λ * ) i g u i ( x * ) 0 , i = 1 , , m 1 , η * , ( λ * ) i 0 , i = 1 , , m 1 ,
where η * , μ * , and λ * are not all zeros. For more details see [18].
If η * 0 , then the point ( x * , 1 , μ * η * , λ * η * ) is called an infeasible KKT point and IFJ conditions are called infeasible KKT conditions.
Lemma 12.
Under standard assumptions S A 1 S A 5 , a subsequence { k i } of the sequence of the iteration satisfies IFJ conditions if the following conditions satisfied:
(i)
lim k i c ( x k i ) = 0 .
(ii)
lim k i P k i g u ( x k i ) > 0 .
(iii)
lim k i m i n s n m 1 + 1 P k i ( g u k i + g u k i T Y k i s ¯ t ) 2 = lim k i P k i g u k i 2 .
Proof. 
For simplification and without loss of generality, let { k i } represents the whole sequence { k } . Assume that s ˜ k is the solution of the subproblem m i n i m i z e s ¯ t P k ( g u k + g u k T Y k s ¯ t ) 2 , then it satisfies the following equation:
Y k T g u k P k g u k + Y k T g u k P k g u k T Y k s ˜ k = 0 .
It also satisfies the right hand side of Condition (iii). That is,
lim k { 2 s ˜ k T Y k T g u k P k g u k + s ˜ k T Y k T g u k P k g u k T Y k s ˜ k } = 0 .
We will consider two cases:
Firstly, if lim k s ˜ k = 0 , then from Equation (53) we have lim k Y k T g u k P k g u k = 0 .
Secondly, if lim k s ˜ k 0 , then by multiplying Equation (53) from the left by 2 s ˜ k T and subtract it from Equation (54), we have lim k P k g u k T Y k s ˜ k 2 = 0 . Hence lim k Y k T g u k P k g u k = 0 . That is, in two cases, we have
lim k Y k T g u k P k g u k = 0 .
Since lim k P k g u k > 0 , then lim k ( P k g u k ) i 0 , for i = 1 , , m 1 and lim k ( P k g u k ) i > 0 , for some i. Let ( λ k ) i = ( P k g u k ) i , i = 1 , , m 1 , then lim k Y k T g u k λ k = 0 . Hence, there exists a sequence of { μ k } such that lim k Y k T { c k μ k + g u k λ k } = 0 . That is, IFJ conditions hold in the limit with η * = 0 , see Definition 3. □
Lemma 13.
Under standard assumptions, S A 1 S A 5 , a subsequence { k i } of the sequence of the iteration satisfies FFJ conditions if the following conditions are satisfied:
(i)
lim k i c ( x k i ) = 0 .
(ii)
For all k i , P k i g u k i > 0 and lim k i P k i g u k i = 0 .
(iii)
lim k i min s n m 1 + 1 P k i ( g u k i + g u k i T Y k i s ¯ t ) 2 P k i g u k i 2 = 1 .
Proof. 
For simplification and without loss of generality, let { k i } represents the whole sequence { k } . Notice that the following equation,
lim k min ν n m 1 + 1 U k + P k g k T Y k ν 2 = 1 ,
is equivalent to Condition (iii), where U k is a unit vector in the direction of P k g u k and ν = s ¯ t P k g u k . Let ν ˜ k be a solution of the following problem:
min ν n m 1 + 1 U k + P k g k T Y k ν 2 .
Hence,
Y k T g u k P k g u k T Y k ν ˜ k + Y k T g u k P k U k = 0 .
Now two cases are considered:
Firstly, if lim k Y k ν ˜ k = 0 and using (58), then lim k Y k T g u k P k U k = 0 .
Secondly, if lim k Y k ν ˜ k 0 , then from (56) and the fact that ν ˜ k is a solution of problem (57) we have:
lim k { ν k ¯ T Y k T g u k P k g u k T Y k ν ˜ k + 2 U k T P k g u k T Y k ν ˜ k } = 0 .
Multiplying Equation (58) from the left by 2 ν ˜ k T and subtracting it from the above limit, we have the following equation: lim k ν ˜ k T Y k T g u k P k g u k T Y k ν ˜ k = 0 . That is lim k Y k g u k P k U k = 0 . Hence in both cases, we have lim k Y k g u k P k U k = 0 . The remnant of the proof follows using cases similar to those in Lemma 12. □
Lemma 14.
Under standard assumptions S A 1 S A 5 , if k represents the index of iteration at which σ k is increased, then we have:
r k c k 2 K 7 ,
where K 7 is a positive constant.
Proof. 
Since σ k is increased, then from Algorithm 1 we have:
σ k 2 [ c k 2 c k + c k T s k 2 ] = [ q k ( s k ) q k ( 0 ) + Δ μ k T ( c k + c k T s k ) ] + β 0 2 [ c k 2 c k + c k T s k 2 ] .
From (42), (50), and using the above equality, we have:
σ k 2 K 4 c k 2 min { δ m i n β 1 , τ 1 ( 1 α 1 ) K 4 4 K 3 , τ 1 , 1 } x k T s k + 1 2 s k T H k s k + Δ μ k T ( c k + c k T s k ) + r k 2 [ P k ( g u k + g u k T s k ) 2 P k g u k 2 ] + β 0 2 [ c k 2 c k + c k T s k 2 ] .
However, σ k r k 2 , then:
r k 2 2 K 4 c k 2 min { δ m i n β 1 , τ 1 ( 1 α 1 ) K 4 4 K 3 , τ 1 , 1 } x ( x k , μ k ) T s k + 1 2 s k T H k s k + Δ μ k T ( c k + c k T s k ) + r k 2 [ P k ( g u k + g u k T s k ) 2 + β 0 2 c k 2 .
Hence,
r k 2 K 4 c k 2 min { δ m i n β 1 , τ 1 ( 1 α 1 ) K 4 4 K 3 , τ 1 , 1 } 1 r k [ x k T s k + 1 2 s k T H k s k + Δ μ k T ( c k + c k T s k ) + β 0 2 c k 2 ] + 1 2 P k ( g u k + g u k T s k ) 2 1 r k [ | x k T s k | + 1 2 | s k T H k s k | + | Δ μ k T ( c k + c k T s k ) | + β 0 2 c k 2 ] + 1 2 P k ( g u k + g u k T s k ) 2 .
From Cauchy–Schwarz inequality, standard assumptions S A 3 S A 5 , and the fact that s k δ m a x , the proof is completed. □
Lemma 15.
Under standard assumptions S A 1 S A 5 , if r k and there is an infinite subsequence { k i } of the sequence of the iteration at which σ k is increased, then:
lim k i c k i = 0 .
Proof. 
From lemma (59) and using r k is unbounded, the proof is completed. □
Theorem 1.
Under standard assumptions S A 1 S A 5 , if r k as k , then
lim k c k = 0 .
Proof. 
See Theorem 4.18 [22]. □
Lemma 16.
Under standard assumptions S A 1 S A 5 , if there exists a subsequence { k j } of indices indexing iterates that satisfy P k g u k ε > 0 for all k { k j } and r k as k . Then a subsequence of the iteration sequence indexed { k j } satisfies IFJ conditions in the limit.
Proof. 
For simplification and without loss of generality, the total sequence { k } denotes to { k j } . This lemma is proved by contradiction, therefor we suppose there is no subsequence of the sequence { k } satisfies IFJ conditions in the limit. Using Lemma 12, we have for all k, | P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 | ε 1 for some ε 1 > 0 . From (55), we have Y k g u k P k g u k ε 2 , for some ε 2 > 0 , hence:
Y k T g u k P k g u k + Y k T g u k P k g u k T s k n Y k T g u k P k g u k Y k T g u k P k g u k T s k n ε 2 K 1 Y k T g u k P k g u k T c k .
Since { c k } converges to zero and Y k T g u k P k g u k T is bounded, then we can write: Y k T g u k P k g u k + Y k T g u k P k g u k T s k n ε 2 2 . Therefore,
Y k T q k ( s k n ) r k Y k T g u k P k g u k + Y k T g u k P k g u k T s k n Y k T x k + Y k T H k s k n r k ε 2 2 Y k T x k + Y k T H k s k n r k [ ε 2 2 1 r k Y k T x k + Y k T H k s k n ] .
From (46), we have:
q k ( s k n ) q k ( s k ) K 5 2 r k [ ε 2 2 1 r k Y k T [ x k + H k s k n ] ] min { Δ k , ε 2 2 1 r k Y k T [ x k + H k s k n ] Y k T g u k P k g u k T Y k + 1 r k Y k T H k Y k } .
For a k sufficiently large, we have:
q k ( s k n ) q k ( s k ) K 5 ε 2 4 r k min { Δ k , ε 2 2 Y k T g u k P k g u k T Y k } .
From Algorithm 3, { r k } is boundless only if there exist an infinite subsequence of indices { k i } , at which:
1 2 [ q k ( s k n ) q k ( s k ) ] < g u k P k g u k min { g u k P k g u k , δ k } .
Since r k , therefore an infinite number of acceptable iterates at which (62) holds and from the way of updating r k , we have r k as k . This gives a contradiction unless r k δ k is bounded and hence δ k 0 . Therefore s k 0 . We will consider two cases:
Firstly, if P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 > ε 1 , then we have
r k { P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 } > r k ε 1 .
Using (63) and standard assumptions S A 3 S A 5 , we have [ q k ( s k n ) q k ( s k ) ] . That is, the left hand side of inequality (62) goes to infinity while the right hand side tends to zero and this is a contradiction in this case.
Secondly, if P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 < ε 1 , then
r k { P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 | } < r k ε 1 ,
where r k as k and similar to the first case, [ q k ( s k n ) q k ( s k ) ] . This is a contradiction with [ q k ( s k n ) q k ( s k ) ] > 0 . The lemma is proved. □
Lemma 17.
Under standard assumptions S A 1 S A 5 , if r k as k , and there exists a subsequence indexed { k j } of iterates that satisfy P k g u k > 0 for all k { k j } and lim k j P k j g u k j = 0 , then a subsequence of the sequence of iterates indexed { k j } satisfies FFJ conditions in the limit.
Proof. 
Without loss of generality, let { k j } be the whole iteration sequence { k } to simplify. This lemma is proved by contradiction and so suppose that there is no subsequence that satisfies FFJ conditions in the limit. From condition (iii) of Lemma 13, for all k sufficiently large, there exists a constant ε 3 > 0 such that:
P k g u k 2 P k ( g u k + g u k T Y k s ¯ k t ) 2 P k g u k 2 ε 3 .
The following two cases are considered:
Firstly, if lim i n f k s ¯ k t P k g u k = 0 , then there is a contradiction with inequality (64).
Secondly, if lim s u p k s ¯ k t P k g u k = , then from subproblem (22) we have:
Y k T q k ( s k n ) = Y k T ( B k + υ k I ) Y k s ¯ k t ,
where υ k 0 represents the Lagrange multiplier vector, which is associated with the constraint Y k s ¯ t Δ k . From (65) and (46), we have:
q k ( s k n ) q k ( s k ) K 5 2 Y k T q k ( s k n ) min { Δ k , Y k T [ 1 r k H k + ( υ k r k I + g u k P k g u k T ) ] Y k s ¯ k t Y k T ( 1 r k H k + g u k P k g u k T ) Y k } .
Since r k as k , then there exists an infinite number of acceptable steps such that inequality (62) holds. However, inequality (62) can be written as follows:
1 2 [ q k ( s k n ) q k ( s k ) ] < β 2 2 P k g u k 2 ,
where β 2 = s u p x Ω Y k g u k .
From (66) and (67), we have:
K 5 2 Y k T q k ( s k n ) min { Δ k P k g u k , Y k T [ 1 r k H k + ( υ k r k I + g u k P k g u k T ) ] Y k s ¯ k t Y k T ( 1 r k H k + g u k P k g u k T ) Y k P k g u k } < 2 β 2 2 P k g u k .
However, in previous inequality, the right hand side tends to zero as k and also lim k i s ¯ k i t P k i g k i = along the subsequence { k i } . Therefore,
Y k i T q k i ( s k i n ) Y k i T [ 1 r k i H k i + ( υ k i r k i I + g k i P k i g k i T ) ] Y k i s ¯ k i t Y k i T ( 1 r k i H k i + g k i P k i g k i T ) Y k i P k i g k i ,
is bounded. That is, either s ¯ k i t P k i g k i lies in the null space of Y k i T ( υ k i r k i I + g k i P k i g k i T ) Y k i T or Y k i q k i ( s k i n ) 0 .
The first possibility occurs only when υ k i r k i 0 as k i and s ¯ k i t P k i g k i lie in the null space of the matrix Y k i T g k i P k i g k i T Y k i which is contradicted with assumption (64). This means that, FFJ conditions are satisfied in the limit. As k i , the second possibility is, Y k i q k i ( s k i n ) 0 and from (65), we have s ¯ k i t 0 which is contradicted with assumption (64). That is FFJ conditions are satisfied in the limit.
In the next section, the convergence of the sequence of the iteration sequence is studied when r k bounded.

3.4. Global Convergence When r k Is Bounded

Our analysis in this section is continued supposing that r k is bounded. Therefore, let k ¯ be an integer at which r k = r ¯ < for all k k ¯ . That is,
1 2 [ q k ( s k n ) q k ( s k ) ] g u k P k g u k min { g u k P k g u k , δ k } .
From assumptions S A 3 and S A 5 , and using (68), then for all k, there is a constant β 3 > 0 such that:
B k β 3 , Y k T B k β 3 , a n d Y k T B k Y k β 3 ,
where B k = H k + r ¯ g u k P k g u k T . □
Lemma 18.
Under standard assumptions S A 1 S A 5 , there exists a constant K 8 > 0 such that:
q k ( 0 ) q k ( s k n ) Δ μ k T ( c k + c k T s k ) K 8 c k .
Proof. 
Since
q k ( 0 ) q k ( s k n ) = x k T s k n 1 2 s k n T H k s k n + r ¯ 2 [ P k g u k 2 P k ( g u k + g u k T s k n ) 2 ] = ( x k + r ¯ g u k P k g u k ) T s k n 1 2 s k n T ( H k + r ¯ g u k P k g u k T ) s k n = ( x k + r ¯ g u k P k g u k ) T s k n 1 2 s k n T B k s k n ,
then we have:
q k ( 0 ) q k ( s k n ) Δ μ k T ( c k + c k T s k ) = ( x k + r ¯ g u k P k g u k ) T s k n 1 2 s k n T B k s k n Δ μ k T ( c k + c k T s k ) x k s k n r ¯ g u k P k g u k s k n B k s k n 2 Δ μ k c k + c k T s k [ x k + r ¯ g u k P k g u k + B k s k n ] s k n Δ μ k c k s k n .
From inequality (35) and the fact that Y k T c ( x k ) = 0 , then we have:
q k ( 0 ) q k ( s k n ) Δ μ k T ( c k + c k T s k ) [ ( x k + r ¯ g u k P k g u k + B k s k n + Δ μ k c k ) K 1 ] c k .
From standard assumptions S A 2 , S A 3 , S A 5 , the fact that s k n δ m a x , and using (69), then there exists K 8 > 0 , such that inequality (70) holds. □
Lemma 19.
Under standard assumptions S A 1 S A 5 , then for all k we have:
P r e d k 1 2 K 5 Y k T q k ( s k n ) min { Δ k , Y k T q k ( s k n ) s ¯ k } + g u k P k g u k min { g u k P k g u k , δ k } K 8 c k + σ k [ c k 2 c k + c k T s k 2 ] .
Proof. 
From (29), we have:
P r e d k = [ q k ( s k n ) q k ( s k ) ] + [ q k ( 0 ) q k ( s k n ) Δ μ k T ( c k + c k T s k ) ] + σ k [ c k 2 c k + c k T s k 2 ] = 1 2 [ q k ( s k n ) q k ( s k ) ] + 1 2 [ q k ( s k n ) q k ( s k ) ] + [ q k ( 0 ) q k ( s k n ) Δ μ k T ( c k + c k T s k ) ] + σ k [ c k 2 c k + c k T s k 2 ] .
Using inequalities (46), (68), and (70), we obtain the desired result. □
Lemma 20.
Under standard assumptions S A 1 S A 5 , if Y k T ( x k + r ¯ g u k P k g u k ) + g u k P k g u k ε > 0 and c k τ δ k where τ is a positive constant given by
τ min ε 6 β 3 K 1 δ m a x , 3 2 K 1 , K 5 ε 24 K 8 min { 2 ε 3 δ m a x , 1 } , ε 4 K 8 min { ε 2 δ m a x , 1 } ,
then there exists a constant K 9 > 0 such that:
P r e d k K 9 δ k + σ k [ c k 2 c k + c k T s k 2 ] .
Proof. 
Since Y k T ( x k + r ¯ g u k P k g u k ) + g u k P k g u k ε , then we can say Y k T ( x k + r ¯ g u k P k g u k ) ε 2 and g u k P k g u k ε 2 . We will consider two cases:
Firstly, if Y k T ( x k + r ¯ g u k P k g u k ) ε 2 , then from inequalities (69), (35) and c k τ δ k , we have:
Y k T ( x k + r ¯ g u k P k g u k + B k s k n ) Y k T ( x k + r ¯ g u k P k g u k ) Y k T B k s k n Y k T ( x k + r ¯ g u k P k g u k ) β 3 K 1 c k ε 2 β 3 K 1 τ δ k .
However, τ ε 6 β 3 K 1 δ m a x , then
Y k T ( x k + r ¯ g u k P k g u k + A k s k n ) ε 3 .
From inequality (35), assumption c k τ δ k , and using the value of τ in (72), we have s k n K 1 c k K 1 τ δ k K 1 3 2 K 1 δ k = 3 2 δ k . That is, Δ k 2 = δ k 2 s k n 2 δ k 2 3 4 δ k 2 = 1 4 δ k 2 . This means that,
Δ k 1 2 δ k .
From inequalities (71), (74), (75), and assumption c k τ δ k , we have the following:
P r e d k 1 2 K 5 Y k T ( x k + r ¯ g u k P k g u k + B k s k n ) min { Y k T ( x k + r ¯ g u k P k g u k + B k s k n ) , 1 2 δ k } K 8 c k + σ k [ c k 2 c k + c k T s k 2 ] K 5 ε 12 δ k min { 2 ε 3 δ m a x , 1 } K 8 τ δ k + σ k [ c k 2 c k + c k T s k 2 ] .
However, τ K 5 ε 24 K 8 min { 2 ε 3 δ m a x , 1 } , then we have
P r e d k K 5 ε 24 min { 2 ε 3 δ m a x , 1 } δ k + σ k [ c k 2 c k + c k T s k 2 ] .
Secondly, if g u k P k g u k ε 2 and using inequality (71), then
P r e d k g u k P k g u k min { g u k P k g u k , δ k } K 8 c k + σ k [ c k 2 c k + c k T s k 2 ] ε 2 min { ε 2 δ max , 1 } δ k K 8 τ δ k + σ k [ c k 2 c k + c k T s k 2 ] ε 4 min { ε 2 δ max , 1 } δ k + σ k [ c k 2 c k + c k T s k 2 ] ,
where τ ε 4 K 8 min { ε 2 δ max , 1 } . Let K 9 = min K 5 ε 24 min { 2 ε 3 δ m a x , 1 } , ε 4 min { ε 2 δ max , 1 } , then the result follows.
From the previous lemma, we notice that either Y k T ( x k + r ¯ g u k P k g u k ) ε 2 > 0 or g u k P k g u k ε 2 > 0 and c k τ δ k , where τ is given by (72) at any iteration k, the value of the penalty parameter σ k is not needed to increase. That is the penalty parameter σ k is increased only when c k τ δ k .
Lemma 21.
Under standard assumptions S A 1 S A 5 , if σ k is increased at kth iteration, then there is a positive constant K 10 such that:
σ k min { c k , δ k } K 10 .
Proof. 
From Algorithm 1, we have:
σ k 2 [ c k 2 c k + c k T s k 2 ] = [ q k ( s k ) q k ( s k n ) ] + [ q k ( s k n ) q k ( 0 ) ] + Δ μ k T ( c k + c k T s k ) + β 0 2 [ c k 2 c k + c k T s k 2 ] = 1 2 [ q k ( s k n ) q k ( s k ) ] 1 2 [ q k ( s k n ) q k ( s k ) ] + [ q k ( s k n ) q k ( 0 ) + Δ μ k T ( c k + c k T s k ) ] + β 0 2 [ c k 2 c k + c k T s k 2 ] ,
where σ k increased at any iteration and r k = r ¯ . From the previous equation, (42), (46), (68), and (70), we have
σ k 2 K 4 c k min { δ k , c k } K 5 2 Y k T q k ( s k n ) min { Δ k , Y k T q k ( s k n ) B ¯ k } g u k P k g u k min { g u k P k g u k , δ k } + K 8 c k + β 0 2 c k 2 K 8 c k + β 0 2 c k 2 .
Using assumption S A 3 , we get the desired result. □
Lemma 22.
Under standard assumptions S A 1 S A 5 and at the jth trial iterate of any iteration k. If σ k j is increased, then there is a constant K 11 > 0 , such that
σ k j c k K 11 .
Proof. 
From (50) and (76), we get the desired result. □
Lemma 23.
Under standard assumptions S A 1 S A 5 , if σ k , then
lim k i c k i = 0 ,
where { k i } is a subsequence indexes the iterates at which σ k is increased.
Proof. 
From Lemma 22 we obtain the desired result. □

3.5. Main Results for Global Convergence

In this section, main global convergence results for FBACTR algorithm are introduced.
Theorem 2.
Under standard assumptions S A 1 S A 5 , the sequence of iterates which is generated by FBACTR algorithm satisfies
lim k c k = 0 .
Proof. 
This theorem is proved by contradiction and so we suppose that lim sup k c k ε > 0 . This means that there exists an infinite subsequence of indices { k j } indexing iterates that satisfy c k j ε 2 . However, there exists an infinite sequence of acceptable steps from Lemma 8. Without loss of generality and to simplify, we suppose that all members of { k j } are acceptable iterates. Now, two cases are considered:
Firstly, if { σ k } is unbounded, then an infinite number of iterates { k i } exists and at which the penalty parameter σ k is increased. So, for k that is sufficiently large and from Lemma 23, let { k i } and { k j } be the two sequences which are not have common elements. Let k ρ 1 and k ρ 2 be two consecutive iterates at which σ k is increased and k ρ 1 < k < k ρ 2 , where k { k j } . The penalty parameter σ k is the same for all iterates that lie between k ρ 1 and k ρ 2 . Since all the iterates of { k j } are acceptable, then for all k { k j } ,
Φ k Φ k + 1 = A r e d k α 1 P r e d k .
Using inequality (45), we have:
Φ k Φ k + 1 σ k α 1 K 4 2 c k min { c k , δ k } .
Summing over all acceptable iterates that lie between k ρ 1 and k ρ 2 , we have:
k = k ρ 1 k ρ 2 1 Φ k Φ k + 1 σ k α 1 K 4 ε 4 min { K 6 ^ , ε 2 } ,
where K 6 ^ is as K 6 in (52)but ε is replaced by ε 2 . Hence,
( x k ρ 1 , μ k ρ 1 ; r ¯ ) ( x k ρ 2 , μ k ρ 2 ; r ¯ ) σ k ρ 1 + [ c k ρ 1 2 c k ρ 2 2 ] α 1 K 4 ε 4 min { K 6 ^ , ε 2 } .
Since σ k , then for k ρ 1 sufficiently large, we have:
( x k ρ 1 , μ k ρ 1 ; r ¯ ) ( x k ρ 2 , μ k ρ 2 ; r ¯ ) σ k ρ 1 < α 1 K 4 ε 8 min { K 6 ^ , ε 2 } .
Therefore,
c k ρ 1 2 c k ρ 2 2 α 1 K 4 ε 8 min { K 6 ^ , ε 2 } .
This leads to a contradiction with Lemma 23 unless ε = 0 .
Secondly, If { σ k } is bounded, then for all an integer k ˜ and k k ˜ , we have σ k = σ ˜ . Hence, for any k ^ { k j } where k ^ k ˜ and using (45), we have:
P r e d k ^ σ ˜ K 4 2 c k ^ min { δ k ^ , c k ^ } ε σ ˜ K 4 4 min { ε 2 δ m a x , 1 } δ k ^ .
Then for any k ^ { k j } , we have:
Φ k ^ Φ k ^ + 1 = A r e d k ^ α 1 P r e d k ^ ,
such that all the iterates of { k j } are acceptable. From above inequality, inequality (80) and using Lemma 11 we have:
Φ k ^ Φ k ^ + 1 α 1 ε σ ˜ K 4 4 min { ε 2 δ m a x , 1 } K 6 ^ > 0 .
However, this is a contradiction of the fact that { Φ k } is bounded when { σ k } is bounded. Therefore, we have a contradiction in both cases. Hence the supposition is not correct and this proves the theorem. □
Theorem 3.
Under standard assumptions S A 1 S A 5 , the sequence of iterates generated by FBACTR algorithm satisfies:
lim inf k [ Y k T x k + g u k P k g u k ] = 0 .
Proof. 
First, we prove that:
lim inf k [ Y k T ( x k + r ¯ g u k P k g u k ) + g u k P k g u k ] = 0 .
The proof of (82) is by contradiction, so, for all k, assume that Y k T ( x k + r ¯ g u k P k g u k ) + g u k P k g u k > ε . Let { k i } be an infinite subsequence at which c k i > τ δ k i , where τ is defined in (72). However, c k 0 , then
lim k i δ k i = 0 .
Let k j be any trial iterate belonging to { k i } and we consider two cases:
Firstly, if { σ k } is unbounded, then for the rejected trial step j 1 of iteration k { k i } , we have c k > τ δ k j = τ 1 τ s k j 1 . Since the trial step s k j 1 is rejected and using inequalities (45) and (41), then
( 1 α 1 ) | A r e d k j 1 P r e d k j 1 | P r e d k j 1 [ 2 κ 1 s k j 1 + 2 κ 2 σ k j 1 s k j 1 c k + 2 κ 3 σ k j 1 s k j 1 2 ] σ k j 1 K 4 min ( τ 1 τ , 1 ) c k 2 κ 1 σ k j 1 K 4 τ 1 τ min ( τ 1 τ , 1 ) + 2 κ 2 τ 1 τ + 2 κ 3 K 4 τ 1 τ min ( τ 1 τ , 1 ) s k j 1 .
However, { σ k } is unbounded, hence for all k k ^ , k ^ is sufficiently large, we have:
σ k j 1 > 4 κ 1 K 4 τ 1 τ min ( τ 1 τ , 1 ) ( 1 α 1 ) .
Therefore, for all k k ^ , we have:
s k j 1 K 4 τ 1 τ min ( τ 1 τ , 1 ) ( 1 α 1 ) 4 ( κ 2 τ 1 τ + κ 3 ) .
From Algorithm 2, we have:
δ k j = τ 1 s k j 1 K 4 τ 1 2 τ min ( τ 1 τ , 1 ) ( 1 α 1 ) 4 ( κ 2 τ 1 τ + κ 3 ) .
This gives a contradiction and this leads to δ k j not being able to go to zero in this case.
Secondly, if the sequence { σ k } is bounded, then there exists an integer k ¯ and σ ¯ such that for all k k ¯ , σ k = σ ¯ . Consider a trial step j of iteration k k ¯ and c k > τ δ k j , we consider three cases:
(i)
If j = 1 , then δ k j δ min , see Algorithm 2. This means that, δ k j is bounded in this case;
(ii)
If j > 1 , and c k l > τ δ k l for l = 1 , , j , then for all rejected trial steps l = 1 , , j 1 of iteration k k ¯ , we have
( 1 α 1 ) | A r e d k l P r e d k l | P r e d k l 2 K 6 s k l K 4 min ( τ , 1 ) c k .
Hence,
δ k j = τ 1 s k j 1 τ 1 K 4 min ( τ , 1 ) ( 1 α 1 ) c k 2 K 3 τ 1 K 4 min ( τ , 1 ) ( 1 α 1 ) τ 2 K 3 δ k 1 τ 1 K 4 min ( τ , 1 ) ( 1 α 1 ) τ 2 K 3 δ min .
That is, δ k j is also bounded in this case.
(iii)
If j > 1 and c k l > τ δ k l does not hold for all l, then there exists an integer ϱ such that c k l > τ δ k l holds for l = ϱ + 1 , , j and c k l τ δ k l holds for all l = 1 , , ϱ . As in case (ii), we can write:
δ k j τ 1 K 4 min ( τ , 1 ) ( 1 α 1 ) 2 K 3 c k τ 1 K 4 min ( τ , 1 ) ( 1 α 1 ) τ 2 K 3 δ k ϱ + 1 .
From Algorithm 2, we have:
δ k ϱ + 1 τ 1 s k ϱ .
From Lemma 20, if c k l τ δ k l and s k ϱ is rejected, then we have:
( 1 α 1 ) | A r e d k ϱ P r e d k ϱ | P r e d k ϱ 2 K 6 r ¯ s k ϱ K 9 .
That is,
s k ϱ K 9 ( 1 α 1 ) 2 K 6 σ ¯ .
This implies that s k ϱ is bounded and from (83) and (84) we have also δ k j is bounded in this case. That is in three cases, we have δ k j is bounded, but this leading to a contradiction. Hence, all the iterates satisfy c k τ δ k j for k j are sufficiently large. From Lemma 20, then the value of the penalty parameter is not needed to increase. Hence, { σ k } is bounded. Using Lemma 20 and for k j k ¯ , we have:
Φ k j Φ k j + 1 = A r e d k j α 1 P r e d k j α 1 K 9 δ k j .
As k , then:
lim k δ k j = 0 .
That is the trust-region radius is not bounded below and this leading to a contradiction. Because at iteration k j > k ¯ , if the previous step was accepted; i.e., at j = 1 , then δ k 1 δ min . That is δ k j is bounded in this case.
If j > 1 , then there exists at least one rejected trial step. From Lemmas 5 and 20, then for the rejected trial step s k j 1 we have:
( 1 α 1 ) < σ ¯ K 3 s k j 1 2 K 9 δ k j 1 .
From Algorithm 2, we have:
δ k j = τ 1 s k j 1 > τ 1 K 9 ( 1 α 1 ) σ ¯ K 3 .
Hence δ k j is bounded and this contradicts (85). That is, the supposition is wrong and hence,
lim inf k [ Y k T ( x k + r ¯ g u k P k g u k ) + g u k P k g u k ] = 0 .
That is, (81) holds and the proof is completed.
From the above two theorems, we conclude that, given any ε > 0 , the algorithm terminates because Y k T x k + g u k P k g u k + c k < ε , for some finite k. □

4. Numerical Results and Comparisons

In this section, we introduce an extensive variety of possible numeric NBLP problems to illustrate the validity of the proposed Algorithm FBACTR Algorithm 5 to solve the NBLP problem. The proposed algorithm FBACTR experimented on 16 benchmark examples given in [4,7,38,39,40].
Ten independent runs with a distinct initial value starting points for every test example are performed to observe the matchmaking of the result. Statistical results of all examples are briefed in Table 1 which displays that the results found by the FBACTR Algorithm 5 are approximate or equal to those by the compered algorithms in method [11] and the literature.
For comparison, the corresponding results of the mean number of iterations (iter), the mean number of function evaluations (nfunc), and the mean value of CPU time (CPUs) in seconds obtained by Methods in [11,41,42] respectively are included and summarized in Table 2. These results show that results of the FBACTR Algorithm 5 are approximate or equal to those of the compared algorithms in the literature.
It is evident from the results that our approach is able to handle NBLP problems even if the upper and the lower levels are convex or not and the computed results converge to the optimal solution which is similar or approximate to the optimal reported in the literature. Finally, it is obvious from the comparison between the solutions obtained using the FBACTR Algorithm 5 with those in the literature, that the FBACTR Algorithm 5 is capable of finding the optimal solution to some problems by a small number of iterations, a small number of function evaluations, and less time.
We offered the numerical results of FBACTR Algorithm 5 using MATLAB (R2013a) (8.2.0.701)64-bit(win64) and a starting point x 0 i n t ( F ˜ ) . The following parameter setting is used: δ m i n = 10 4 , δ 0 = m a x ( s 0 c p , δ m i n ) , δ m a x = 10 4 δ 0 , α 1 = 10 3 , α 2 = 0.8 , τ 1 = 0.5 , τ 2 = 2 , ε 1 = 10 10 , and ε 2 = 10 12 .

5. Conclusions

In this paper, the FBACTR Algorithm 5 is presented to solve the NBLP problem (1). A KKT condition is used with the Fischer–Burmeister function and an active-set strategy to convert the NBLP problem to an equivalent smooth equality constrained optimization problem. To ensure global convergence for the FBACTR algorithm, a trust-region globalization strategy is used.
A global convergence theory for the FBACTR algorithm is introduced and applications to mathematical programs with equilibrium constraints are provided to clarify the effectiveness of the proposed approach. Numerical results reflect the good behavior of the FBACTR algorithm and the computed results converge to the optimal solutions. It is clear from the comparison between the solutions obtained using the FBACTR algorithm with algorithms [11,41,42] that the FBACTR can find the optimal solution to some problems with a small number of iterations, small number of function evaluations, and in less time.
Test Problem 1 [41]:
min v f u = w 1 2 + w 2 2 + v 2 4 v s . t . 0 v 2 , min w f l = w 1 2 + 0.5 w 2 2 + w 1 w 2 + ( 1 3 v ) w 1 + ( 1 + v ) w 2 , s . t . 2 w 1 + w 2 2 v 1 , w 1 0 , w 2 0 .
Test Problem 2 [41]:
min v f u = w 1 2 + w 3 2 w 1 w 3 4 w 2 7 v 1 + 4 v 2 s . t . v 1 + v 2 1 , v 1 0 , v 2 0 min w f l = w 1 2 + 0.5 w 2 2 + 0.5 w 3 2 + w 1 w 2 + ( 1 3 v 1 ) w 1 + ( 1 + v 2 ) w 2 , s . t . 2 w 1 + w 2 w 3 + v 1 2 v 2 + 2 0 , w 1 0 ; w 2 0 w 3 0 .
Test Problem 3 [41]:
min v f u = 0.1 ( v 1 2 + v 2 2 ) 3 w 1 4 w 2 + 0.5 ( w 1 2 + w 2 2 ) s . t . min w f l = 0.5 ( w 1 2 + 5 w 2 2 ) 2 w 1 w 2 v 1 w 1 v 2 w 2 , s . t . 0.333 w 1 + w 2 2 0 , w 1 0.333 w 2 2 0 , w 1 0 , w 2 0 ,
Test Problem 4 [41]:
min v f u = v 1 2 2 v 1 + v 2 2 2 v 2 + w 1 2 + w 2 2 s . t . v 1 0 , v 2 0 min w f l = ( w 1 v 1 ) 2 + ( w 2 v 2 ) 2 , s . t . 0.5 w 1 1.5 , 0.5 w 2 1.5 ,
Test Problem 5 [41]:
min v f u = v 2 + ( w 10 ) 2 s . t . v + w 0 , 0 v 15 , min w f l = ( v + 2 w 30 ) 2 , s . t . v + w 20 , 0 w 20 ,
Test Problem 6 [41]:
min v f u = ( v 1 1 ) 2 + 2 w 1 2 2 v 1 s . t . v 1 0 , min w f l = ( 2 w 1 4 ) 2 + ( 2 w 2 1 ) 2 + v 1 w 1 , s . t . 4 v 1 + 5 w 1 + 4 w 2 12 , 4 v 1 5 w 1 + 4 w 2 4 , 4 v 1 4 w 1 + 5 w 2 4 , 4 v 1 + 4 w 1 + 5 w 2 4 , w 1 0 , w 2 0 ,
Test Problem 7 [41]:
min v f u = ( v 5 ) 2 + ( 2 w + 1 ) 2 s . t . v 0 , min w f l = ( 2 w 1 ) 2 1.5 v w , s . t . 3 v + w 3 , v 0.5 w 4 , v + w 7 , w 0 .
Test Problem 8 [41]:
min v f u = v 1 2 3 v 1 + v 2 2 3 v 2 + w 1 2 + w 2 2 s . t . v 1 0 , v 2 0 , min w f l = ( w 1 v 1 ) 2 + ( w 2 v 2 ) 2 , s . t . 0.5 w 1 1.5 , 0.5 w 2 1.5 ,
Test Problem 9 [3]:
min v f u = 16 v 2 + 9 w 2 s . t . 4 v + w 0 , v 0 , min w f l = ( v + w 20 ) 4 , s . t . 4 v + w 50 0 , w 0 .
Test Problem 10 [3]:
min v f u = v 1 3 w 1 + w 2 s . t . 0 v 1 1 , min w f l = w 2 s . t . v 1 w 1 10 , w 1 2 + v 1 w 2 1 , w 2 0 .
Test Problem 11 [42]:
min v f u = 2 v 1 + 2 v 2 3 w 1 3 w 2 60 s . t . v 1 + v 2 + w 1 2 w 2 40 , 0 v 1 50 , 0 v 2 50 , min w f l = ( w 1 v 1 + 20 ) 2 + ( w 2 v 2 + 20 ) 2 , s . t . v 1 2 w 1 10 , v 2 2 w 2 10 , 10 w 1 20 , 10 w 2 20 .
Test Problem 12 [3]:
min v f u = ( v 3 ) 2 + ( w 2 ) 2 s . t . 2 v + w 1 0 , v 2 w + 2 0 , v + 2 w 14 0 , 0 v 8 , min w f l = ( w 5 ) 2 s . t . w 0 .
Test Problem 13 [42]:
min v f u = v 1 2 3 v 2 2 4 w 1 + w 2 2 s . t . v 1 2 + 2 v 2 4 , v 1 0 , v 2 0 , min w f l = 2 v 1 2 + w 1 2 5 w 2 , s . t . v 1 2 2 v 1 + 2 v 2 2 2 w 1 + w 2 3 , v 2 + 3 w 1 4 w 2 4 , w 1 0 , w 2 0 .
Test Problem 14 [42]:
min v f u = ( v 1 ) 2 + ( w 1 ) 2 s . t . v 0 , min w f l = 0.5 w 2 + 500 w 50 v w s . t . y 0 .
Test Problem 15 [42]:
min v f u = 8 v 1 4 v 2 + 4 w 1 40 w 2 4 w 3 s . t . v 1 0 , v 2 0 min w f l = v 1 + 2 v 2 + w 1 + w 2 + 2 w 3 , s . t . w 2 + w 3 w 1 1 , 2 v 1 w 1 + 2 w 2 0.5 w 3 1 , 2 v 2 + 2 w 1 w 2 0.5 w 3 1 , w i 0 , i = 1 , 2 , 3 .
Test Problem 16 [42]:
min v f u = 8 v 1 4 v 2 + 4 w 1 40 w 2 4 w 3 s . t . v 1 0 , v 2 0 min w f l = 1 + v 1 + v 2 + 2 w 1 w 2 + w 3 6 + 2 v 1 + w 1 + w 2 3 w 3 , s . t . w 1 + w 2 + w 3 + w 4 = 1 , 2 v 1 w 1 + 2 w 2 0.5 w 3 + w 5 = 1 , 2 v 2 + 2 w 1 w 2 0.5 w 3 + w 6 = 1 , w i 0 , i = 1 , , 6 .

Author Contributions

B.E.: Conceptualization and software G.A.: formal analysis and writing. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Acknowledgments

The author would like to thank the anonymous referees for their valuable comments and suggestions which have helped to greatly improve this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bialas, W.; Karwan, M. On two-level optimization. IEEE Trans. Autom. Control. 1982, 27, 211–214. [Google Scholar] [CrossRef]
  2. Dempe, S. Foundation of Bilevel Programming; Kluwer Academic: London, UK, 2002. [Google Scholar]
  3. Gumus, H.; Flouda, A. Global Optimization of Nonlinear Bilevel Programming Problems. J. Glob. Optim. 2001, 20, 1–31. [Google Scholar] [CrossRef]
  4. Muu, D.; Quy, N. A Global Optimization Method for Solving Convex Quadratic Bilevel Programming Problems. J. Glob. Optim. 2003, 26, 199–219. [Google Scholar] [CrossRef]
  5. Abo-Elnaga, Y.; El-Shorbagy, M. Multi-Sine Cosine Algorithm for Solving Nonlinear Bilevel Programming Problems. Int. J. Comput. Intell. Syst. 2020, 13, 421–432. [Google Scholar] [CrossRef] [Green Version]
  6. Abo-Elnaga, Y.; Nasr, S. Modified Evolutionary Algorithm and Chaotic Search for Bilevel Programming Problems. Symmetry 2020, 12, 767. [Google Scholar] [CrossRef]
  7. Falk, J.; Liu, J. On bilevel programming, Part I: General nonlinear cases. Math. Program. 1995, 70, 47–72. [Google Scholar] [CrossRef]
  8. Ma, L.; Wang, G. A Solving Algorithm for Nonlinear Bilevel Programing Problems Based on Human Evolutionary Model. Algorithms 2020, 13, 260. [Google Scholar] [CrossRef]
  9. Savard, G.; Gauvin, J. The steepest descent direction for the nonlinear bilevel programming problem. Oper. Res. Lett. 1994, 15, 265–272. [Google Scholar] [CrossRef]
  10. Edmunds, T.; Bard, J. Algorithms for nonlinear bilevel mathematical programs. IEEE Trans. Syst. Man Cybern. 1991, 21, 83–89. [Google Scholar] [CrossRef]
  11. El-Sobky, B.; Ashry, G. An interior-point trust-region algorithm to solve a nonlinear bilevel programming problem. AIMS Math. 2022, 7, 5534–5562. [Google Scholar] [CrossRef]
  12. Chen, J. The semismooth-related properties of a merit function and a descent method for the nonlinear complementarity problem. J. Glob. Optim. 2006, 36, 565–580. [Google Scholar] [CrossRef]
  13. Chen, J. On some NCP-functions based on the generalized Fischer–Burmeister function. Asia-Pac. J. Oper. Res. 2007, 24, 401–420. [Google Scholar] [CrossRef] [Green Version]
  14. Chen, J.; Pan, S. A family of NCP-functions and a descent method for the nonlinear complementarity problem. Comput. Optim. Appl. 2008, 40, 389–404. [Google Scholar] [CrossRef] [Green Version]
  15. Facchinei, F.; Jiang, H.; Qi, L. A smoothing method for mathematical programming with equilibrium constraints. Math. Program. 1999, 85, 107–134. [Google Scholar] [CrossRef]
  16. Byrd, R.; Hribar, M.; Nocedal, J. An interior point algorithm for largescale nonlinear programming. SIAM J. Optim. 1999, 9, 877–900. [Google Scholar] [CrossRef]
  17. Byrd, R.; Gilbert, J.; Nocedal, J. A trust region method based on interior point techniques for nonlinear programming. Math. Program. 2000, 89, 149–185. [Google Scholar] [CrossRef] [Green Version]
  18. Bazaraa, M.; Sherali, H.; Shetty, C. Nonlinear Programming Theory and Algorithms; John Wiley and Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  19. Curtis, F.E.; Schenk, O.; Wachter, A. An interior-point algorithm for large-scale nonlinear optimization with inexact step computations. Siam J. Sci. Comput. 2010, 32, 3447–3475. [Google Scholar] [CrossRef]
  20. Esmaeili, H.; Kimiaei, M. An efficient implementation of a trust-region method for box constrained optimization. J. Appl. Math. Comput. 2015, 48, 495–517. [Google Scholar] [CrossRef]
  21. El-Sobky, B. A Multiplier active-set trust-region algorithm for solving constrained optimization problem. Appl. Math. Comput. 2012, 219, 127–157. [Google Scholar] [CrossRef]
  22. El-Sobky, B. An active-set interior-point trust-region algorithm. Pac. J. Optim. 2018, 14, 125–159. [Google Scholar] [CrossRef] [Green Version]
  23. El-Sobky, B.; Abotahoun, A. An active-set algorithm and a trust-region approach in constrained minimax problem. Comput. Appl. Math. 2018, 37, 2605–2631. [Google Scholar] [CrossRef]
  24. El-Sobky, B.; Abotahoun, A. A trust-region Algorithm for Solving Mini-Max Problem. J. Comput. Math. 2018, 36, 881–902. [Google Scholar]
  25. El-Sobky, B.; Abouel-Naga, Y. A penalty method with trust-region mechanism for nonlinear bilevel optimization problem. J. Comput. Appl. Math. 2018, 340, 360–374. [Google Scholar] [CrossRef]
  26. El-Sobky, B.; Abo-Elnaga, Y.; Mousa, A.; El-Shorbagy, A. trust-region based penalty barrier algorithm for constrained nonlinear programming problems: An application of design of minimum cost canal sections. Mathematics 2021, 9, 1551. [Google Scholar] [CrossRef]
  27. Kouri, D.; Heinkenschloss, M.; Ridzal, D.; van Waanders, B. A trust-region Algorithm with Adaptive Stochastic Collocation for PDE Optimization under Uncertainty. SIAM J. Sci. Comput. 2020, 35, 1847–1879. [Google Scholar] [CrossRef] [Green Version]
  28. Li, N.; Xue, D.; Sun, W.; Wang, J. A stochastic trust-region method for unconstrained optimization problems. Math. Probl. Eng. 2019, 2019, 8095054. [Google Scholar] [CrossRef] [Green Version]
  29. Niu, L.; Yuan, Y. A new trust region algorithm for nonlinear constrained optimization. J. Comput. Math. 2020, 28, 72–86. [Google Scholar]
  30. Wang, X.; Yuan, Y. A trust region method based on a new affine scaling technique for simple bounded optimization. Optim. Methods Softw. 2013, 28, 871–888. [Google Scholar] [CrossRef]
  31. Wang, X.; Yuan, Y. An augmented Lagrangian trust region method for equality constrained optimization. Optim. Methods Softw. 2015, 30, 559–582. [Google Scholar] [CrossRef]
  32. Zeng, M.; Ni, Q. A new trust region method for nonlinear equations involving fractional mode. Pac. J. Optim. 2019, 15, 317–329. [Google Scholar]
  33. Byrd, R. Robust trust-region methods for nonlinearly constrained optimization. In Proceedings of the Second SIAM Conference on Optimization, Houston, TX, USA, 18–20 May 1987. [Google Scholar]
  34. Omojokun, E. Trust-Region Strategies for Optimization with Nonlinear Equality and Inequality Constraints. Ph.D. Thesis, Department of Computer Science, University of Colorado, Boulder, CO, USA, 1989. [Google Scholar]
  35. El-Sobky, B.; Abouel-Naga, Y. Multi-objective optimal load flow problem with interior-point trust-region strategy. Electr. Power Syst. Res. 2017, 148, 127–135. [Google Scholar] [CrossRef]
  36. Dennis, J.; El-Alem, M.; Williamson, K. A trust-region approach to nonlinear systems of equalities and inequalities. SIAM J. Optim. 1999, 9, 291–315. [Google Scholar] [CrossRef]
  37. Dennis, J.; Heinkenschloss, M.; Vicente, L. trust-region interior-point SQP algorithms for a class of nonlinear programming problems. SIAM J. Control. Optim. 1998, 36, 1750–1794. [Google Scholar] [CrossRef] [Green Version]
  38. Bard, J.F. Convex two-level optimization. Math. Program. 1988, 40, 15–27. [Google Scholar] [CrossRef]
  39. Oduguwa, V.; Roy, R. Bi-level optimization using genetic algorithm. In Proceedings of the IEEE international Conference Artificial Intelligence Systems, Divnomorskoe, Russia, 5–10 September 2002; pp. 123–128. [Google Scholar]
  40. Shimizu, K.; Aiyoshi, E. A new computational method for Stackelberg and min-max problems by use of a penalty method. IEEE Trans. Autom. Control 1981, 26, 460–466. [Google Scholar] [CrossRef]
  41. Li, H.; Jiao, Y.; Zhang, L. Orthogonal genetic algorithm for solving quadratic bilevel programming problems. J. Syst. Eng. Electron. 2010, 21, 763–770. [Google Scholar] [CrossRef] [Green Version]
  42. Wang, Y.; Jiao, Y.; Li, H. An evolutionary algorithm for solving nonlinear bilevel programming based on a new constraint-Handling scheme. IEEE Trans. Syst. Man Cybern. Part C 2005, 35, 221–232. [Google Scholar] [CrossRef]
Table 1. Comparisons of the results of FBACTR Algorithm 5 with the method [11] and methods in the reference.
Table 1. Comparisons of the results of FBACTR Algorithm 5 with the method [11] and methods in the reference.
Problem Name ( v * , w * )
Method [11]
f u *
f l *
Method [11]
( v * , w * )
FBACTR
Algorithm 5
f u *
f l *
FBACTR Algorithm 5
( v * , w * )
Ref.
f u *
f l *
Ref.
TP1(0.8503, 0.0227,−2.6764(0.8465, 0.7695, 0)−2.0772(0.8438, 0.7657, 0)−2.0769
0.03589)0.0332 −0.5919 −0.5863
TP2(0.609, 0.391, 0,0.6086(0.6111, 0.3890, 0,0.64013(0.609, 0.391, 0,0.6426
0, 1.828)1.67130, 1.8339)1.68160, 1.828)1.6708
TP3(0.97, 3.14,−8.92(0.97, 3.14−8.92(0.97, 3.14,−8.92
2.6, 1.8)−6.052.6, 1.8)−6.052.6, 1.8)−6.05
TP4(0.5, 0.5, 0.5, 0.5)−1(0.5, 0.5, 0.5, 0.5)−1(0.5, 0.5, 0.5, 0.5)−1
0 0 0
TP5(9.839, 10.059)96.809(9.9953, 9.9955)99.907(10.03, 9.969)100.58
0.0019 1.8628 × 10 4 0.001
TP6(1.6879, 0.8805, 0)−1.3519(1.8889, 8.8889 × 10 1 ,−1.4074NA3.57
7.4991 6.8157 × 10 6 )7.6172 2.4
TP7(1, 0)17(1, 0)17(1, 0)17
1 1 1
TP8(0.75, 0.75,−2.25(0.7513, 0.7513,−2.2480( 3 / 2 , 3 / 2 ,−2.1962
0.75, 0.75)00.752, 0.752)0 3 / 2 , 3 / 2 )0
TP9(11.138, 5)2209.8(11.25, 5)2250(11.25, 5)2250
222.52 197.753 197.753
TP10(1, 0, 6.6387 × 10 6 ) 6.6387 × 10 6 (1, 0, 1)1(1, 0, 1)1
6.6387 × 10 6 −1 −1
TP11(24.972, 29.653,4.9101(25, 30, 5, 10)5(25, 30, 5, 10)5
5.0238, 9.7565)0.01332 0 0
TP12(3, 5)9(3, 5)9(3, 5)9
0 0 0
TP13(0, 1.7405,−15.548(0, 2, 1.875, 0.9063)−12.68(0, 2, 1.875, 0.9063)−12.68
1.8497, 0.9692)−1.4247 −1.016 −1.016
TP14(10.016, 0.81967)81.328(10, 0.011) 8.1978 × 10 1 (10.04, 0.1429)82.44
−0.3359 0 0.271
TP15(0, 0.9, 0, 0.6, 0.4)−29.2(0, 0.9, 0, 0.6, 0.4)−29.2(0, 0.9, 0, 0.6, 0.4)−29.2
3.2 3.2 3.2
TP16(0, 0.9, 0, 0.6,−29.2(0, 0.9, 0, 0.6,−29.2(0, 0.9, 0, 0.6,−29.2
0.4, 0, 0, 0)0.31480.4, 0, 0, 0)0.31480.4, 0, 0, 0)0.3148
Table 2. Comparisons of the results of FBACTR Algorithm 5 with method [11], method [41] and method [42] with respect to the number of iterations, the number of function evaluations, and time/s.
Table 2. Comparisons of the results of FBACTR Algorithm 5 with method [11], method [41] and method [42] with respect to the number of iterations, the number of function evaluations, and time/s.
Problem NameIter Method [11]nfunc Method [11]CPUs Method [11]Iter FBACTR Algorithmnfunc FBACTR AlgorithmCPUs FBACTR AlgorithmCPUs Method [41]CPUs Method [42]
TP111121.4310131.621.734-
TP210141.9879121.872.375-
TP3682.9782.523.31511.854
TP410141.6812131.921.576-
TP5691.635671.5231.8255.888
TP66114.18103.954.68925.332
TP712131.911121.6521.769-
TP810111.00211120.9531.124-
TP910131.958101.87--
TP10572.987563.31--
TP119123.74210133.632-37.308
TP12891.23791.33--
TP13572.1581.998-14.42
TP14682.12561.97-4.218
TP155620.5126720.125-45.39
TP165740.3194535.21-107.55
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Elsobky, B.; Ashry, G. An Active-Set Fischer–Burmeister Trust-Region Algorithm to Solve a Nonlinear Bilevel Optimization Problem. Fractal Fract. 2022, 6, 412. https://doi.org/10.3390/fractalfract6080412

AMA Style

Elsobky B, Ashry G. An Active-Set Fischer–Burmeister Trust-Region Algorithm to Solve a Nonlinear Bilevel Optimization Problem. Fractal and Fractional. 2022; 6(8):412. https://doi.org/10.3390/fractalfract6080412

Chicago/Turabian Style

Elsobky, Bothina, and Gehan Ashry. 2022. "An Active-Set Fischer–Burmeister Trust-Region Algorithm to Solve a Nonlinear Bilevel Optimization Problem" Fractal and Fractional 6, no. 8: 412. https://doi.org/10.3390/fractalfract6080412

APA Style

Elsobky, B., & Ashry, G. (2022). An Active-Set Fischer–Burmeister Trust-Region Algorithm to Solve a Nonlinear Bilevel Optimization Problem. Fractal and Fractional, 6(8), 412. https://doi.org/10.3390/fractalfract6080412

Article Metrics

Back to TopTop