Next Article in Journal
Research on Multi-UAV Cooperative Dynamic Path Planning Algorithm Based on Conflict Search
Next Article in Special Issue
A Systematic Survey of Transformer-Based 3D Object Detection for Autonomous Driving: Methods, Challenges and Trends
Previous Article in Journal
Optimizing Disaster Response through Efficient Path Planning of Mobile Aerial Base Station with Genetic Algorithm
Previous Article in Special Issue
CooPercept: Cooperative Perception for 3D Object Detection of Autonomous Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles

1
School of Civil Aviation, Northwestern Polytechnical University, 127 Youyi West Street, Xi’an 710072, China
2
Yangtze River Delta Research Institution, Northwestern Polytechnical University, 27 Zigang Road, Taicang 215400, China
3
School of Aeronautics and Astronautics, Tiangong University, 399 Binshui West Road, Tianjin 300387, China
4
School of Astronautics, Northwestern Polytechnical University, 127 Youyi West Street, Xi’an 710072, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(6), 273; https://doi.org/10.3390/drones8060273
Submission received: 27 April 2024 / Revised: 17 June 2024 / Accepted: 17 June 2024 / Published: 20 June 2024

Abstract

:
In this paper, we explore the problem of smooth braking at preset points in autonomous vehicles using model predictive control (MPC) with a receding horizon extended state observer (RHESO) and a neural network (NN). An NN-based modeling method is proposed to intuitively describe the relationship between vehicle speed and the vehicle controllers (brake and throttle), and establish a dynamic model of autonomous vehicles. A sufficient condition is put forward to guarantee the convergence of the proposed NN. Furthermore, a composite MPC strategy based on RHESO is designed, which optimizes a given cost function over the receding horizon while mitigating the effects of modeling inaccuracies and disturbances. Additionally, easily verifiable conditions are provided to ensure the autonomous driving vehicles’ uniform boundedness. Numerically illustrative examples are given to demonstrate the effectiveness of the proposed approach.

1. Introduction

Generally, non-stationary braking in vehicles poses major safety hazards, which may cause vehicle imbalance and accidents [1], resulting in uncomfortable riding experiences for passengers [2], damaging brake system components, increasing repair and maintenance costs [3], and even posing a threat to pedestrians on the road [4].
Unlike manual braking, the emergence of autonomous vehicles (AVs) makes it easier for vehicles to brake smoothly and accurately. Accuracy and smoothness are the two key performance criteria of smooth braking in AVs, meaning low tracking errors can be achieved without aggressive braking/throttle operations. Furthermore, smooth and precise braking in AVs has potential applications in many fields, such as bus passenger pick-up and drop-off [5], truck loading and unloading of goods at preset points [6], and passenger elevator docking with airplanes [7]. In these application scenarios, AVs must stop smoothly and precisely at preset points to ensure comfortable riding experiences for passengers and avoid collisions. Moreover, smooth braking at preset points in AVs significantly improves both parking efficiency and parking space utilization.
Smooth braking at preset points in AVs is a typical position and speed (PS) tracking control problem, which is also a dynamic control problem [8,9]. Existing PS tracking methods can be broadly categorized into three groups. The first group of methods does not consider vehicle dynamics or disturbance models. Control commands are solely generated based on tracking errors, employing designs such as proportional–integral–derivative (PID) controllers and their variants [10,11]. The disadvantage is that control precision heavily depends on adjustable parameters with high conservation and poor portability. The second method type relies on precise dynamic models that overlook disturbances to better predict PS information [12]. In [13], the authors analyzed a smooth braking torque control strategy for a brushless DC motor. In [14], a preview of a servo-loop speed control algorithm was proposed to achieve smooth, accurate, and computationally inexpensive speed tracking for AVs. Furthermore, cooperative braking control was introduced to ensure that AVs maintained safe spacing and stopped rapidly, smoothly, and accurately at the designated target positions with zero velocity [15]. These studies range from torque control based on brushless DC motor models to designing smooth and accurate braking methods using brushless DC motor models to designing smooth and accurate braking methods based on vehicle kinematic models, achieving various conditions for smooth and accurate braking of autonomous vehicles from hardware to software levels. However, precise dynamic models of AVs are difficult to obtain, and the real-time performance of control calculations using these models is often poor. A common drawback of the first two method types is their inability to handle the inherent constraints of AVs, such as speed limits, front-wheel deviation limits, and control input constraints
Different from the first two groups of methods, the third type of control strategy is based on model predictive control (MPC), which effectively handles various constraints. By minimizing the difference between the actual and predicted PS, the optimal control signal can be generated through repeated online optimization [16,17]. Under the MPC framework, there are schemes that use dynamic models of AVs [18,19], as well as schemes that use kinematic models of AVs [20,21]. Note that schemes using dynamic models for tracking PS constitute a servo control problem [22,23], with the advantage of directly tracking speed as a state variable. However, the disadvantage is that accurate model parameters are hard to obtain, and the real-time solution of the control signal cannot be guaranteed due to strong nonlinearity. In contrast, the scheme using kinematic models of AVs essentially addresses a speed planning problem [24], with the advantage of easily obtaining precise model parameters and ensuring real-time control signal solutions; the disadvantage is that the planned speed must be tracked by the servo layer, which leads to delays in speed tracking.
To maintain the high tracking accuracy of servo control while reducing the optimization load, the best approach is to combine the strengths of both the kinematic and dynamic models to establish a new dynamic model that can effectively handle the servo control of PS. One of the most direct methods is to add the speed variable as a state variable into the kinematic model. Correspondingly, throttle and brake openings are used as control inputs to manage the AVs’ speed servo control. However, due to unknown nonlinearities and disturbances, the relationship between the throttle (or brake) opening and AVs’ speed is challenging to determine. As an excellent tool for approximating nonlinear functions, neural networks (NNs) can describe the relationship between AVs’ speed and the throttle (or brake) opening [25,26]. There is significant research on using NNs to approximate unknown nonlinear system models, but few studies have applied NNs to autonomous vehicle modeling.
In addition, distractions such as pedestrians in vehicles, uneven road surfaces, and the coupling between equipment inside the vehicle often affect the accurate tracking of AVs’ PS. To minimize the impact of these disturbances, it is necessary to design disturbance observers for autonomous vehicles. The existing observers are primarily divided into two categories. The first type of observer is designed using the dynamic characteristics of the disturbance, which usually requires prior knowledge of the dynamic model parameters of the disturbance [27,28,29,30]. Another type of disturbance observer is the extended state observer (ESO); it can estimate the disturbance with a bounded rate of change and it has a wider application range [31,32,33]. However, these studies mainly focus on the final observation accuracy of the disturbance observer, while neglecting the stationarity of the disturbance observation process (i.e., the disturbance observation process may have a significant overshoot). One of the goals of this paper is to comprehensively consider the disturbance estimation process and smooth braking, as well as realize smooth and accurate braking at preset points in AVs while canceling the disturbance effects.
This paper aims to address the problem of smooth braking at preset points in AVs with a receding horizon extended state observer (RHESO)-based composite MPC strategy. The main contributions of this paper are as follows:
(1)
A novel extended kinematic model of AVs is established using a proposed NN such that the relationship between the AVs’ speed and throttle (and brake) opening is more clearly described;
(2)
A novel composite MPC strategy based on the RHESO is put forward, which is optimized over the receding horizon while eliminating the effects of the model’s inaccuracy and disturbance;
(3)
Easily verifiable sufficient conditions are established to ensure the recursive feasibility of the MPC scheme and the stability of the closed-loop system.
Finally, the notation used in this article are presented in Table 1.

2. Problem Formulations and Preliminaries

2.1. System Model

A general vehicle kinematic model is given as follows:
ξ c ˙ ( t ) = x ˙ ( t ) y ˙ ( t ) θ ˙ ( t ) = v ( t ) cos ( θ ( t ) ) v ( t ) sin ( θ ( t ) ) ω ( t )
where ξ c ( t ) R 3 indicates the state vector, x ( t ) R and y ( t ) R are the reference coordinates on the rear axle center in a geodetic coordinate system, θ ( t ) R denotes the heading angle, u c ( t ) = [ ω ( t ) v ( t ) ] T R 2 is the control input with the linear speed v ( t ) R , the angular speed is ω ( t ) v ( t ) tan ( δ ( t ) ) L R , where L R is the vehicle wheelbase, and δ ( t ) R is the wheel deflection angle. The vehicle model is presented in Figure 1.
Generally, the inner and outer ring structures are used to control the intelligent driving vehicle. The function of the outer loop is to generate the desired longitudinal speed v ( t ) and lateral speed ω ( t ) according to the desired trajectory, while the function of the inner loop is to obtain the real control input signal, i.e., the throttle opening (or brake opening) and steering wheel angle, by tracking the desired longitudinal speed v ( t ) and lateral speed ω ( t ) , respectively.
Note that if the desired longitudinal speed v ( t ) and lateral speed ω ( t ) change rapidly, the tracking of v ( t ) and ω ( t ) in the inner loop may experience hysteresis and inaccuracies. Under such conditions, it becomes challenging to achieve smooth braking at preset points for AVs. Since the variation in longitudinal velocity during braking typically exceeds that of lateral velocity, it can be assumed that ω ( t ) = 0 and θ ( t ) = θ ¯ with θ ¯ being a known parameter. Considering these characteristics, this paper proposes a neural network (NN)-based longitudinal kinematic model, as follows:
ξ ˙ ( t ) = x ˙ ( t ) y ˙ ( t ) v ˙ ( t ) = v ( t ) cos ( θ ¯ ) v ( t ) sin ( θ ¯ ) W T σ ( u tb ( t ) ) + ε ( t ) ,
where the extended dynamic model v ˙ ( t ) = W T σ ( u tb ( t ) ) + ε ( t ) is based on an NN to be designed; W R 3 × 1 is an unknown ideal matrix consisting of weight and bias; σ ( · ) is a vector of smooth activation functions; ε ( t ) is the external disturbance satisfying ε ( t ) ε ¯ and ε ˙ ( t ) ε ˜ with ε ¯ , and ε ˜ being known positive constants; u tb ( t ) [ v ( t ) u o ( t ) 1 ] T is the input signal of the NN, and u o ( t ) is the longitudinal speed control variable, where
u o ( t ) = u to ( t ) , if u to ( t ) ( 0 , 1 ] , u bo ( t ) = 0 u bo ( t ) , if u bo ( t ) ( 0 , 9 ] , u to ( t ) = 0 0 , if u to ( t ) = 0 , u bo ( t ) = 0
with u to ( t ) and u bo ( t ) being the throttle opening and brake opening, respectively.
As stated in [34], it is assumed that there exists a known constant, ϵ W , such that W ϵ W . In addition, according to [35], the gradient of σ ( · ) is assumed to satisfy σ ( · ) ϵ σ with ϵ σ being a positive constant. For simplicity, we denote σ ( t ) σ ( u tb ( t ) ) .
Note that the AV is driven by the control input u ( t ) u o ( t ) in the form of a segment constant in each sampling interval [ i T , ( i + 1 ) T ) , i Z 0 , with  sampling period T (in seconds). Based on (2), a discrete-time AV kinematic model is given as follows:
ξ ( k + 1 ) = x ( k ) y ( k ) v ( k ) + T v ( k ) cos ( θ ¯ ) T v ( k ) sin ( θ ¯ ) T W T σ ( k ) + T ε ( k ) .
In the following, the control framework diagram of this paper is presented in Figure 2. In Figure 2, the control framework is divided into offline and online parts. The offline part is used to collect data on the speed and acceleration of AVs under different throttle and brake openings and to train NN parameters using the collected data to prepare for subsequent system control. The online part is a closed-loop control loop, including the RHESO and MPC that need to be designed later.

2.2. Estimation of the Unknown NN Parameter

The ideal NN parameter W is an unknown matrix that exists in theory, but an estimated matrix W ^ * that is close enough to W can be obtained through estimation methods. Therefore, an adaptive estimator is designed to obtain the estimation W ^ * .
Consider the following adaptive estimator:
v ^ ( k + 1 ) = v ( k ) + T W ^ T ( k ) σ ( k ) ,
where v ^ ( k ) is the estimation of v ( k ) ; W ^ ( k ) is the estimation of W at time instant k. Subtracting v ( k + 1 ) from v ^ ( k + 1 ) leads to the following:
v ^ ( k + 1 ) v ( k + 1 ) = T ( W ^ T ( k ) W T ) σ ( k ) T ε ( k ) ,
which yields the following:
1 T ( v ^ ( k + 1 ) v ( k + 1 ) ) = ( W ^ T ( k ) W T ) σ ( k ) ε ( k ) .
Denote
e ( k + 1 ) 1 T ( v ^ ( k + 1 ) v ( k + 1 ) ) , W ˜ ( k ) W ^ ( k ) W ,
then, (4) can be rewritten as follows:
e ( k + 1 ) = W ˜ T ( k ) σ ( k ) ε ( k ) .
To minimize E ( k ) = 1 2 e T ( k + 1 ) e ( k + 1 ) , using the Levenberg–Marquardt algorithm [36], the tuning law of W ^ ( k ) is designed as follows:
W ^ ( k + 1 ) = W ^ ( k ) Λ 1 + σ T ( k ) σ ( k ) E ( k ) W ^ ( k ) T = W ^ ( k ) Λ σ ( k ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) ,
where Λ is a positive constant to be designed.
In the following, a theorem is given to demonstrate that the proposed tuning law (5) can make the estimation W ^ ( k ) converge to a neighborhood of ideal value W.
Theorem 1. 
If σ ( k ) in (5) is persistently exciting and 0 < Λ < 1 holds, then W ˜ ( k ) is uniformly bounded to the set W where W { W ˜ | W ˜ 2 κ ¯ } with the following:
κ ¯ 1 κ 1 + 1 1 Λ 2 ε ¯ 2 , κ λ ̲ ( σ ( [ 0 0 1 ] T ) σ T ( [ 0 0 1 ] T ) ) .
Proof. 
We select a Lyapunov candidate function, as follows:
V ( W ˜ ( k ) ) = W ˜ T ( k ) Λ 1 W ˜ ( k ) ,
thus, we have the following:
V ( W ˜ ( k + 1 ) ) V ( W ˜ ( k ) ) = e ( k + 1 ) σ T ( k ) W ˜ ( k ) 1 + σ T ( k ) σ ( k ) W ˜ T ( k ) σ ( k ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) + e ( k + 1 ) σ T ( k ) Λ σ ( k ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) 2 = 2 e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) e ( k + 1 ) ε T ( k ) 1 + σ T ( k ) σ ( k ) ε ( k ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) + σ T ( k ) Λ σ ( k ) e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) 2 e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) + ε ( k ) ε T ( k ) 1 + σ T ( k ) σ ( k ) + σ T ( k ) Λ σ ( k ) e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) 2 1 σ T ( k ) Λ σ ( k ) 1 + σ T ( k ) σ ( k ) e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) + ε ( k ) ε T ( k ) 1 + σ T ( k ) σ ( k ) ( 1 Λ ) e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) + ε ( k ) ε T ( k ) 1 + σ T ( k ) σ ( k ) .
Select a proper Λ , such that 0 < Λ < 1 . By letting V ( W ˜ ( k + 1 ) ) V ( W ˜ ( k ) ) < 0 , we can obtain the following:
ε ( k ) ε T ( k ) 1 + σ T ( k ) σ ( k ) < ( 1 Λ ) e ( k + 1 ) e T ( k + 1 ) 1 + σ T ( k ) σ ( k ) ,
which indicates the following:
ε T ( k ) ε ( k ) < ( 1 Λ ) e T ( k + 1 ) e ( k + 1 ) .
This means that V ( W ˜ ( k ) ) will remain convergent until condition (7) is not satisfied. Due to e ( k + 1 ) R and ε ( k ) R , we have the following:
1 1 Λ ε 2 ( k ) < ( W ˜ T ( k ) σ ( k ) ε ( k ) ) 2 ,
which yields the following:
1 + 1 1 Λ 2 ε 2 ( k ) < W ˜ T ( k ) σ ( k ) σ T ( k ) W ˜ ( k ) .
To ensure that inequality (8) always holds, we require the following:
1 + 1 1 Λ 2 ε ¯ 2 < λ ̲ ( σ ( k ) σ T ( k ) ) W ˜ T ( k ) W ˜ ( k ) .
Considering σ ( k ) = σ ( u tb ( k ) ) , we have the following:
κ λ ̲ ( σ ( k ) σ T ( k ) ) = λ ̲ ( σ ( [ 0 0 1 ] T ) σ T ( [ 0 0 1 ] T ) ) .
Then, (9) can be rewritten as follows:
1 κ 1 + 1 1 Λ 2 ε ¯ 2 < W ˜ T ( k ) W ˜ ( k ) .
Denote W { W ˜ ( k ) | W ˜ ( k ) 2 κ ¯ } with κ ¯ 1 κ 1 + 1 1 Λ 2 ε ¯ 2 . Based on the fact that σ ( k ) is persistently exciting [37], it can be obtained that W ˜ ( k ) would converge to the set W . Therefore, W ˜ ( k ) is uniformly bounded. This completes the proof.    □
From Theorem 1, it can be obtained that W ˜ ( k ) will finally converge to the set W . Therefore, if there exists W ˜ ( k s ) W for k s Z 0 , then W ˜ ( k s + 1 ) W holds. Hence, it is easy to obtain the following:
W ˜ ( k s + 1 ) W ˜ ( k s )   = W ^ ( k s + 1 ) W ^ ( k s )     W ˜ ( k s + 1 )   +   W ˜ ( k s )     2 κ ¯ .
Inequality (10) can be used as a termination condition of the tuning law in (5). When (10) is satisfied, W ^ * = W ^ ( k s ) can be obtained.
In the following, the process of approximating AVs’ dynamics using a NN is given in Algorithm 1.
Algorithm 1 first collects data on changes in AV velocity and acceleration under different throttle and brake opening conditions. Secondly, these data are divided into inputs and outputs to train the NN. Thirdly, the NN structure, loss function, and activation function are designed. Finally, the NN is trained until its parameters converge. At this stage, W ^ * is obtained and the approximate dynamic model of the AV can be obtained.
Remark 1. 
The process of training W ^ ( k ) involves first tuning the throttle and brake, that is, recording the impact of the throttle and brake on speed. Then, the collected data are adopted for training to obtain W ^ * . Therefore, the process of training the NN parameter is completed offline.

2.3. Linearization Model and Constraints of the AV

Note that (3) can be rewritten in the following linear form:
ξ ( k + 1 ) = A ξ ( k ) + B u ( k ) + D d ( k ) ,
where
A = 1 0 T cos ( θ ¯ ) 0 1 T sin ( θ ¯ ) 0 0 1 + T W ^ 1 * a σ , B = 0 0 T W ^ 2 * b σ = D , d ( k ) = 1 W ^ 2 * b σ W ^ * T σ ( k ) e ( k + 1 ) W ^ 1 * a σ v ( k ) u ( k ) ,
where W ^ i * ( i = 1 , 2 , 3 ) indicates the i-th element of W ^ * ; and a σ and b σ represent the linearization coefficients of σ ( k ) with u o ( k ) = 0 . Assume that | d ( k ) | d ¯ with d ¯ > 0 being a known parameter.
Algorithm 1 Algorithm for approximating AVs’ dynamics using a NN
Brake data collection:
1. Divide V max into n equal parts for AVs with the maximum speed of V max . Denote a i = i V max m , where ( i = 1 , 2 , , m ) is the step size for each increase in velocity. Divide the maximum brake opening, 9, equally into n parts, and the size of each brake opening is 9 n . Denote b j = j 9 n , where ( j = 1 , 2 , , n ) denotes the step size for each increase in the brake opening value. Set i = 1 and j = 1 . Go to step 2.
2. Set the AV to travel at velocity a i . Then, apply the AV brake with the brake opening of b j . Go to step 3.
3. Collect the data of the velocity v ( a i , b j ) and acceleration a ( a i , b j ) from the start of braking until the AV comes to a complete stop. Go to step 4.
4. If i + 1 m , i = i + 1 . Go to step 2. Otherwise, go to step 5.
5. If j + 1 9 n , set i = 1 and j = j + 1 . Go to step 2. Otherwise, go to step 6.
Throttle data collection:
6. Divide V max into p equal parts for AVs with the maximum speed of V max . Denote c i = i V max p , ( i = 0 , 1 , 2 , , p ) as the step size for each increase in velocity. Divide the maximum throttle opening 1 equally into q parts, and the size of each brake opening is 1 q . Denote d j = j 1 q , ( j = 1 , 2 , , q ) as the step size for each increase in the brake opening value. Set i = 0 and j = 1 . Go to step 7.
7. Set the AV to travel at a fixed velocity c i and accelerate at a throttle opening of d j . Go to step 8.
8. Collect the data of the velocity v ( c i , d j ) and acceleration a ( c i , d j ) from the beginning of acceleration until the AV reaches V max . Go to 9.
9. If i + 1 p , i = i + 1 . Go to step 7. Otherwise, go to step 10.
10. If j + 1 1 q , set i = 1 and j = j + 1 . Go to step 7. Otherwise, go to 11.
Training and testing the NN:
11. Set the input of the NN as b j and v ( a i , b j ) , with the output being a ( a i , b j ) . Alternatively, the input of NN can be set as c i and v ( c i , d j ) , with the output being a ( c i , d j ) . Go to step 12.
12. Extract  70 % of these data as training data, and the remaining data as test data for backup. Go to step 13.
13. Define the NN structure, loss function, and activation function σ ( · ) . Using the training data obtained in step 12, the NN is trained until the network parameters converge. Go to step 13.
14. Using the NN obtained in step 12. If the loss meets the requirements, it indicates successful training. Terminate the algorithm and export the NN parameters W ^ * . If the loss does not meet the requirements, reset the training parameters and retrain. Go to step 13.
Considering the driving safety of the AV, its state and control input are subject to the following constraints:
ξ ( k ) X { ξ ( k ) : h ˜ i , min , ξ ξ i ( k ) h ˜ i , max , ξ } ,
u ( k ) U { u ( k ) : h ˜ min , u u ( k ) h ˜ max , u } ,
where ξ i ( k ) ( i = 1 , 2 , 3 ) indicates the i-th element of ξ ( k ) ; h ˜ i , min , ξ and h ˜ i , max , ξ represent the given minimum and maximum values of ξ i ( k ) , respectively; and h ˜ min , u and h ˜ max , u denote the given minimum and maximum values of u ( k ) , respectively. Then, (12) and (13) can be rewritten as follows:
ξ ( k ) X { ξ ( k ) : b ξ T ξ ( k ) h ξ } ,
u ( k ) U { u ( k ) : b u T u ( k ) h u } ,
where b ξ , b u , h ξ , and h u are matrices obtained from (12) and (13).

2.4. The Design of RHESO

In this subsection, a RHESO is designed to accurately estimate the unknown disturbance d ( k ) , as follows:
m ( k + 1 ) = ( I L d D ) ( m ( k ) + L d ξ ( k ) ) L d ( A ξ ( k ) + B u ( k ) ) + b ( k ) ,
d ^ ( k ) = m ( k ) + L d ξ ( k ) ,
where m ( k ) R is an intermediate variable; d ^ ( k ) R is an estimation of d ( k ) ; L d R 1 × 3 is an observer gain; and b ( k ) R is a decision variable to be designed. As a result, both L d and b ( k ) need to be designed subsequently.
Furthermore, system (11) can be extended to the following system:
ξ ( k + 1 ) = A ξ ( k ) + B u ( k ) + D d ( k ) ,
d ( k + 1 ) = d ( k ) + h ( k ) ,
where h ( k ) R is an unknown bounded nonlinear term satisfying | h ( k ) | H , which is equivalent to b h T h ( k ) h h with b h = [ 1 1 ] and h h = [ H H ] T .
Considering e d ( k ) = d ( k ) d ^ ( k ) , we obtain the following:
e d ( k + 1 ) = ( I L d D ) e d ( k ) b ( k ) + h ( k ) .
To restrict the transient performance of the RHESO (16a) and (16b), the following constraint of e d ( k ) is given as follows:
e d ( k ) E { e d ( k ) : h ˜ e d e d ( k ) h ˜ e d }
where h ˜ e d > 0 is a known parameter. Then, (19) can be rewritten as follows:
e d ( k ) E = { e d ( k ) : b e d T e d ( k ) h e d }
where b e d = [ 1 1 ] and h e d = [ h ˜ e d h ˜ e d ] T .
Remark 2. 
Different from the traditional ESO, the RHESO proposed in (16a) and (16b) allows us to obtain the optimal estimation by optimizing the decision variable b ( k ) . Furthermore, it is possible to avoid significant overshoot caused by the estimation error e d ( k ) in terms of the constraint (19), thereby ensuring that system (11) does not experience performance degradation due to a large estimation error e d ( k ) before convergence.

2.5. Control Input and System Reconstruction

In this subsection, the control input is designed for system (11). Firstly, the control input of system (11) is designed as follows:
u ( k ) = K ξ ( k ) + c ( k ) V d ^ ( k ) , ξ ( k ) X T K ξ ( k ) V d ^ ( k ) , ξ ( k ) X T
where V [ 0 1 ] T ; X T is a terminal set to be designed; K R 2 × 3 is the feedback gain to be designed; and c ( k ) R 2 is the decision variable to be designed.
Secondly, based on (11), (18) and (21), the extended state system is obtained as follows:
η ˇ ( k + 1 ) = A ˜ η η ˇ ( k ) + B ˜ η u η ( k ) + d η ( k ) + ϖ ( k ) ,
where
u η ( k ) = K η η ˇ ( k ) + c η ( k ) [ d ^ ( k ) 0 ] T , η ˇ ( 1 ) ξ ( 1 ) 0 , η ˇ ( k ) ξ ( k ) e d ( k 1 ) , A ˜ η A 0 0 I , B ˜ η B 0 0 I , d η ( k ) D d ( k ) 0 , K η K 0 0 L d D ,
ϖ ( k ) 0 h ( k 1 ) , k Z [ 2 , + ) ,
c η ( k ) = c ( k ) b ( k 1 ) , k Z [ 2 , + ) .
Then, we obtain the following:
F η ˇ ( k + 1 ) = ( A ˜ η + B ˜ η K η ) η ˇ ( k ) + B ˜ η c η ( k ) + ϖ ( k ) = Φ ˜ η η ˇ ( k ) + B ˜ η c η ( k ) + ϖ ( k )
where
F I D 0 I , Φ ˜ η Φ 0 0 Φ e d , Φ A + B K , Φ e d I L d D .
Since F is a full rank matrix, (26) is converted as follows:
η ˇ ( k + 1 ) = ( F 1 A ˜ η + F 1 B ˜ η K η ) η ˇ ( k ) + F 1 B ˜ η c η ( k ) + F 1 ϖ ( k ) = ( A η + B η K η ) η ˇ ( k ) + B η c η ( k ) + F 1 ϖ ( k ) = Φ η η ˇ ( k ) + B η c η ( k ) + F 1 ϖ ( k ) ,
where Φ η = A η + B η K η , A η = F 1 A ˜ η and B η = F 1 B ˜ η . According to (14), (15) and (19), the constraints of η ˇ ( k ) and c η ( k ) are obtained as follows:
η ˇ ( k ) { η ˇ ( k ) : b η T η ˇ ( k ) h η } , c η ( k ) { c η ( k ) : b ˜ u T η ˇ ( k ) + b c η T c η ( k ) h c η + h d ( k ) }
where
b η b ξ 0 0 b e d , b ˜ u K T b u 0 0 0 , b c η b u 0 0 0 , h η h ξ h e d , h c η h u 0 , h d ( k ) b u T d ^ ( k ) 0 .
Based on | h ( k ) | H , the constraint of ϖ ( k ) is obtained as follows:
ϖ ( k ) W { ϖ ( k ) : b ϖ T ϖ ( k ) h ϖ } ,
where
b ϖ 0 0 0 b h , h ϖ 0 h ¯ h , h ¯ h H H .
The nominal system of (27) is denoted as follows:
η ( k + 1 ) = Φ η η ( k ) + B η c η ( k ) ,
where η ( k ) is the nominal state of η ˇ ( k ) .
Denoting e η ( k ) η ˇ ( k ) η ( k ) , we have the following:
e η ( k + 1 ) = Φ η e η ( k ) + F 1 ϖ ( k ) .
Then, constraints of η ( k ) and c η ( k ) are obtained as follows:
η ( k ) M { η ( k ) : b η T η ( k ) h η b η T e η ( k ) } , c η ( k ) N { c η ( k ) : b ˜ u T η ( k ) + b c η T c η ( k ) h c η ( b ˜ u T e η ( k ) h d ( k ) ) } .

2.6. The MPC Scheme

To achieve smooth braking at preset points in AVs under constraints, an MPC scheme is proposed in this subsection. Firstly, the cost function of system (27) is denoted as follows:
J ( ȷ ^ ( k ) , c ^ η ( k ) ) s = 0 N 1 η ^ ( k + s | k ) Q 2 + K η η ^ ( k + s | k ) + c ^ η ( k + s | k ) R 2 + η ^ ( k + N | k ) P 2 ,
where η ^ ( k + s | k ) is the predictive state with η ^ ( k | k ) = η ˇ ( k ) ; c ^ η ( k + s | k ) is the predictive decision variable; ȷ ^ ( k ) [ η ^ T ( k | k ) η ^ T ( k + 1 | k ) η ^ T ( k + N | k ) ] T is the predictive state sequence; c ^ η ( k ) [ c ^ η T ( k | k ) c ^ η T ( k + 1 | k ) c ^ η T ( k + N 1 | k ) ] T is the predictive decision variable sequence with c ^ η ( k | k ) c η ( k ) ; Q > 0 and R > 0 are the weighting matrices with appropriate dimensions; P > 0 is the terminal penalty matrix to be designed; and N is the predictive horizon. Then, the optimization problem of MPC is denoted as follows:
Prob 1:
( ȷ ^ * ( k ) , c ^ η * ( k ) ) arg min J ( ȷ ^ ( k ) , c ^ η ( k ) )
subject to
η ^ ( k + s + 1 | k ) = Φ η η ^ ( k + s | k ) + B η c ^ η ( k + s | k ) ,
b η T η ^ ( k + s | k ) h η ϕ ( k + s | k ) , s Z [ 0 , N 1 ] ,
b ˜ u T η ^ ( k + s | k ) + b c η T c ^ η ( k + s | k ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + s | k ) , s Z [ 0 , N 1 ] ,
η ^ ( k + N | k ) X T ( k + N | k ) ,
where
ϕ ( k + s | k ) j = 0 s 1 max ϖ ( k + s 1 j | k ) W b η T Φ η j F 1 ϖ ( k + s 1 j | k ) , ϕ ( k | k ) 0 , χ ( k + s | k ) j = 0 s 1 max ϖ ( k + s 1 j | k ) W b ˜ u T Φ η j F 1 ϖ ( k + s 1 j | k ) , χ ( k | k ) 0 , e η ( k + s + 1 | k ) = Φ η e η ( k + s | k ) + F 1 ϖ ( k + s | k ) , e η ( k | k ) = 0 , X T ( k + N | k ) X ^ ( k + N | k ) C ^ ( k + N | k ) , X ^ ( k + N | k ) { η ^ ( k + N | k ) : b η T η ^ ( k + N | k ) h η ϕ ( k + N | k ) } , C ^ ( k + N | k ) { η ^ ( k + N | k ) : b ˜ u T η ^ ( k + N | k ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + N | k ) } , ȷ ^ * ( k ) [ η ^ * T ( k | k ) η ^ * T ( k + 1 | k ) η ^ * T ( k + N | k ) ] T , c ^ η * ( k ) [ c ^ * T ( k | k ) c ^ * T ( k + 1 | k ) c ^ * T ( k + N 1 | k ) ] T .
Remark 3. 
The polynomial complexity of the proposed RHESO-based MPC consists of three parts: (1) the polynomial complexity of the RHESO; (2) the polynomial complexity of Prob 1; and (3) the polynomial complexity of the input signal. The polynomial complexity of the RHESO is obtained as follows:
O ( I L d D ) ( m ( k ) + L d ξ ( k ) ) L d ( A ξ ( k ) + B u ( k ) ) + b ( k ) + O m ( k ) + L d ξ ( k ) = O 24
The polynomial complexity of Prob 1 is obtained as follows:
O s = 0 N 1 η ^ ( k + s | k ) Q 2 + K η η ^ ( k + s | k ) + c ^ η ( k + s | k ) R 2 + η ^ ( k + N | k ) P 2 = O 52 N + 31
The polynomial complexity of the input signal is obtained as follows:
O K ξ ( k ) + c ( k ) V d ^ ( k ) = O 13
Then the polynomial complexity of the proposed RHESO-based MPC is obtained as follows:
O 24 + O 52 N + 31 + O 13 = O 52 N + 31
Definition 1 
([38]). If for all Ξ ¯ 1 > 0 , there exists a constant Ξ ¯ 2 ( Ξ ¯ 1 ) such that we have the following:
η ˇ ( 0 ) Ξ ¯ 1 η ˇ ( k ) Ξ ¯ 2 ( Ξ ¯ 1 )
holds for all η ˇ ( 0 ) R n and k Z 0 , then system (27) is said to be uniformly bounded.

3. Results

In this part, a lemma is presented firstly to demonstrate the feasibility of Prob 1, and then a theorem is presented to verify the uniformly bounded stability of the closed-loop system.
Lemma 1. 
If a feasible solution exists for Prob 1 at time instant k, then it also has at least one feasible solution at time instant k + 1 .
Proof. 
Denote a candidate decision variable sequence c ˜ η ( k + 1 ) as
c ˜ η ( k + 1 ) c ˜ η ( k + 1 | k + 1 ) c ˜ η ( k + N | k + 1 ) = c ^ η * ( k + 1 | k ) c ^ η * ( k + N 1 | k ) 0 ,
where c ˜ η ( k + 1 + s | k + 1 ) is a candidate decision variable for s Z [ 0 , N 1 ] . Then, there exists
η ^ ( k + 1 | k + 1 ) = Φ η η ^ ( k | k ) + B η c ^ η * ( k | k ) + F 1 ϖ ( k ) .
Considering the predictive dynamics of η ˇ ( k + 1 ) using c ˜ η ( k + 1 ) at time instant k + 1 , we have the following:
η ˜ ( k + 1 + s + 1 | k + 1 ) = Φ η η ˜ ( k + 1 + s | k + 1 ) + B η c ˜ η ( k + 1 + s | k + 1 ) + F 1 ϖ ( k + 1 + s | k + 1 ) ,
where η ˜ ( k + 1 + s + 1 | k + 1 ) is the feasible state for s Z [ 0 , N ] .
Combining (29), (33) and (34), we have the following:
η ˜ ( k + s | k + 1 ) = η ^ * ( k + s | k ) + Φ η s 1 F 1 ϖ ( k ) .
Denote
ȷ ˜ ( k + 1 ) = η ˜ ( k + 1 | k + 1 ) η ˜ ( k + 1 + N | k + 1 )
as the feasible predictive state sequence. To guarantee that Prob 1 is feasible at time instant k + 1 , the following conditions should be satisfied:
b η T η ˜ ( k + 1 + s | k + 1 ) h η ϕ ( k + 1 + s | k + 1 ) , s Z [ 0 , N ]
b ˜ u T η ˜ ( k + 1 + s | k + 1 ) +   b c η T c ˜ η ( k + 1 + s | k + 1 ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + 1 + s | k + 1 ) , s Z [ 0 , N 1 ]
η ˜ ( k + 1 + N | k + 1 ) X T ( k + 1 + N | k + 1 )
In the following, conditions (36)–(38) are proved one by one. Firstly, the proof of condition (36) is given. Assume that Prob 1 is feasible at time instant k. Then, the constraint (30) holds, which indicates the following:
b η T η ^ * ( k + 1 + s | k ) h η ϕ ( k + 1 + s | k )
for s Z [ 0 , N 2 ] . Combining (35) and (39), the following exists:
b η T ( η ˜ ( k + 1 + s | k + 1 ) Φ η s F 1 ϖ ( k ) ) h η ϕ ( k + 1 + s | k ) ,
which yields
b η T η ˜ ( k + 1 + s | k + 1 ) h η ϕ ( k + 1 + s | k ) + b η T Φ η s F 1 ϖ ( k ) = h η j = 0 s 1 max ϖ ( k + s 1 j | k ) W b η T Φ η j F 1 ϖ ( k + s 1 j | k ) + b η T Φ η s F 1 ϖ ( k ) max ϖ ( k + s j | k ) W ϖ ( k + s j | k ) h η j = 0 s 1 max ϖ ( k + s 1 j | k ) W b η T Φ η j F 1 ϖ ( k + s 1 j | k ) = h η ϕ ( k + 1 + s | k + 1 )
for s Z [ 0 , N ] . Therefore, condition (36) is satisfied for s Z [ 0 , N ] .
Secondly, the proof of condition (37) is given. By using the candidate variable sequence c ˜ η ( k + 1 ) , we have that c ˜ η ( k + 1 + s | k + 1 ) = c ^ η ( k + 1 + s | k ) . Based on (31) and (35), we have the following:
b ˜ u T ( η ˜ ( k + 1 + s | k + 1 ) Φ η s F 1 ϖ ( k ) ) + b c η T c ^ η ( k + 1 + s | k ) h c η | h e d + d ¯ | [ b u 0 ] T χ ( k + 1 + s | k ) ,
which indicates the following:
b ˜ u T η ˜ ( k + 1 + s | k + 1 ) + b c η T c ˜ η ( k + 1 + s | k + 1 ) h c η | h e d + d ¯ | [ b u 0 ] T χ ( k + 1 + s | k ) + b ˜ u T Φ η s F 1 ϖ ( k ) = h c η | h ˜ e d + d ¯ | [ b u 0 ] T j = 0 s 1 max ϖ ( k + s 1 j | k ) W b ˜ u T Φ η j F 1 ϖ ( k + s 1 j | k ) + b ˜ u T Φ η s F 1 ϖ ( k ) max ϖ ( k + s j | k ) W ϖ ( k + s j | k ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T j = 0 s 1 max ϖ ( k + s 1 j | k ) W b ˜ u T Φ η j F 1 ϖ ( k + s 1 j | k ) = h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + 1 + s | k + 1 )
for s Z [ 0 , N 2 ] . For s = N 1 , there exists c ˜ η ( k + N | k + 1 ) = 0 . Based on (32), (35) and c ˜ η ( k + N | k + 1 ) = 0 , it is easy to obtain the following:
b ˜ u T ( η ˜ ( k + N | k + 1 ) Φ η N 1 F 1 ϖ ( k ) ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + N | k ) ,
which means that we have the following:
b ˜ u T η ˜ ( k + N | k + 1 ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + N | k ) + b ˜ u T Φ η N 1 F 1 ϖ ( k ) = h c η | h ˜ e d + d ¯ | [ b u 0 ] T j = 0 N 2 max ϖ ( k + N 2 j | k ) W b ˜ u T Φ η j F 1 ϖ ( k + N 2 j | k ) b ˜ u T Φ η N 1 F 1 max ϖ ( k + N 1 j | k ) W ϖ ( k + N 1 j | k ) + b ˜ u T Φ η N 1 F 1 ϖ ( k ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T j = 0 N 2 max ϖ ( k + N 2 j | k ) W b ˜ u T Φ η j F 1 ϖ ( k + N 2 j | k ) = h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + N | k + 1 ) .
Therefore, condition (37) is satisfied for s Z [ 0 , N 1 ] .
Thirdly, the proof of condition (38) is proved, as X T ( k + 1 + N | k + 1 ) X ^ ( k + 1 + N | k ) C ^ ( k + 1 + N | k + 1 ) . In terms of (36), it follows that η ˜ ( k + 1 + N | k + 1 ) X ^ ( k + 1 + N | k ) . Furthermore, based on (32), it is easy to obtain that η ˜ ( k + 1 + N | k + 1 ) C ^ ( k + 1 + N | k ) . Therefore, η ˜ ( k + 1 + N | k + 1 ) X T ( k + 1 + N | k + 1 ) is satisfied, such that condition (38) holds.
As a result, if Prob 1 is feasible at time instant k, then Prob 1 is also feasible at time instant k + 1 . This completes the proof. □
Theorem 2. 
If Ξ > 0 and Υ η exist, which make the linear matrix inequality (LMI),
Ξ Ξ A η T + Υ η B η T Υ η Ξ A η Ξ T + B η Υ η T Ξ 0 0 Υ η T 0 R 1 0 Ξ T 0 0 Q 1 < 0
then system (22) is uniformly bounded and K η = Υ η T Ξ 1 with Ξ 1 = P . Furthermore, the state of system (22) converges to the set S { η ˇ ( k ) : η ˇ ( k ) Q 2 σ ¯ } , where
σ ¯ s = 1 N 1 λ ¯ 2 ( Q ) Ω 2 ( s ) + 2 λ ¯ 2 ( Q ) Ω ( s ) F ( s ) + λ ¯ 2 ( R ) Ξ 2 ( s ) + 2 λ ¯ 2 ( R ) Ξ ( s ) Ψ ( s ) + λ ¯ 2 ( P ) Ω 2 ( N ) + λ ¯ ( P ) Ω ( N ) π ( N ) ,
with
Ω ( s ) max ϖ ( k ) W Φ η s 1 F 1 ϖ ( k ) , F ( s ) h η ϕ ( k + s | k ) , Ξ ( s ) max ϖ ( k ) W K η Φ η s 1 F 1 ϖ ( k ) , Ψ ( s ) h c η | h e d + d ¯ | [ b u 0 ] T χ ( k + s | k ) .
Proof. 
Considering the optimal solution of Prob 1 is obtained at time instant k, it can be concluded that Prob 1 has at least one feasible solution at time instant k + 1 . Denote J ( ȷ ˜ ( k + 1 ) , c ˜ η ( k + 1 ) ) as the feasible cost of Prob 1 at time instant k + 1 with ȷ ˜ ( k + 1 ) and c ˜ η ( k + 1 ) given in Lemma 1.
The optimal cost of Prob 1 at time instant k is denoted as J ( ȷ ^ * ( k ) , c ^ η * ( k ) ) . Choose Δ J ( k ) J ( ȷ ˜ ( k + 1 ) , c ˜ η ( k + 1 ) ) J ( ȷ ^ * ( k ) , c ^ η * ( k ) ) . Split Δ J ( k ) = Δ 1 + Δ 2 + Δ 3 with
Δ 1 = s = 1 N 1 η ˜ ( k + s | k + 1 ) Q 2 η ^ * ( k + s | k ) Q 2 + K η η ˜ ( k + s | k + 1 ) + c ˜ η ( k + s | k + 1 ) R 2 K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) R 2 , Δ 2 = η ˜ ( k + T s | k + 1 ) Q 2 + K η η ˜ ( k + T s | k + 1 ) R 2 + η ˜ ( k + 1 + T s | k + 1 ) P 2 η ^ * ( k + T s | k ) P 2 , Δ 3 = η ^ * ( k | k ) Q 2 K η η ^ * ( k | k ) + c ^ η * ( k | k ) Q 2 .
For Δ 1 , we have the following:
Δ 1 = s = 1 N 1 η ˜ ( k + s | k + 1 ) Q 2 η ^ * ( k + s | k ) Q 2 + K η ( η ^ * ( k + s | k ) + Φ η s 1 F 1 ϖ ( k ) ) + c ˜ η ( k + s | k + 1 ) R 2 K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) R 2 = s = 1 N 1 η ˜ ( k + s | k + 1 ) Q 2 η ^ * ( k + s | k ) Q 2 + K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) + K η Φ η s 1 F 1 ϖ ( k ) R 2 K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) R 2 s = 1 N 1 η ˜ ( k + s | k + 1 ) Q η ^ * ( k + s | k ) Q η ˜ ( k + s | k + 1 ) Q + η ^ * ( k + s | k ) Q + K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) + K η Φ η s 1 F 1 ϖ ( k ) R K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) R × K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) + K η Φ η s 1 F 1 ϖ ( k ) R + K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) R s = 1 N 1 λ ¯ ( Q ) Φ η s 1 F 1 ϖ ( k ) × 2 λ ¯ ( Q ) η ^ * ( k + s | k ) + λ ¯ ( Q ) Φ η s 1 F 1 ϖ ( k ) + λ ¯ ( R ) K η Φ η s 1 F 1 ϖ ( k ) 2 λ ¯ ( R ) K η η ^ * ( k + s | k ) + c ^ η * ( k + s | k ) + λ ¯ ( R ) K η Φ η s 1 F 1 ϖ ( k ) s = 1 N 1 λ ¯ ( Q ) Φ η s 1 F 1 ϖ ( k ) 2 λ ¯ ( Q ) h η ϕ ( k + s | k ) + λ ¯ ( Q ) Φ η s 1 F 1 ϖ ( k ) + λ ¯ ( R ) K η Φ η s 1 F 1 ϖ ( k ) 2 λ ¯ ( R ) h c η [ ( h ˜ e d + d ¯ ) 0 ] T χ ( k + s | k ) + λ ¯ ( R ) K η Φ η s 1 F 1 ϖ ( k ) s = 1 N 1 λ ¯ 2 ( Q ) Ω 2 ( s ) + 2 λ ¯ 2 ( Q ) Ω ( s ) F ( s ) + λ ¯ 2 ( R ) Ξ 2 ( s ) + 2 λ ¯ 2 ( R ) Ξ ( s ) Ψ ( s ) ,
where
Ω ( s ) max ϖ ( k ) W Φ η s 1 F 1 ϖ ( k ) , F ( s ) h η ϕ ( k + s | k ) , Ξ ( s ) max ϖ ( k ) W K η Φ η s 1 F 1 ϖ ( k ) , Ψ ( s ) h c η | h ˜ e d + d ¯ | [ b u 0 ] T χ ( k + s | k ) .
For Δ 2 , the following can be obtained:
Δ 2 = η ˜ ( k + N | k + 1 ) Q 2 + K η η ˜ ( k + N | k + 1 ) R 2 + η ˜ ( k + 1 + N | k + 1 ) P 2 η ^ * ( k + N | k ) P 2 + η ˜ ( k + N | k + 1 ) P 2 η ˜ ( k + N | k + 1 ) P 2 ( η ˜ ( k + N | k + 1 ) P 2 η ^ * ( k + N | k ) P 2 ) η ˜ ( k + N | k + 1 ) Q ¯ 2 η ˜ ( k + N | k + 1 ) η ^ * ( k + N | k ) P ( η ˜ ( k + N | k + 1 ) P + η ^ * ( k + N | k ) P ) η ˜ ( k + N | k + 1 ) Q ¯ 2 η ˜ ( k + N | k + 1 ) Q ¯ 2 + λ ¯ ( P ) Φ η N 1 F 1 ϖ ( k ) × 2 λ ¯ ( P ) η ^ * ( k + N | k ) + λ ¯ ( P ) Φ η N 1 F 1 ϖ ( k ) ,
where Q ¯ = Φ η T P Φ η + Q + K η T R K η P and π ( N ) min { F ( N ) , Ψ ( N ) } . Letting Q ¯ < 0 , we obtain the following:
Φ η T P Φ η + Q + K η T R K η P < 0
which can be equivalently converted to LMI (40), such that K and L d can be obtained by solving LMI (40). Thus, η ˜ ( k + N | k + 1 ) Q ¯ 2 > 0 , and we have the following:
Δ 2 λ ¯ 2 ( P ) Ω 2 ( N ) + λ ¯ ( P ) Ω ( N ) π ( N ) .
For Δ 3 , there exists Δ 3 η ^ * ( k | k ) Q 2 . Combining Δ 1 , Δ 2 , and Δ 3 , we obtain the following:
Δ J ( k ) η ^ * ( k | k ) Q 2 + σ ¯ .
with
σ ¯ s = 1 N 1 λ ¯ 2 ( Q ) Ω 2 ( s ) + 2 λ ¯ 2 ( Q ) Ω ( s ) F ( s ) + λ ¯ 2 ( R ) Ξ 2 ( s ) + 2 λ ¯ 2 ( R ) Ξ ( s ) Ψ ( s ) + λ ¯ 2 ( P ) Ω 2 ( N ) + λ ¯ ( P ) Ω ( N ) π ( N ) .
According to the optimality principle, it is easy to obtain the following:
J ( ȷ ^ * ( k + 1 ) , c ^ η * ( k + 1 ) ) J ( ȷ ^ * ( k ) , c ^ η * ( k ) ) J ( ȷ ˜ ( k + 1 ) , c ˜ η ( k + 1 ) ) J ( ȷ ^ * ( k ) , c ^ η * ( k ) ) η ^ * ( k | k ) Q 2 + σ ¯ = η ˇ ( k ) Q 2 + σ ¯ .
Therefore, J ( ȷ ^ * ( k ) , c ^ η * ( k ) ) can be used as the Lyapunov function. As a result, η ˇ ( k ) can converge to the set S as k + , suggesting that system (22) is uniformly bounded, such that the uniform boundedness of system (11) and error system (18) can also be guaranteed. □
Remark 4. 
Compared to existing smooth and precise braking methods [13,14,15], the RHESO-based MPC proposed in this paper has the following advantages: (1) A new neural network-based vehicle kinematic model is proposed, which simulates the dynamic characteristics of vehicles through neural networks to supplement the shortcomings of the kinematic model in accurately describing the throttle and brake control on the vehicle speed; (2) the proposed RHESO can estimate the unmodeled dynamics and potential disturbances in AV kinematic modeling, providing preliminary theoretical support for the AV to achieve smooth and accurate preset point braking; and (3) design an AV speed control method based on MPC, taking into account the inherent speed and control constraints of AV to enable smooth and accurate braking at preset points.

4. Numerical Example

An electric AV model provided by the CarSim software is employed to obtain the associated data of the throttle/brake on vehicle speed and acceleration. Therefore, the corresponding approximate dynamic model can be further obtained by using the proposed NN. Partial parameters of the electric AV used in this example are given in Table 2.
As this paper focuses on longitudinal control, the driving angle θ ¯ = 60 ° is selected in (2). The external disturbance is selected as ε ( t ) = sin ( t + 0.2 ) . Other simulation parameters are given in Table 3.
The activation function vector and the weight matrix for the NN are reconstructed as follows:
σ ( t ) = [ v ( t ) u o ( t ) 1 ] T , W ^ * = [ 0.21 1.58 1.09 ] T .
The matrices L d , K, and P are obtained by solving LMI (40) as follows:
L d = [ 0 0 6.1 ] , K = [ 1.4 2.6 3.6 ] , P = 2576.1 1376.5 18.3 0 1376.5 1003.7 33.9 0 18.3 33.9 43.9 0 0 0 0 20.9 .
The initial state of the electric AV is selected as follows:
ξ ( 0 ) = [ 15.1 26.2 8 ] T .
To display the effectiveness of the proposed RHESO-based MPC for achieving smooth braking at preset points in AVs, its numerical simulation result is compared with those obtained by the traditional ESO-based MPC, active disturbance rejection control (ADRC), and robust MPC (RMPC). The simulation results are displayed in Figure 3, Figure 4, Figure 5 and Figure 6. Figure 3 shows the position response of the electric AV using the proposed RHESO-based MPC, as well as traditional ESO-based MPC, ADRC, and RMPC. It can be clearly seen that the position components of the AV, i.e., x ( t ) and y ( t ) , enable it to converge well to the origin when the proposed RHESO-based MPC is employed. However, if the traditional ESO-based MPC, ADRC, and RMPC are adopted, the AV’s position components x ( t ) and y ( t ) cannot converge to the origin. Similarly, Figure 4 exhibits the velocity response of the electric AV obtained by the proposed RHESO-based MPC, traditional ESO-based MPC, ADRC, and RMPC. The simulation results in Figure 3 and Figure 4 reveal that the proposed RHESO-based MPC allows for achieving smooth braking at designated locations in the electric AV. On the contrary, smooth braking at designated positions in the electric AV is difficult to achieve if the traditional ESO-based MPC, ADRC, and RMPC are applied. Figure 5 plots the control input of the electric AV. Furthermore, the history of estimation errors obtained by the proposed RHESO and the traditional ESO are shown in Figure 6. Note that the error convergence speed of the traditional ESO is faster than that of the proposed RHESO, but to some extent, the existence of optimized estimation errors makes the electric AV’s braking process smoother and more accurate, which is another advantage of the proposed RHESO-based MPC.
In the following, simulation results under another set of parameters are shown. The initial conditions are given as follows:
ξ ( 0 ) = [ 30 17.3 18 ] T , θ ¯ = 30 , ε ( t ) = 5 sin ( 0.2 t ) + cos ( t ) .
Figure 7, Figure 8, Figure 9 and Figure 10 show the impacts of the proposed RHESO-MPC and traditional ESO-based MPC, ADRC, and RMPC on AV when the initial distance and initial speed are relatively large. From Figure 7, Figure 8, Figure 9 and Figure 10, it is easy to observe that the RHESO-MPC outperforms traditional ESO-based MPC, ADRC, and RMPC in terms of position control accuracy and speed convergence accuracy when both the initial position and speed increase. As a conclusion, from Figure 3, Figure 4, Figure 5, Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10, the effectiveness of the proposed method is verified.

5. Conclusions

This paper analyzed the problem of smooth braking at preset points in AVs. Given the difficulty of obtaining dynamic model parameters for AVs and the limitations of using kinematic models for servo control, an NN was designed to approximate the velocity order model within the dynamic model, enabling the use of throttle/brake openings as control signals. A new extended kinematic model of AVs was proposed by integrating the proposed NN model with the traditional kinematic model, which facilitates longitudinal speed servo control. An RHESO-based MPC method was also proposed to achieve smooth braking at preset points in AVs. Furthermore, sufficient conditions were established to guarantee the recursive feasibility of the optimization problem and the practical stability of the closed-loop system. Finally, the effectiveness of the proposed method was validated through an illustrative numerical simulation, demonstrating that the proposed extended kinematic model can effectively control velocity and position and that the RHESO-based MPC method ensures smooth braking at preset points.

Author Contributions

Conceptualization, J.C.; methodology, Y.X.; software, Y.X.; validation, J.C., Y.X. and Z.Z.; formal analysis, J.C.; investigation, Y.X.; resources, Y.X.; data curation, Z.Z.; writing—original draft preparation, Y.X.; writing—review and editing, J.C.; supervision, Z.Z.; project administration, J.C.; funding acquisition, J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work is partially supported by the Shaanxi Province Natural Science Basic Research Plan 2023-JC-QN-0007, Taicang Basic Research Plan TC2023JC02, and the Fundamental Research Funds for the Central Universities.

Data Availability Statement

Data used in this paper are presented in the Section 4.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
AVautonomous vehicle
MPCmodel predictive control
NNneural network
RHESOreceding horizon extended state observer
PSposition and speed
PIDproportional–integral–derivative
ESOextended state observer

References

  1. Fu, Y.; Li, C.; Yu, F.R.; Luan, T.H.; Zhang, Y. A Decision-making strategy for vehicle autonomous braking in emergency via deep reinforcement learning. IEEE Trans. Veh. Technol. 2020, 69, 5879–5888. [Google Scholar] [CrossRef]
  2. Guo, J.; Hu, P.; Wang, R. Nonlinear coordinated steering and braking control of vision-based autonomous vehicles in emergency obstacle avoidance. IEEE Trans. Intell. Transp. Syst. 2016, 17, 3230–3240. [Google Scholar] [CrossRef]
  3. Wahid, S.M.S. Automotive brake wear: A review. Environ. Sci. Pollut. Res. 2018, 25, 174–180. [Google Scholar] [CrossRef] [PubMed]
  4. Rasouli, A.; Tsotsos, J.K. Autonomous vehicles that interact with pedestrians: A survey of theory and practice. IEEE Trans. Intell. Transp. Syst. 2020, 21, 900–918. [Google Scholar] [CrossRef]
  5. Kong, X.; Li, M.; Tang, T.; Tian, K.; Matias, L.M.; Xia, F. Shared subway shuttle bus route planning based on transport data analytics. IEEE Trans. Autom. Sci. Eng. 2018, 15, 1507–1520. [Google Scholar] [CrossRef]
  6. Serrano, C.; Delorme, X.; Dolgui, A. Scheduling of truck arrivals, truck departures and shop-floor operation in a cross-dock platform, based on trucks loading plans. Int. J. Prod. Econ. 2017, 194, 102–112. [Google Scholar] [CrossRef]
  7. Zhang, Z.; Ding, M.; Ding, Y.; Ma, G. Research on special vehicle detection and passenger elevator docking behavior recognition in intelligent monitoring. In Proceedings of the 6th International Conference on Information Science, Computer Technology and Transportation, Xishuangbanna, China, 26–28 November 2021. [Google Scholar]
  8. Gerdes, J.C.; Hedrick, J.K. Vehicle speed and spacing control via coordinated throttle and brake actuation. Control Eng. 1997, 5, 1607–1614. [Google Scholar] [CrossRef]
  9. Attia, R.; Orjuela, R.; Basset, M. Combined longitudinal and lateral control for automated vehicle guidance. Vehicle Syst. Dyn. 2014, 52, 261–279. [Google Scholar] [CrossRef]
  10. Santis, R.M.D. A novel PID configuration for speed and position control. J. Dyn. Syst. Meas. Control 1994, 116, 542–549. [Google Scholar] [CrossRef]
  11. Ioannou, P.; Xu, Z.; Eckert, S.; Clemons, D.; Sieja, T. Intelligent cruise control: Theory and experiment. In Proceedings of the IEEE Conference on Decision and Control, San Antonio, TX, USA, 15–17 December 1993; pp. 1885–1890. [Google Scholar]
  12. Chen, Y.; Wang, J. Adaptive vehicle speed control with input injections for longitudinal motion independent road frictional condition estimation. IEEE Trans. Veh. Technol. 2011, 60, 839–848. [Google Scholar] [CrossRef]
  13. Cao, Y.; Shi, T.; Niu, X.; Li, X.; Xia, C. A smooth torque control strategy for brushless DC motor in braking operation. IEEE Trans. Energy Convers. 2018, 33, 1443–1452. [Google Scholar] [CrossRef]
  14. Xu, S.; Peng, H.; Song, Z.; Chen, K.; Tang, Y. Accurate and smooth speed control for an autonomous vehicle. In Proceedings of the 2018 IEEE Intelligent Vehicles Symposium, Suzhou, China, 26–30 June 2018; pp. 26–30. [Google Scholar]
  15. Liu, Y.; Xu, B.; Ding, Y. Convergence analysis of cooperative braking control for interconnected vehicle systems. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1894–1906. [Google Scholar] [CrossRef]
  16. Wu, X.; Wei, C.; Tian, H.; Wang, W.; Jiang, C. Fault-tolerant control for path-following of independently actuated autonomous vehicles using tube-based model predictive control. IEEE Trans. Intell. Transp. Syst. 2022, 23, 20282–20297. [Google Scholar] [CrossRef]
  17. Yang, H.; Wang, Z.; Xia, Y.; Zuo, Z. EMPC with adaptive APF of obstacle avoidance and trajectory tracking for autonomous electric vehicles. ISA Trans. 2023, 135, 438–448. [Google Scholar] [CrossRef] [PubMed]
  18. Luan, Z.; Zhang, J.; Zhao, W.; Wang, C. Trajectory tracking control of autonomous vehicle with random network delay. IEEE Trans. Veh. Technol. 2020, 69, 8140–8150. [Google Scholar] [CrossRef]
  19. Nguyen, A.T.; Rath, J.; Guerra, T.M.; Palhares, R.; Zhang, H. Robust set-invariance based fuzzy output tracking control for vehicle autonomous driving under uncertain lateral forces and steering constraints. IEEE Trans. Intell. Transp. Syst. 2021, 22, 5849–5860. [Google Scholar] [CrossRef]
  20. Wang, X.; Sun, W. Trajectory tracking of autonomous vehicle: A differential flatness approach with disturbance-observer-based control. IEEE Trans. Intell. Veh. 2023, 8, 1368–1379. [Google Scholar] [CrossRef]
  21. Maghenem, M.; Loría, A.; Panteley, E. A cascades approach to formation-tracking stabilization of force-controlled autonomous vehicles. IEEE Trans. Autom. Control 2018, 63, 2662–2669. [Google Scholar] [CrossRef]
  22. Xu, S.; Li, S.E.; Peng, H.; Cheng, B.; Zhang, X.; Pan, Z. Fuel-saving cruising strategies for parallel HEVs. IEEE Trans. Veh. Technol. 2016, 65, 4676–4686. [Google Scholar] [CrossRef]
  23. Xu, S.; Li, S.E.; Cheng, B.; Li, K. Instantaneous feedback control for a fuel-prioritized vehicle cruising system on highways with a varying slope. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1210–1220. [Google Scholar] [CrossRef]
  24. Wu, J.; Zhou, H.; Liu, Z.; Gu, M. Ride comfort optimization via speed planning and preview semi-active suspension control for autonomous vehicles on uneven roads. IEEE Trans. Veh. Technol. 2020, 69, 8343–8355. [Google Scholar] [CrossRef]
  25. Patil, O.S.; Le, D.M.; Greene, M.L.; Dixon, W.E. Lyapunov-derived control and adaptive update laws for inner and outer layer weights of a deep neural network. IEEE Control Syst. Lett. 2022, 6, 1855–1860. [Google Scholar] [CrossRef]
  26. Chai, R.; Tsourdos, A.; Savvaris, A.; Chai, S.; Xia, Y.; Chen, C.L.P. Design and implementation of deep neural network-based control for automatic parking maneuver process. IEEE Trans. Neural Netw. Learn. Syst. 2022, 33, 1400–14143. [Google Scholar] [CrossRef] [PubMed]
  27. Guo, L.; Chen, W. Disturbance attenuation and rejection for systems with nonlinearity via DOBC approach. Int. J. Robust Nonlinear Control 2005, 15, 109–125. [Google Scholar] [CrossRef]
  28. Cui, Y.; Qiao, J.; Zhu, Y.; Yu, X.; Guo, L. Velocity-tracking control based on refined disturbance observer for gimbal servo system with multiple disturbances. IEEE Trans. Ind. Electron. 2022, 69, 10311–10321. [Google Scholar] [CrossRef]
  29. Liu, Y.; Wang, H.; Guo, L. Composite robust H control for uncertain stochastic nonlinear systems with state delay via a disturbance observer. IEEE Trans. Autom. Control 2018, 63, 4345–4352. [Google Scholar] [CrossRef]
  30. Xie, Y.; Qiao, J.; Yu, X.; Guo, L. Dual-disturbance observers-based control for a class of singularly perturbed systems. IEEE Trans. Syst. Man Cybern. Syst. 2022, 52, 2423–2434. [Google Scholar] [CrossRef]
  31. Zhao, L.; Zhang, B.; Yang, H.; Wang, Y. Observer-based integral sliding mode tracking control for a pneumatic cylinder with varying loads. IEEE Trans. Syst. Man Cybern. Syst. 2020, 50, 2650–2658. [Google Scholar] [CrossRef]
  32. Chang, S.; Wang, Y.; Zuo, Z.; Yang, H. Fixed-time formation control for wheeled mobile robots with prescribed performance. IEEE Trans. Control Syst. Technol. 2022, 30, 844–851. [Google Scholar] [CrossRef]
  33. Xu, H.; Zhang, J.; Yang, H.; Xia, Y. Extended state functional observer-based event-driven disturbance rejection control for discrete-time systems. IEEE Trans. Cybern. 2022, 52, 6949–6958. [Google Scholar] [CrossRef]
  34. Jagannathan, S. Neural Network Control of Nonlinear Discrete-Time Systems; CRC Press: Boca Raton, FL, USA, 2006. [Google Scholar]
  35. Hornik, K.; Stinchcombe, M.; White, H. Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks. Neural Netw. 1990, 3, 551–560. [Google Scholar] [CrossRef]
  36. Li, M.; Wu, H.; Wang, Y.; Handroos, H.; Carbone, G. Modified Levenberg-Marquardt algorithm for backpropagation neural network training in dynamic model identification of mechanical systems. J. Dyn. Syst. Meas. Control 2017, 139, 031012. [Google Scholar] [CrossRef]
  37. Tao, G. Adaptive Control Design and Analysis; John Wiley & Sons: Hoboken, NJ, USA, 2003. [Google Scholar]
  38. Peuteman, J.; Aeyels, D.; Sepulchre, R. Boundedness properties for time-varying nonlinear systems. SIAM J. Control Optim. 2000, 39, 1408–1422. [Google Scholar] [CrossRef]
Figure 1. The vehicle model.
Figure 1. The vehicle model.
Drones 08 00273 g001
Figure 2. The control framework diagram.
Figure 2. The control framework diagram.
Drones 08 00273 g002
Figure 3. Curves of position components obtained by four methods.
Figure 3. Curves of position components obtained by four methods.
Drones 08 00273 g003
Figure 4. Curves of velocity components obtained by four methods.
Figure 4. Curves of velocity components obtained by four methods.
Drones 08 00273 g004
Figure 5. Curves of the control input obtained by four methods.
Figure 5. Curves of the control input obtained by four methods.
Drones 08 00273 g005
Figure 6. Curves of the estimation error obtained by two methods.
Figure 6. Curves of the estimation error obtained by two methods.
Drones 08 00273 g006
Figure 7. Curves of position components obtained by four methods.
Figure 7. Curves of position components obtained by four methods.
Drones 08 00273 g007
Figure 8. Curves of velocity components obtained by four methods.
Figure 8. Curves of velocity components obtained by four methods.
Drones 08 00273 g008
Figure 9. Curves of the control input obtained by four methods.
Figure 9. Curves of the control input obtained by four methods.
Drones 08 00273 g009
Figure 10. Curves of the estimation error obtained by two methods.
Figure 10. Curves of the estimation error obtained by two methods.
Drones 08 00273 g010
Table 1. Notation used in this paper.
Table 1. Notation used in this paper.
R Set of Real Numbers
Z 0 Set of nonnegative integers
Z + Set of positive integers
Z [ a , b ] The positive integer set { a , a + 1 , , b }
A R m × n The dimensions of A are m × n
A > 0 A is a positive definite matrix
A < 0 A is a negative definite matrix
IThe unit matrix with appropriate dimensions
· The Euclidean norm
0 The matrix full of 0 with appropriate dimensions
λ ¯ ( A ) The maximum eigenvalue of A
λ ̲ ( A ) The minimum eigenvalue of A
x ( k + s | k ) The prediction of x at the time instant k + s from the time instant k
Table 2. Partial parameters of the electric AV.
Table 2. Partial parameters of the electric AV.
ParameterValue
Sprung mass1270 kg
Wheelbase2910 mm
Vehicle height1730 mm
Vehicle width2082 mm
Roll inertia536.6 kg· m 2
Pitch inertia1536.7 kg· m 2
Yaw inertia1536.7 kg· m 2
Brake torque at front wheel250 N· m / Mpa
Brake torque at rear wheel150 N· m / Mpa
Table 3. Parameters used in the simulation.
Table 3. Parameters used in the simulation.
Parameter Λ κ ε ¯ H
Value0.51.193.50
Parameter W ^ 1 * W ^ 2 * a σ b σ
Value−0.21−1.5811
Parameter h ˜ 1 , min , ξ h ˜ 1 , max , ξ h ˜ 2 , min , ξ h ˜ 2 , max , ξ
Value−100 m100 m−100 m100 m
Parameter h ˜ 3 , min , ξ h ˜ 3 , max , ξ h ˜ min , u h ˜ max , u
Value0 m/s50 m/s09
ParameterNQRT
Value10 150 I I0.1 s
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, J.; Xu, Y.; Zheng, Z. Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles. Drones 2024, 8, 273. https://doi.org/10.3390/drones8060273

AMA Style

Chen J, Xu Y, Zheng Z. Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles. Drones. 2024; 8(6):273. https://doi.org/10.3390/drones8060273

Chicago/Turabian Style

Chen, Jianlin, Yang Xu, and Zixuan Zheng. 2024. "Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles" Drones 8, no. 6: 273. https://doi.org/10.3390/drones8060273

APA Style

Chen, J., Xu, Y., & Zheng, Z. (2024). Neural Network and Extended State Observer-Based Model Predictive Control for Smooth Braking at Preset Points in Autonomous Vehicles. Drones, 8(6), 273. https://doi.org/10.3390/drones8060273

Article Metrics

Back to TopTop